

Why Sam Altman is Scared & Why People Are Giving Up on MCP | EP99.23
This Day in AI
What You'll Learn
- ✓OpenAI is shifting its strategy to become an AI platform and ecosystem provider, rather than a developer of first-party applications
- ✓This move signals a shift away from competing at the application level and towards providing the underlying infrastructure and models for third-party developers
- ✓Companies like Cursor are building their own specialized models, which could pose an existential threat to OpenAI by allowing them to bypass reliance on large providers
- ✓There is a lack of a clear platform mode for OpenAI, as the core components like chips, racks, and data centers can be replicated by other cloud providers
- ✓Customers are increasingly seeking flexibility in model choice, rather than being locked into a single provider's offerings
Episode Chapters
OpenAI's Strategic Shift
The podcast discusses OpenAI's announcement about its strategic shift towards becoming an AI platform and ecosystem provider.
Threat from Specialized Models
The discussion explores the potential threat posed by companies like Cursor, which are building their own specialized models to bypass reliance on large providers like OpenAI.
Lack of Platform Mode
The podcast examines the lack of a clear platform mode for OpenAI, as the core components can be replicated by other cloud providers.
Customer Demand for Flexibility
The discussion highlights the increasing customer demand for flexibility in model choice, rather than being locked into a single provider's offerings.
AI Summary
The podcast discusses OpenAI's recent announcement about its strategic shift towards becoming an AI platform and ecosystem provider, rather than a developer of first-party applications. This signals a move away from competing at the application level and towards providing the underlying infrastructure and models for third-party developers. The discussion also touches on the potential threat posed by companies like Cursor, which are building their own specialized models to bypass the reliance on large providers like OpenAI and Anthropic, potentially impacting OpenAI's business model.
Key Points
- 1OpenAI is shifting its strategy to become an AI platform and ecosystem provider, rather than a developer of first-party applications
- 2This move signals a shift away from competing at the application level and towards providing the underlying infrastructure and models for third-party developers
- 3Companies like Cursor are building their own specialized models, which could pose an existential threat to OpenAI by allowing them to bypass reliance on large providers
- 4There is a lack of a clear platform mode for OpenAI, as the core components like chips, racks, and data centers can be replicated by other cloud providers
- 5Customers are increasingly seeking flexibility in model choice, rather than being locked into a single provider's offerings
Topics Discussed
Frequently Asked Questions
What is "Why Sam Altman is Scared & Why People Are Giving Up on MCP | EP99.23" about?
The podcast discusses OpenAI's recent announcement about its strategic shift towards becoming an AI platform and ecosystem provider, rather than a developer of first-party applications. This signals a move away from competing at the application level and towards providing the underlying infrastructure and models for third-party developers. The discussion also touches on the potential threat posed by companies like Cursor, which are building their own specialized models to bypass the reliance on large providers like OpenAI and Anthropic, potentially impacting OpenAI's business model.
What topics are discussed in this episode?
This episode covers the following topics: AI platform strategy, AI infrastructure and ecosystem, Specialized AI models, AI provider business models, AI model flexibility.
What is key insight #1 from this episode?
OpenAI is shifting its strategy to become an AI platform and ecosystem provider, rather than a developer of first-party applications
What is key insight #2 from this episode?
This move signals a shift away from competing at the application level and towards providing the underlying infrastructure and models for third-party developers
What is key insight #3 from this episode?
Companies like Cursor are building their own specialized models, which could pose an existential threat to OpenAI by allowing them to bypass reliance on large providers
What is key insight #4 from this episode?
There is a lack of a clear platform mode for OpenAI, as the core components like chips, racks, and data centers can be replicated by other cloud providers
Who should listen to this episode?
This episode is recommended for anyone interested in AI platform strategy, AI infrastructure and ecosystem, Specialized AI models, and those who want to stay updated on the latest developments in AI and technology.
Episode Description
<p>Join Simtheory to experience MCPs: <a href="https://simtheory.ai">https://simtheory.ai</a><br>----<br>00:00 - OpenAI's State of the Union & Why Cursor's Composer Model is a Threat<br>44:26 - Does MCP Need To Die? Our Thoughts on State of MCP and Why The Client Implementations are the Problem<br>1:07:53 - 1X NEO The Home Robot LOLZ<br>1:28:05 - Greg Brockman, A Sad Song.<br>----<br>Thanks for listening and your continued support. We appreciate you.</p>
Full Transcript
So Chris, this is one of those weeks where I would generally tell people, tune out, there's nothing to see here. I've got a lot of criticism about this at the start of the show, but in fact, it might seem like a quiet week in the world of AI, but indeed, I don't actually think it is. I think it could be one of the most meaningful weeks we've ever had in terms of strategic signals from some of the bigger AI businesses out there. One of those being the bigliest open AI who had what felt like a random sort of town hall state of the union event. It was none other than Sam Altman and the guy who kind of looks like Elon Musk. He's just like he looks exactly like Elon Musk. And he's foreign. So, you know, he's smart. I think they always need someone with a non-American accent in there to make them smarter. Yeah, it's like this is chief gaggle nerd here that's come out to sort of validate what I'm going to say to you. And really, they said at the start of this YouTube presentation, Sam Altman came out and said at the start that it was really about the structural updates to the business because they've made some structural changes where Microsoft is now the single largest shareholder. But they didn't really mention that in the update. and what interested me about this most of all is the fact that he sort of said they're going to pivot like they're changing their entire strategy and it it hasn't got a lot of pickup like no one's really talked about how meaningful this might actually be i'll play two little excerpts and then i want to tell you about some other news that happened this week that um is interesting as well There was a time earlier on in OpenAI where we thought that AI or AGI would be sort of this oracular thing in the sky and it would make all these wonderful things for us. And we now have a sharper view of that, which is we want to create tools and then we want people to use them to create the future. We want to empower people with AI as much as possible and then trust that the process that has been working for human history of people building better and better things with newer and better tools will continue to go on. We can now see a vision where we help build a personal AGI that people can use anywhere with all of these different tools, access to all these different services and systems to help with work and personal life, and their personal life. And as AI gets better and better, as AI can even do things like discover or help discover new science, what people will be able to create with that to make all of society better and their own lives more fulfilled, we think should be quite incredible. There are three core pillars we think about for open AI. research, product, and infrastructure. We have to succeed at the research required to build AGI. We have to build a platform that makes it easy and powerful to use. And we have to build enough infrastructure such that people can use at a low cost all of this amazing AI that they'd like. Here's a little cartoon of how we think about our world. So at the bottom layer here, we have chips, racks, and the systems around them, the data centers that these go into, and the energy. We'll talk more about the first three today and energy another time. Then we train models on top of these. Then we have an OpenAI account on top of that. We have a browser now called Atlas, and we have devices coming in the next few years that you'll be able to take AI with you every day. Alright, that's enough of the presentation. But essentially what was said is that they're moving from this world of them building first-party applications on top of the chips, the racks, the data center, the trained models, and the OpenAI account. And the useless web browser that nobody needs. Yeah, and so they're building these applications on top, And now this huge puzzle piece is going to be the third party applications. And really what it signaled to me is they're saying, hey, we don't want to be out there competing at the application level anymore. We just want all the applications on top of us so we can have the platform, the ecosystem, and essentially be like an AI, AGI cloud provider thingy. And so, you know, nothing like I don't think it's like groundbreaking to come out and say this. Of course, they want to be a platform and be like the Apple tax model on top of everything out there in existence. But what I did find interesting about this is Cursor came out during the week and announced Cursor 2.0. And I'll unpack a little bit of that in a minute. But why I think it contrasts interestingly to this State of the Union is the Cursor announcement said, we're announcing our first model called Composer, a proprietary model that from all accounts versus train, potentially on users interacting with their agent to be a faster and more efficient agent and obviously to eventually make their COGS, their cost of goods sold, go down because right now they're just paying OpenAI and Anthropic basically and don't really have a a fundamental business. But what it says to me is if you build an application that gets really popular and people do a particular workflow on it, it's now becoming possible if you hire the right people to train a model that can do the things that that application is well known for. And so I think that this could be an existential threat to open AI because what it says is maybe there's really no mode in the next couple of years with these models. You can just go So build an app, build it on the frontier models, and then learn how people use it, and then fine-tune for those use cases and improve it based on the very specific workflow or industry niche that you work in. So anyway, there's a lot to unpack here. I just wanted to set it up so everyone's in the loop. But I think this is bigger than we can understand right now. We know, we've known for a while that large amounts of the demand of these big providers like Anthropic and OpenAI do come from these applications that are incredibly heavy users of it. And I said from the very beginning, OpenAI is not a software company. They're great with the research and the models, and that's what really led the way. But if you look at their history of producing software, it's not very good. Like GPT was a disaster. Their interface is okay, and I know everybody's cloned it, but it's not exactly like setting world records in terms of usability and that kind of stuff. Really what they did was came out with the best models and kept them there for the longest time. And then people built on that and they had a good brand, so people would use their stuff. But they've never really come out with something that's blown everyone's socks off and made it their go-to system. I think people de facto use ChatGPT because they know about it, but the people who discover other platforms would prefer to use them. And I actually think them leaning into it is smart. Be the best and cheapest models out there and just have everyone dependent on you for that. It's a much better business model in terms of, like he says, selling chips and electricity and then the value add, which is their models, rather than trying to compete with every single business that comes along. And I think that's been the challenge, right? Like you look at ChatGPT Atlas, where I think a lot of people in the know that have used AI browsers are saying that Perplexity's Comet and then Dia, the other browsers are just better. Like they're better at being AI browsers. And look, I'm sure those teams are capable of refining that and making Atlas as a browser better over time. But this is what it feels like. Like they go and build these products and they're half-baked. And there's other specialist companies that are just dedicated to this problem. And, you know, your point about Cursor is a very interesting one, because if you have a business where the primary input cost is you're transferring a huge amount of the customer spend directly to another company, and it's also something that can really blow out and make an individual customer unprofitable. so let's say someone's paying you a fixed amount of monthly money and you've got to spend a certain amount of that on token usage with open ai or anthropic or whoever then your margin is thin because like you've got to give that person value people generally know what the tokens are worth so you can't have like a 90 markup on the tokens or they're just going to use another platform But if you as cursor can say, look, I can give you 80%, 90% of the functionality, but it's going to cost you half the price, right? People will take that and their margins will go up in the process. So it totally makes sense for them ultimately to run their own specialized models for this and just bypass the big guys. Like, why would you use them? But then that gets me to the case of why would you build outside of the infrastructure, like the cloud infrastructure on OpenAI stack, right? Like they talk about chips. It's like, so what? You bought the most NVIDIA chips. Like that's really it, right? And then you built some racks and data centers, sure. But it's like AWS can do that. Azure can do that. So what's unique there apart from capital? Like they just can raise the most money. That might be one of the things. Yeah, I think when he talks these big numbers like so many trillion dollars in agreed spend and all that sort of stuff, he's really probably talking to the investors and trying to get more money invested rather than customers. Customers don't really care as long as it works, right? Like it doesn't really matter how big their data center is as long as your requests get through. But I think honing in on this point, right? So it's like, where's the platform mode? This is what I don't really understand here because you've got chips, racks, data centers, sure, but anyone can do those things. And the models, I mean, you can run their models on Azure today unless they're going to restrict their models to their own platform eventually, which I highly doubt. So that's, you know, not that relevant either. Also, you know, there's equivalent models now. Some people might prefer, say, a GPT-5, but I know a lot in the sort of commercial sector or enterprise prefer Anthropics models because they're just better at instruction following and tool use. And so there's no real – I don't think there's a real mode around the model anymore. Well, not only is there not a mode, I think there's value in switching. And I think we've seen it, obviously, because our product does that as its main thing. However, even outside of that, I think people like the model flexibility. They want to be able to try whatever the latest one is in their platform of choice. They don't want to go, oh, we're pure open AI and we'll always be that way. And I think that people don't want to miss out on the ability to use different models. So why would you voluntarily tie yourself into them? There's no reason to. This is the thing. I frequently am talking to businesses and enterprises, and the same thing is said to me over and over again, is that their teams maybe have multiple subscriptions and try different models. And they're not stupid. They know, like, oh, one model's better at this, another model's better at that. And so the idea, especially in such a state of flux or in such a state of chaos of locking into a single model, I don't really see the point, right? So then the next layer up, open AI accounts, sure, they've got distribution, and you can't underestimate that. Like, everyone has an account, but only a small portion of that total addressable market for them is paying. And sure, like, they're making a lot of money. Let's not gloss over that. And they have, you know, they're in the zeitgeist. Like, everyone thinks of AI as ChatGPT. So I think that's their biggest asset. That's not what we're talking about here. were talking about what Altman is saying their strategy is in the long term. And he more or less said, oh, well, originally we were going to make AGI that would control the whole world. So don't bother building anything, guys. And now he's like, oh, actually, we're not going to do that. We're going to let everyone have their own personal AGI, which kind of makes no sense. It's either AGI or it isn't. You don't like go, oh, my AGI is slightly different to yours or something like that. The guy also famously said in India, remember that, like, you can compete with us, but you'll lose. Like, basically just mocked them, like mocked all of India saying, oh, you'll never be able to compete or build a model as good as ours. And, like, I would argue they possibly could in the future now, all the tooling's out there. So that entire moat and that entire argument is just evaporated into thin air. And so it's like, what do they really have that's of value left? and it's just that the eyeballs on chat GPT. Yeah, I agree. I think brand recognition and just the fact that so many people have accounts and use it and that if AI is mentioned in media, it is chat GPT that is mentioned. It's not even LLMs or a more general term. It's always chat GPT. I think that's what they have and they need to capitalize on that while they can and lock people in. I just don't think that they really have anything that can lock people in. And this is the challenge, right? as a developer, and I'm not sure what's coming. They talk about an enterprise product, but I would suspect the only way of really getting enterprises to adopt something like this is getting, you know, well, first of all, having the security and just trust factor, which I think they lack as a company, to be quite honest. Well, and they don't even have regional deployments. Like, if they were just like, we're going to build a data center in every major jurisdiction, you know, for international business around the world, then it's going to have the best out-of-the-box security and encryption, storage, all that sort of stuff. And you could do private deployments within those data centers. That would be a way to get to win the enterprise, right? Because you're like, okay, we can do this. But the truth is now the only people doing that are Google, Microsoft, and Amazon. They can do that to some degree, right? But it's similar to regular cloud computing the way you would have done it in the past. OpenAI simply can't compete on that front. But let's go to the things we know to be true today, right? Like if you think about, I think the biggest unlock for enterprise and business today with AI is MCP. Like 100%, like I think it's the biggest unlock because you can securely connect to data wherever it is. You can unlock the value of having all of that data considered when making decisions where in the past you just wouldn't have had the time or the ability to go and do it. You also can unlock the value from what was once a problem, which was all these disparate systems and data stores. And instead of these like pie in the sky projects, like one day we'll have it all in a beautiful data center and perfectly coupled together. It doesn't matter anymore because you just with a series of MCPs link it all together, teach your business, like put a sort of business workflow as the tools on top of the MCP layer. and now you can just query data and make decisions across the org and then test those decisions across various models. So to me, that is probably the biggest unlock. And then you think about third-party applications and where all the data is stored and those companies that would need to unlock it to be out of the box, right? Yeah. And are they going to do it? Are they stupid enough to do this? That's the real question. Because why would you unlock it and make it so simple and build on top of this platform when you essentially, especially something like Oracle or Salesforce, because you're just exceeding control to open AI. You're like, here's our customer base. We offer no value. Yeah, I think that's the thing. I do think that is where all the value is. Who can provide a platform that allows enterprises to do what you just described securely? I think that is the key. is like, how do you get it so the staff can access all of the data and use it to leverage their decision making or take actions or pipe it into other systems through MCP as a platform? Because I think that is actually the crucial part of it that actually adds all the value. And I think even with lesser models, you're getting more value by working in that way than you are just being like, well, we have one that's closer to AGI by 4% than the other model. but it's hard to use the tools with it and it's not yeah i just feel like a lot of the things they're now saying everyone is a bit tired of like these claims about oh don't worry we're also gonna invest compute into solving cancer and all these other things like it's just it it's actions speak louder than words like you're releasing a browser you're releasing some doom scrolling video service where's where's everything else whereas i look at anthropic and it's like Like, they're experimenting around with skills, a better way to package up, like, essentially little applications that the AI has access to so it's better at instruction following. Let's be honest. I mean, that's basically what it is. And they're thinking through the problems in the workplace and the problems people have with AI today and trying to make their models better around that. and the actions there, if the goal is sort of an enterprise platform or these third-party applications, far outweigh the actions of OpenAI today. That's right. Like you can't talk a big game around enterprise and then take actions that are completely opposing that. Like a browser doesn't help with that, definitely. Making TikTok-style videos is sort of an antithesis to what you're trying to portray now, one week later or whatever it is, as your company's main mission. It's just, yeah, it's just all talk and it doesn't really add up. And then the whole like solving, like helping people solve diseases and stuff and like a large language model is really the thing that are going to solve the diseases? I wouldn't think so. You know, like even though I have a lot of faith in them, their ability to create new knowledge from existing data is limited. and I feel like a lot of the scientific discoveries are like maths and chemistry based. It's not really something a large language model is great at. I mean, I could be wrong. I'm not a scientist. Every time they've claimed a scientific breakthrough or some novel thing. Remember when ChatGPT5 came out, someone claimed it solved some novel math problem. That was completely discredited like two weeks ago saying, oh, no, it existed in the training data. Here's the solution. This is completely wrong. And then Greg Brockman and others went and deleted all of the tweets related to this discovery. And I remember reading it at the time thinking, wow, this is a huge moment if this is true. But of course it was not. And so I think... Yeah, and I have no doubt that it can help in the science fields. Like, for example, packaging things into a report, visualizing the data, you know, organizing things, piping things through MCP calls to get things done more efficiently. But it isn't going to, on its own, solve diseases. It isn't going to be like, okay, one day the people trying to figure out the solution to cancer type in, how do I solve cancer? And then one day it just has the answer. That's not really how it works. And he's sort of purporting as if they get enough electrons in their data centers that it will. Yeah, this whole idea of like, oh, don't worry, guys, by March of 2028, this is like literally what he said. by March of 2028, they expect to have a meaningful, fully automated AI research. Like, what a weird specific day. It would be bold, though, to say by March 2028, we will solve cancer with our model. Like, that'd be pretty cool. It's just like, yeah, no, we're pretty sure it'll be done by then. But, like, anyone that's using these models today knows, especially researchers who use them all the time, like, of course there's benefits. That's why they're using them. but it's not yeah it's not coming up with novel ideas like you say it's just helping them make connections and making them more efficient in their research yeah yeah exactly it's like it is individual productivity leverage based on your own expert knowledge like you the expert are guiding it through a process that you normally do and i would argue what mcps give you the ability to do is take steps that you otherwise might not because it would be too slow like for example say I want to make a company decision, do I want to take into account market sentiment about my company and our recent sales figures and our historical sales figures and all of these factors where I may not take that into account on a small decision on a day-to-day basis normally because the amount of work required to gather and organize that information is so high. But if my AI system using MCPs can do that in, say, two minutes and then make the decision, then I might actually do it. So I've got more leverage, smarter decision making and things from this AI system. But the AI system isn't the one setting the goals. It isn't the one driving the narrative or driving what you're trying to accomplish. And I think that is where we want to get to with the next step. But to act like, okay, suddenly the model is going to just do that stuff is wrong because it's never going to be the model that does it. It's going to be the system around it that does it. The real innovation, in my opinion, is going to be around how we package up agentic behaviors and the environment in which we give an agent to operate so that it can succeed. The model improvements will help, of course, because it'll make less mistakes and be more capable of doing things like vision, for example. But the model itself is never going to suddenly just be like, I am AGI today now and I've done everything. Don't worry about all that stuff you built. Yeah, and so the other sort of going, just going back quickly to the platform piece, the other thing is into the video he says about two minutes in, people are going to be able to build incredible services with AI starting with our API, with apps and ChatGPT and our new enterprise platform, right? But this pitch is basically like we're gonna lock you into our ecosystem everyone's gonna use their open ai account i just don't get what the pitch is to the companies and developers who are expected i know what the pitch is it's the people yeah like you build on our proprietary store you will make money imagine the influences and stuff you're gonna see on on youtube like here's how i made $10,000 a day using my fucking, you know, GPT thing I built. And, you know, everybody subscribes because it's got keywords in the title and whatever. It's going to be another sort of junk store. My argument is I don't think OpenAI has what it takes in software to accomplish that. I think Apple does and companies like that, but they're smart enough to know that it's not ready for that yet. I know we're hanging on this topic probably a bit too long, but again sure the audience right But you got to have a good product Audiences aren stupid Like if you go there and the current proposition is like hey ChatGPT make me a Spotify playlist and then the Spotify playlist comes up, and I can only work with one app at a time, which basically forgoes all the benefits of MCP with context gathering and action-taking and asynchronous tool calling, that product's going to be bad. So there's got to be more to this. Like, are they going to have like basic CRUD interfaces, like the SAS sort of create, destroy, edit, update stuff in chat? I mean, I reckon it's almost certainly going to be like a dynamic UI producing style platform. So it'll be tooling around. You can use their models to build something that's going to have dynamic input UI and dynamic output UI that is customized for the thing. And I think if that's what it is, it's totally misguided. That is not what people want. People want to be able to automate common processes at certain times, on a regular basis, proactively, reactively, that kind of thing. And they want the AI to work within little frameworks where they're able to start to accomplish things automatically. Like the example I gave before, where it gives you great leverage in decision making and getting through your own workday, I think the next step is, okay, I'm commonly doing this task, but it requires some intelligence in order to get through the task. I couldn't just give it to, like, let's say you have a junior employee. I couldn't just give it to them and say, go do this. You'd need to train them, right? And I think these trained skills where an AI can accomplish the task that, say, a trained employee could do on a regular basis when required is what everyone wants to get to. Like that is the next step in this journey we're all on with the AI technology. And I think anything that doesn't do that is a misstep by them. Like if they were the first ones to come out with, here's an agentic framework that is going to allow you to automate common tasks and train a system on doing that, that for me would be an advancement. But anything other than that, I don't think is right. Yeah, I think the real benefit today or the part that's lacking with all of this stuff, right is just people figuring out reproducible workflows skills context actions compiling them together as like essentially ai apps that other people can utilize like i i mentioned it last week or the week before on the show around someone could totally build a an mcp that fully replaces i mean you'd need like a companion app for sure like somewhere where you set up your inbox and a few settings and stuff. But outside of that, I think today it's very capable, especially with like MCPUI as proposed by them. You could totally build like a full Zendesk replacement as an MCP, right? And I think that is a good example of an application you could potentially build on a platform like this and get good distribution for. But I would still argue the downsides in a large enterprise are you still want the interface. You still want to be able to go and access the data, and there's still workflows and approvals and permissions and very specific things to different industries and markets. And you're going to then go, okay, well, I still need my core app. There's just a small interface into the AI model, but then I can deliver that similar to Cursor in a very specialist way in my own application. Like, I just, there's no example in my mind, unless someone's listening and thinks I'm incorrect, where this has worked before. Like, look at, in our era, when we were in San Francisco, you had Facebook and you had that gaming company, I think it was Zygna Games or whatever, built a whole business. Like, at one point, valued at billions based on making games on the Facebook sort of SDK like this. It's the exact same playbook, right? Like, there was a bunch of, like, Farmville and all those things on that Facebook ecosystem. And that was dubbed as the future. Like, build on this ecosystem. And what happened? Like, it doesn't even exist anymore. And all of those companies, I mean, they rose up and some did quite well. But then they, you know, everything completely disappeared. And it sort of goes against the whole open web as well. what strikes me as crazy though is either open ai is not capable of making the most of their technology when it comes to the ai systems around it or they think their audience is stupid and wouldn't get it like in terms of what it's there for like if you look at the way anthropic is packaging up skills for example that is a real meaningful step towards a better way of working with the AI. It may not hold, it may not last, but it's a step in that direction, right? Like, it's actually a useful thing that would get us closer to what I just described earlier with the automated task. Whereas OpenAI sort of is a few steps behind. Like, he's talking about next year providing the abilities for enterprises to, say, build apps on there. But enterprises are building apps now, like with any model. Like, it doesn't even have to be OpenAI's model. I just feel like they're kind of behind and I just don't really see how that helps them catch up. Like, again, I don't understand the economics of big business. Like maybe they're just waiting and they want to get KPMG as a customer or Coke or whatever they said in the early days. And that doesn't matter because they'll make so many billions. But in terms of like the forefront of technology, they're just not at it. Like they're just not even close right now. I think if you're losing $8 billion, I think that's what it says a year and you've got to invest trillions in infrastructure, someone's got to pick up the tab, right? It can't be investors long term. I remember when we had committed to Amazon to spend some insane amount of money, and you're looking at it being like, oh my God, we owe these guys so much money. I better be nice to them so I can get out of this. And imagine owing trillions. Our business basically has this massive anchor that's dragging us to the bottom of the sea, and we have to invent something to overcome that and everyone's kind of lost interest in us what concerns me is sam altman single-handedly has set up probably let's be honest the next financial bubble i don't think ai is a bubble in and of itself like i think it has so much practical use and if adopted and set up right in organizations is totally transformative right and as you say i just don't think it's being done very well and but yeah all the hype all the sort of like forward investment the sort of we invest in you you invest in us and we pay you in gpus and you pay us in like tokens like the whole thing is a house of cards and then you like when enron started selling bandwidth at night or whatever it was like we're gonna sell broadband and trade it on the web yeah it's starting to get really bubbly and we'll get to the uh the neo robot later but the the fact that they're losing so much money they're committing fake money like 1.4 trillion he's like we're gonna spend 1.4 trillion it's like they have a lot of billies don't get me wrong but not that many billies like that's a that's a trilly not a billy like these are different these are different when we get the numbers that high i quit it's just not worth it like let's just get gold chains and retire it's just so fake. I said we'd cover the cursor update, and normally we don't touch on these individual tools, but I do want to talk about it and sort of contrast it to what we just said, right? So here's a company that, sure, has built on the back of, let's be honest, Claude 3.5 Sonnet. They just exploited the fact that it was such a good coding model. Forked Visual Studio, which is the the interface that most developers were using at the time, and then slowly iterated and then rapidly iterated to build coding agents, better autocomplete, and just, let's be honest, just better tools for developers. And despite a lot of people using Claude Code and then Codex, which is the OpenAI competitor, Cursors held strong, and a lot of developers I know still use it. so they release cursor 2 and it has some interesting features but i don't think anything groundbreaking like now it can go off and use all models at the same time to solve a problem and then let you decide which approach um especially for large code bases you would like to adopt so this is the people that are wearing gold chains in large companies and just don't care they're and and this sort of validates our point of wanting access to all the models they're like just run it across all the best models, will pick the best variant, and it makes it so much more productive, it's worth the spend, right? They don't care. It's not like people trying to conserve tokens. It's completely different. Think about the cost of a developer. For a senior developer, at least in Australia, it's $200,000 a year, right? Like front-end, back-end, that's about the price. I mean, it might have gone down a bit lately, but around that. So what's that in American, like 140,000, something like that? So if you can replace them with a piece of software and you've got to spend a bit of money on tokens, it's extremely worth it. It's not replacing. It's just speeding up. Like one dev will get you the output of two. Well, I mean, okay, well, it's effectively replacing them then because you don't need to hire them. Think about you trying to pump out three different features at once in your product. You would have to have two different small, like two or three different small teams of devs. If one guy can do what one team used to do, that is replacing people. Yeah, okay. Because they're piers you wouldn't make. But anyway, my point around it basically is they added that feature, multi-file editing, things like deeper code-based understanding, like improvements to semantic search, things that I don't think are that relevant to this show. But what they did also say is, you know, they've announced this Composer tool. They say it achieves frontier coding results with generation speed four times faster than similar models. And they mean similar frontier models, Sonnet 4.5 and GPT-5. They also say that GPT-5 and Sonnet 4.5 offer a slightly higher level of intelligence. But this is really a fast frontier option. And you and I have talked about this before. it's why at the moment i'm addicted to sonnet uh sorry yeah haiku sorry um 3.5 as i know a lot of our listeners who use sim theory are also addicted to for mcps because it's just so fast and it's intelligent enough it doesn't really matter anymore yeah um and so i think they've realized the same thing is when you're in this agent interaction speed's critical like test something try it it doesn't work try it again and they know that these even the frontier models need to do like six iterations sometimes to get it right so really it's a speed game of like going back and forth playing yeah it might it might sound silly but at least for me my personal productivity excuse me is made better by having a faster model because i don't get distracted i'm not like waiting for it to do whatever the hell it's doing and going off and doing something else i'm i'm getting it straight away so i can move on to the next step like it's a far more uh frantic pace when you work with a faster model like that. And as you know, like I'm working on a new feature for Sim Theory, like Vibe doc editing, which is a lot different to I think what's out there today. And when using it, it's just so critical to have a fast model because you just want to, you know, quickly iterate. But then sometimes you want it to go and do longer things like go off in the background and research something and then add that chunk into the document at the bottom while I'm working on the top bit, right? And that's okay. But if I'm like working in line with it and need quick answers, I don't want to sit there waiting for this document to think and update and change. And so anyway, I think the speed is important and the reactions on X just from the quick assessment I did is that people are really impressed with the speed. They're saying tasks are completing in under 30 seconds um the parallel like being able to put agents in parallel is much faster yada yada anyway i don't really want this to be some sort of free promotion for their tool but what i would say about this is i think that this is an example of a successful product that's been built on the backs of anthropic and open ai's models you know and they can call it ecosystem and platform or whatever they want but what they've done is they've gone out, figured how AI can be helpful to developers, honed in on those use cases and workflows. Now they've gotten to a point where they're like, okay, now we can just train our own model because we have all the data for this specialist thing. And that's what they've gone and done. And yeah, okay, it's not the best, but it will get better, won't it? Exactly. But think about the brilliance of the economics of that. You've basically used an available tool at a high cost, knowing that it's a high cost and just accepting that burn, right? That you're spending more than you're making probably on the tokens for a while. Then they would have reached some level of economies of scale where they're like 50% of Anthropics usage or something like that. The reason none of us could get quota for a while, but then they've negotiated a better rate. Then they're like, hey, how about we just cut you guys out completely, run our own stuff. All we need then is the GPUs themselves. And now we have a fixed cost because it's just the same ongoing GPU spend, no marginal cost for tokens. And we just bank money. It's a brilliant, brilliant business strategy in terms of lowering your input costs, but gaining the market share while the technology is evolving. It's not become the playbook for everyone now. So imagine you're building the cursor for finance or financial, like the new Bloomberg terminal that everyone seems to get a heart on over. Wait, I meant to beat that out. You beat the wrong bit. Yeah. But you know what I mean? So, like, they go and they have some maybe MCPs connecting to all the financial data. You can vibe it out with this thing and make investments, and that becomes the new sort of investment platform call it. I'm just making this up. Maybe it exists. Maybe someone's listening and going to build it. So this product comes out. everyone starts to use it because it's just better for investing, say. And then what's the next step? Okay, well, they go train their own model. And so you're really just benefiting on the backs of the R&D that OpenAI and Anthropic have poured billions into. Yeah, and I think that when you think about it like that, if you were a major institution, why the hell would you ever sign a deal directly with one of the major providers like OpenAI or Anthropic? Or who's the other one? Google. because something better and cheaper might come along. You might want to develop your own. You might want to run your own. You need to keep that flexibility there because locking yourself into one means that you're guaranteed to pay the maximum and not necessarily have the best model in the long run. So I think that it presents a major challenge to those people as platforms in the sense that you've got this very, very healthy competition. And like you say, there's pretty much nothing stopping except for money. Someone coming along with their almost frontier level model. I mean, XAI proved that, right? They just came along, they're like, let's throw some money at this and just make our own. And then they got one that's almost as good as the others. It's probably obvious to OpenAI too that ChatGBT, while it's in the lead now, like AskJeeves was at one point, just kidding. But, you know... this lead can be eroded quite quickly right like if google ever get their act together and release a decent assistant um or apple get their act together and release a decent assistant and put it everywhere it just starts to erode the like yeah maybe it's not going to kill their lead immediately right but it starts to erode it because people are like oh ai's here ai's there ai's everywhere like i can get this chatbot experience literally anywhere yeah and it's it's good enough. Like the image editing, the image, um, creation, the video creation, like whatever it is, it's all the same. And so it gets to that point where it starts to get eroded. And then you think, okay, well, what, what doesn't get eroded? It's specialist workflows and applications. These do not get eroded. Like they, they never will be. And he acknowledges that in the excerpt I played at the top of the podcast where he says, you know, humans have always been building better tools and, you know, things to do stuff, mine gold. And so he's now saying, oh, we admit that that's the future. It's not some magical AGI box that we fantasized about. And so they've realized that it's these specialist applications and their pitch is like build them on top of us. But I don't get the pitch. The pitch doesn't make sense. I think like for getting open AI for a second, I think that people in companies who are now building these AI systems around processes in their organizations or exposing data are basically taking the steps needed to get to some level of automated intelligence in the future. Like these are not throwaway things that will need to be replaced. Like if you can teach the AI to do a fundamental component of what is done in your industry with intelligence, then that can be a building block later of a wider AI and smarter AI system. you're building towards it. Having a smarter chatbot is not going to accomplish that. Like it's the stuff around it. It's the arrangement of the data. And I think you had a topic we were going to talk about today about building smarter MCPs. And I think that is a big part of it is like, it's actually the architecture of how you arrange these skills and tools that will lead to, you know, better intelligence later. Yeah. And to really just sum up that whole conversation and why I think it is a big week in AI news or at least what will impact the future of everything is that this cursor model is a big blow to OpenAI and Anthropic and Google and whoever else, because what it says is there's no moat. Like we can basically build products on top of your APIs, get to economies of scale, and then make our own model. And look, I'm not sure how this plays out, if people will eventually just prefer this composer model or not. But I would think maybe they will. And then you think about Mira Moretti and her Thinking Labs company that allegedly is building it so you can just train your own models, right? Like you can just build your own custom models for your apps. if a company comes along like that and offers a service to future app developers where they're like, hey, you're currently using OpenAI and Anthropic. We can study the workflows that your users use and then train a custom model for you. And then it's yours. Like that's our service and we maintain it for you. Yeah. And I mean, to put that into perspective, let's say your business is spending 30 to 50,000 US dollars a month on the big players models. And someone could come along and be like, well, we can actually reduce that down to like 5,000 a month. running on a couple of GPUs. That's very appealing for a business. And you're going to get most of the performance. And because it's specialized to what you do, it'll actually be faster and more accurate. I mean, it's a very appealing proposition. And I don't really see how the big players defend against that, except for with good marketing, talking about how hold us back from taking over the world. One thing that I questioned is, where was Greggy boy? Where's Greg Brockman, co-founder? The only remaining co-founder left at OpenAI other than Sam. Sam's bringing out the AI Elon clone to the table. Oh, Greg. And he's up there talking about company structure, vision. Anyway, there'll be a sad song at the end of the episode about Greg if you want to listen after the credits. Oh, sweet. I haven't heard it. That'll be good. It's going to be good. All right. So next topic, as you alluded to, mine's a bit more clickbaity than what you said. which is, does MCP need to die? And I base this whole thing on a tweet by, you know, informer SigKitten over on X. He said, MCP just needs to die. Anthropic's already abandoned it. And he's showing in the Anthropic UI trying to add a custom MCP connector and their interface just crashes. It doesn't work. And then he said, you don't want to see the issues with the MCP in ChatGPT. And he says, in ChatGPT, you need to enable developer mode from the obscure place and settings, if your account type even has it. Then you have to go to the connector screen and not miss the weirdly placed create button. I agree with him. It's very weirdly placed. And then he says, then have your server running so it can maybe connect to it. And guess what? if you restart your server, you need to come back in here, disconnect, connect manually, or your chats are going to fail trying to connect. In Claude, it still sucks, but at least the MCP connection is done in the browser, so you can just refresh the page. Then make sure you select it each time from the overloaded plus button, or ChatTard won't know about it. So you've got to focus on the MCP you're using, or ChatGBT has no idea. And then he says, oh, and my favorite part, when you click disconnect if you want to do anything that closes the dialogue it's gone your mcb connection disappears and you have to use a different wow i mean this makes this actually increases my self-esteem quite a bit because i feel like in sim theory we're like a hundred times better than that i um i can't believe that that's the state of it in there yeah he's like literally no one's gonna go through and use this shit and i think this is like the bit with mcps right is it's so hard, so inaccessible. These companies have completely failed to make them work well that everyone blames MCP, like it's MCP's fault as a protocol or an app. Yeah, and I think this is the thing. Every time I show someone how easy it is to set up their own and how well it can work and quickly, they're blown away. Maybe that's why. Because even if you're setting it up to, say, claw desktop, there's a lot of steps. And I think the other thing is that because people inconsistently adhere to the protocols that there a lot of fiddling around on a per MCP basis Like you hear one announced you like that great I love to add that And then you realize that there all these different things you have to go through to actually get it going. It's not as plug and play as it appears on the surface, even though it should be. But I don't think that dismissing the whole thing because of a poor implementation like that or a poor way of working with it is right, obviously. because I think that if you work with it well, we've shown like the ability to have parallel tool calls, one-click install and nice OAuth connection processes on them. The power is massive, like of MCP. It's just that unfortunately they're being presented poorly in these major providers. Yeah, it's just like two of the biggest companies. Yeah, NN Throbic who invented it. And you see this reply I had up on the screen earlier. Someone replied, looks fine to me. And it's a screenshot of like 17 errors. that come up based on MCPs not connecting correctly when they load up Claude. So, yeah, it's a mess. And quite frankly, and I think to say that, as users of Sim3 will know, it's taken us a good two months to refine the experience, to get it now where I think it's mostly reliable. And sometimes it's not. You're relying on so many providers and so many different variables that it can be quite difficult. But I think we're at a point now where it's reliable enough that I use it all the time and don't really think about it that much anymore, especially if you refine assistance with a limited MCP tool set rather than just trying to smash them all out of one. But I mean, some of that has been us having to rebuild the MCPs ourselves. Like we're like, OK, that's a poor implementation. Let's make our own version of it that is better. So it's taken a bit of that. But I feel strongly enough about it that it's worth it. Like it's worth getting them right because the productivity gains from people using them is so high. And quite frankly, they're not that hard to build properly. It's something that it's more about the way you work with it that's important than the thing itself. However, I think I mentioned earlier, the other thing that's bad about MCPs is that I think the people making them in some instances don't really understand the way they're being used in that they're building it like it's just an API with a series of functions and sort of hoping that the AI will figure out, okay, well, I can get this parameter from this tool call, and then I'll pipe it into this tool call, and then after both of those are done, I'll call the third one. But the problem is that isn't a really nice way to work with the model because it's going to have to go through three or four iterations to actually get to the thing you asked it to do. My opinion is the tools need to be more like individual procedures or accomplish a sort of range of things to get a task done. And then the AI can then feed them all the information they get to use that. So an example of that is, let's say it's your Outlook calendar, and you say, make a meeting with Mike next week at 9 a.m. If you think about the steps involved, it needs to check that it can access your calendar and that you're free. Check which calendar it's referring to in case you've got multiples connected. to look up the contact of the person they're referring to to check that it has it and it can add it. You know, that's three or four steps. Now, if you just have an MCP that has those as individual things, it's going to take a while. And there's each time it takes a step, there's a chance that the AI will make a mistake, right? Or pause or do something dumb. If you can combine that as one thing, it's like, this is how you make an appointment. You do the following four things. It's much more efficient and much more likely to succeed. And I think that the MCP design needs to get more intelligent in that way in order for it to be perceived as better. And it's interesting you say the design because there was this post on X by David Kramer. He says, I'm so tired of would-be AI influencers, that's us, chilling GitHub as an example of why MCP is bad. And that's why CLI is better. You have all done a disservice to the community by miseducating people. GitHub is a really poorly designed MCP. The other thing he says that I'll call out because we can explain why it's bad is, so just because GitHub and some others have built bad implementations on top of MCP doesn't mean the need for tools and the metadata required to operate said tools is the problem. We are still early in the lifecycle, and there's a good chance we'll be fine tuning agents on top of our tools to improve on this, but any version of training is still lossy compression and will be more error prone. so please stop spewing MCP is bad because some random dude on the internet who doesn't work for a living told you so. Yeah, so the problem with GitHub is it has like 100 tools in it alone. So if you enable GitHub as an MCP, your model has to contend with going through the definitions of 100 tools to decide if it should call one or not. And then even then, they're tools like I described where it's got to like look up the repository name from scratch every time. It's got to look up something else. There's no idea of like, okay, we're working in this context, so this is the set of tools I should be working with. We have that in Sim Theory. I filter the tool calls to give it a more curated selection of things that, you know, here, sir, here is your tray of whiskeys or tray of cigars. Pick your favorite rather than like there's a wall of books and games and garbage behind me. Go read them all and decide which one you're going to use. So straight out of the box, enabling the GitHub MCP is going to make your experience working with AI worse. There's no other way to describe it. You're moneying the waters. It's going to cost more. And it doesn't even work that great when you use it. So it's a canonical example of a poor MCP. And it's basically useless for that reason. Absolutely useless. And it's one where we're like, okay, we'll have to rebuild that. We'll have to do their job for them because they've done it so poorly. like do you really want every time you type a command to your ai model it to decide if it wants to comment on a public issue on github because it has that ability you know like do you really want it to create a gist of what you're talking about and things like that don't understand coding lingo but understand like this sort of um you know ai chat style interface imagine every time you like But like the classic example of hello, saying hello without any form of filtering would need to do the following. If you had the GitHub installed, you're talking about a prompt that guides the tool use alone and the tools and like the, all the examples and stuff in them of pasting like a full blown essay in. When you say hello, above that there's an essay of like, Hey, here's all the stuff you can do with this tool and then multiply that by like 20 or 10. and this is why I think obviously OpenAI have just restricted it to one, and with Anthropic I think it's like a couple, is because you're putting essays in every single time. And so that bit is just so dumb. But you can fix it so easily. Like one of the experiments I did during the week is if you have a tool, which is like Get Schema or Get Info or Get More Tools, so you just have the most critical tool listed, like for say image editing you might have um create image uh you know or edit image sorry one tool listed and then have another one more image tools and so then the ai can decide to go and discover more tools and you've basically just described anthropic skills yeah which is which is exactly the way that works like they give a sort of top level here's a thing it can do And then all the additional prompting and mechanisms for running that are encapsulated within that skill, but not included in the primary prompt. And so, yeah, you've made that, but in tool call form. And I think this is why people are excited about flawed skills, because you're encapsulating or you're teaching people and you're getting the help of the AI if you build it through the AI to encapsulate those things that people would be building. So instead of me as the developer with my MCP trying to figure out like your workflow, everyone's workflow, you now can then take, well, at least in the implementation we want to do, a series of MCPs and instructions and then tune your own skills so that you're teaching it how you want it to operate it. Yeah, and then keep those tools within the context of that skill so you're not presenting them to your main AI model every time you do anything. Like, it doesn't make sense to do that. It makes sense that they would be clustered and working in a certain domain rather than just this global thing where you have all these tools. I mean, we have people with literally 500 tools enabled. The AI won't even take that. I think the limit's like 125 on Anthropic, for example. So there has to be a level of filtering there. But I also just don't think that filtering is the future. I think it's what you're describing, which is these curated skills that contain sets of tools within those to accomplish specific procedures or specific decisions or research or whatever it is. It's this evolution and iteration where I think people are quick to come out and say, oh, MCPs are terrible. But I think it's just the mainstream experience with it. I've seen in the Sim Theory community firsthand people starting to say, like, hey, the MCPs are good in here now. Like, it's running well. We're doing asynchronous stuff. We're getting stuff done with these. And I think that sort of validates all the work we've been doing, figuring out, like, how do people want to use these, listening to, like, quite frankly, just constant criticism. Yeah. And responding to it and just taking it on the chin and going, okay, we agree. Like, let's make it better and better. And then now I'm at a point where like, you know, I'm using it again for literally everything. Like, you know, when we do anything in the overall system or respond to a support ticket now, you can just say go off and figure all of it out. Come back to me and I'll approve it. And I'm using it like I shared an example with you earlier in the week. I'll try and bring it up on the screen for everyone that watches all two of you. And basically what I was able to do is, let me just bring it up here. Oh, wrong screen. Hang on. Yeah. So basically what I was able to do is I had a power bill that was sent to me by my energy provider. I've talked about it on the show. You can tell the problems we have in our country that both of our main examples of huge expenses in life is our power bill. Yeah, and so... We couldn't afford to run trillions of GPUs, that's for sure. Not here, no, even though we have all the energy in the world. But anyway, so I quit this energy provider. They sent me a bill for what would be less than a single month, and normally my bill might be like, say, 200 bucks a month, and the bill was $732. Oh, you have solar. That's right. Yeah, anyway, but this is not my point. So there's this huge disparity, right? 500 bucks. Now, I am so busy in my life. I normally would just pay it and just accept it and be like, oh, they couldn't have got this wrong. There's no way. But I was curious. And so I asked the AI with MCPs to go off and research it. I cut and paste in my solar data, like which actually tracks my energy consumption. And the AI is like, it doesn't match. Like this doesn't match up. like your energy usage tracked in this application does not match what they're saying. And then they sent me a few days later an email saying, you know, we got it wrong. We apologize. We've basically stuffed up your bill. And I thought, oh, good. You know, they've corrected this mistake. I'm glad I hadn't paid it. And then it said, we owe you an apology. And then they came back and reduced the bill by $4 and said we stuffed up on the solid credit. So anyway, I'm like, that's it. I'm going all in. So I fire up the AI lawyer. I have an assistant called Lawyer, which says you're a really, really, really good lawyer. I think that's seriously the prompt. There's a prompt engineering guide for you. And I'm like, here's all your tools. You got access to my email. So it went through my email. And in the new Gmail that you build, thankfully, it can read attachments. So it went through, read all the historical power bills. I think in a couple of seconds, it pulled up like 20 or 30 of them. It was able to go through those PDFs, read them all, put together a case that quite frankly, I was like this, you know, maybe it's hallucinating. Maybe it's not true. But I sent it to their, you know, I sent it. It also looked at the legislation and my rights as a consumer. So I emailed them the most legally summary and breakdown of like every problem on earth and was like, you know, I could basically take you to this tribunal and get $100,000 payout. And then it listed all the case examples of where this particular company had lost when this had occurred in the past. Wow. Anyway, then they came back to this email, not realizing how good my really, really, really good AI lawyer is. How much money it would cost a real lawyer to do that research and draft that letter for you. It would be thousands of dollars. It would never happen, let's be honest. So anyway, they come back and they're like, you know, as a gesture of goodwill, like we haven't done anything wrong here, but as a gesture of goodwill, we'll give you 50% off the bill if you pay it and end this matter. So I wrote back, Rahul, your 50% offer is rejected. And then, you know, whatever the AI lawyer wrote. I didn't write any of these emails. I don't have time to do this. I was just like, respond. You're just like a spectator in your AI agent doing your dirty work. My wife was like, oh, you probably should just settle this. Like, you're wasting a lot of time on it. And I'm like, no, I'm not. I'm just telling you what it's wasting a lot of time on. So anyway, it replies. And sure enough, it gave them a deadline. It decided that it was this Wednesday at 5 p.m. And I was like, wow, that's really aggressive. That is a really, really, really good lawyer. Great lawyer. Really, really, really good lawyer. And so then I get back to you, Michael, thanks for your email. I have waived the debt of $732.13 as a once-off resolution for you. Based on the above, I have now closed this matter, and it is resolved. Yeah, as once-off, like, don't use this as a precedent for other people. So my point is, around this, is the power of working MCPs. This isn't saying, like, obviously AI is the equalizer, the great equalizer for the consumer, where these companies could basically take you for a ride in the past because they had all the resources you didn't. Now you have all the resources they do and more. And without bragging, I think that Sim Theory actually made that possible with the way we implement the MCPs, because if you think about it, let's say you were using ChatGPT, you would have had to do way more copying and pasting. You would have had to go into every email and get the contents from the PDF or upload the PDF or whatever, right? Or you would have had to be directing it to call each MCP individually, like look up the laws, you know, analyze this PDF. You know, it would have been you just basically guiding it through the process. You probably, again, even that probably would have been too much effort to actually go to, I would imagine. Yeah, I mean, there's just no way I would have invested the time otherwise. And I guess this is what I mean, because our experience of MCP is this. And everyone else's, like, the 99.999% experience has been like, how do I even access it? What even is it? Like, I would argue if you're just a regular ChatGPT user, you would have no idea what it even is. And even in Anthropic. It looks like Chicago for you. But even in Anthropic, like, sure, you can connect to email. can't read attachments. And I'm not saying like that's some killer feature. I'm just saying these are the workflows and things that people actually want to do with AI, right? And just expect to work. And this isn't to sort of, like this isn't, I know it sounds very promotional, but it's really not. I'm just saying like, why haven't these larger companies with far more resources than us and quite frankly, a lot of talent realize that this is the benefit. Like, this is what is good. The one thing I wonder about it is, I mean, the whole reason we started this podcast was because we wanted to actually use this technology, and therefore we're talking about it. And I think a lot of what's good about the MCPs and sim theory has come from us and, by extension, our community actually using them. And like you say, constant criticism. like this didn't work, this was crap, the output was garbled, it did something dumb, and then us just going back and be like, well, we know the model is capable of better than that, and going back and making it better. Or where necessary, completely replacing the MCPs. Just this week, for example, I replaced the YouTube one to be much better at searching and much better at getting transcripts and analyzing them because people weren't using the previous one because it sucked. And so I think that you can't just go, oh, well, that's the YouTube MCP and just end it there. I think that we have to see it as very, very foundling technology that needs some attention on it to get better. But people need to look at the possibilities. You really need to think, what if I gave it a series of tools that could do this for my company or whatever system I want to work with? would it be able to do good stuff with it? I think in most cases the answer is yes. And therefore to dismiss it outright and say, oh, MCPs are dead, is ridiculous and wrong. That's the Sim Theory promotional song I made. So it was a promotion after all, disguised as content. Do you want to know something really interesting? I always am alluding to the fact that the tracks are now all on Spotify. We've actually risen in the rankings. People are getting songs like Love Rat. You know when they do Spotify DJ? Yeah. They are actually getting Love Rat. For those that don't know the beauty of that song. So they're listening. They're on their commute home from work. Here's the play. you know they're on the train you know if I'm ever at a cafe and I hear that song come on I'll know we've made it they're on the train they're commuting home and all of a sudden they're listening and this song sounds sort of you know like a mainstream track and then they're hearing stuff about like Jeffrey Hinton I had it on in the car the other day and everyone was just like enjoying it and then it said something about like GPT-4 or something and then the kids like is this AI like yup Anyway, all the songs available in the album, AI diss track collection and then average tracks from the show. You know what we should do? We should do like a Triple J Hottest 100, but for AI songs and let everyone submit them. And we, like no voting, we pick. You know, the other day someone reminded me about, remember that original? It's on one of the shows like from long ago. that song um it was like that first country one we heard about ai that really gave us it spawned the idea for all these songs i think oh yeah yeah have you got it i remember that i'm not that good i'll try and pull it up where it's like sorry but i won't do that yeah as an ai language model i can't do that you know back when rejections were a thing uh wait they still are refusing and probably uh all right moving on uh moving on because this has gone on way longer than i thought for a quiet week. At least you didn't tell everyone to skip it. Well, I sort of did. I mean, I gave them a little warning. So in the future, Chris, in the future, there's going to be very muted colours. I'm thinking like, you know, tans, whites, all that kind of stuff. Like non-confrontational colours. Yeah, non-confrontational sort of like beta colours is how I would describe this. So this company, One X, they have investors, including OpenAI, is one of their investors, of course, because they've got all the money. And so they released the, or allegedly are releasing in a year from now, the Neo Home Robot, which you can order today or subscribe to for $500-y dues a month. And I'll mute it because the audio is painful. Can Australians do it, or is it just America as usual? I don't know, maybe we should I think we need a company one I'll get on that I'm sure someone, one of our listeners will accept the shipping of the robot to their house in the US if we can't but anyway, it's this like whimsical sort of old-fashioned-y video, we're in San Francisco and this robot's yeah, like doing stuff with them, like, you know, it gets a cup it opens the door to the house and they alleged that it'll be able to fold clothes, tidy the living room when you're out, pack the dishwasher, unpack the dishwasher. And don't get me wrong, I want to live in a world where all of this is true. But in the demo videos, there's two scenes, one where it takes a cup off the guy and another one where I think it opens the door where it's actually using the AI model to operate the robot. the rest of it is tele-operated with a guy literally using a VR headset in another room just acting like he's a robot and I get the whole fake it till you make it tele-operator thing but what I believe they're proposing and it's all speculation on my part is that you could in the future for some tasks have like a tele-operated robot in your house so there's some guy in India with a VR on I'm like, I don't want to do the action. He just probably looking for crypto where you store your crypto keys Yeah Something like that So like anyway why would anyone on earth let this thing into your house if it can be teleoperated? But, you know, it's been... See, I thought, because I read it that it was a paid out-on to do some stuff. I didn't realize the majority of the stuff was done in that way. Well, I think they're claiming that it'll be able to fold laundry, vacuum, and carry groceries. but apparently to fold it took um so demos showing it taking two minutes to fold a single shirt while struggling with balance so it's sort of like a drunk elderly person yeah pretty much and it says critics are calling it a modern day mechanical turk the 90 the 1700s chess playing automation that was secretly a human inside and demos show it taking out two minutes to fold laundry, the privacy concerns, people are calling it creepy, a privacy nightmare, and questioning why anyone would pay $20,000 for what's essentially a remote-controlled robot operated by strangers who can see inside your home. I bet you it sells out, though. so humanoid robotics is a category companies, this is outside of Tesla's robot and some of the internal projects 3.2 billion in 2025 alone this is the first robot I've seen where the main demo isn't someone just kicking it it seems like all of the main robots their main job is to just take a punch and take a kick I believe this will happen I don't believe this company is Look at what happens to this robot if I hit it in the head with a cricket bat, it responds perfectly So you can have your very own next year apparently for $20,000 or $500 a month The one thing I'll say about this is I got to experience last weekend the Australian version of FSD right because my car's too poor to um get that software and so a friend of mine he had it we were away for the weekend and he was like you got to check this out so we get in the car and i'm thinking you know my expectations were very low not because i didn't think it would be good i just thought oh maybe it's like one of those things like you watch the videos on youtube of people in the cars and you and and you're like wow that's amazing but then when you're experiencing real life there's all these quirks and problems with it and you're like it's not that great yeah anyway he's like yeah the door keeps falling off but we're working on that yeah here's my genuine experience with it i had an existential crisis it drove around roundabouts it drove us to woolworths in this remote town there was like road work where you had to cross onto the other side of the lane like people with stop signs signaling it drove us all the way there comfortably more comfortably than a human um drives into woolworths which is like a safe way in australia goes into the parking lot and finds a park and then reverse parks into the park. And I don't know, I would equate it to like losing your virginity. Like that was the experience I felt after. I was like, hang on, what? I was like, this is like, I'm sorry, like a late adopter here. But I'm like, this is going to change everything. Like every other manufacturer, like if no one else can copy this, get to this standard. Eventually, once everyone experiences this and they perfect it, which I'm certain will not be that long, maybe five, ten years. I know people might disagree with my timeline, but I reckon it'll still take another decade before you can go to sleep. Sorry, let me clarify. Go to sleep with your family in the car on a long trip and young children. And trust that you're safe. I was like, this is going to change economics like how how much could this company be worth like trillions because if if this if they truly are ahead right and and this thing is this software can be applied to like other vehicles like trucks and transport and everything big one right like if you can do road freight with no drivers like that's going to be pretty incredible so i'm a huge believer in this stuff and i think the humanoid thing sure will eventually get there but think how long it's taken tesla to get to this point it it's been like over i think over a decade um or maybe just under a decade and there's still edge cases from what i understand where like it freaked you know it can freak people out or make mistakes sometimes and in fact it did make one mistake i'll call out where it caught fire there was a left turn and um my uh my friend he he's he does like admit he never washed his car ever like he he does he has he he has a certain color choice in cars he's both these cars are the same color so they can remain dirty and no one knows like that sort of like dark gray color where it just looks dirty naturally and uh anyway we were driving into the sun it's making a left-hand turn and it panics and slightly goes onto the what you would i mean the wrong side of the road but in the sense it's one of those like estates with a really narrow road and it it just sort of goes over to the right when it should have gone to the left and around this like sort of hub thing in the road the wrong way anyway like if it's something a drunk driver would do basically so outside of that like one thing everything else was incredible may have just like killed someone or a pedestrian i don't think so i don't think it would have killed anyone but yeah so so that one thing but i think similar to what we experienced with llm's right is moments of brilliance like 98 brilliance and then that other 2% where it fails and you're like, God, this thing's dumb and we'll never get better. And I think the self-driving car thing, that's sort of where it's at right now. And overcoming that last sort of 2% is going to be the hardest challenge ever. Similar to these robots, they're at like the 1%. So they've got to get to the 98% before they're useful. I just think this stuff is still probably like 15 years away if I had to put a number on it yeah this is like that what was that uh rabbit rabbit ai thing it's like the equivalent but in robotics right the scams are just getting bigger yeah it's like we know that people want this so we can sell something that kind of looks like it does it but actually because i just think about things like in the house that you know they're not easy like our door lock right is a bit stiff it's kind of hard to use and i'm like is the robot just going to persist with trying to open the door for like 45 minutes, trying different combinations like, oh, sorry, I appear to be having an issue with the door. And then like, just keep trying. Like, I like the idea of one that's just going to escalate things like I'm going to go get a hammer and that will open the door or something like that. The other problem I noticed with it, this sort of fabric look it has, I can imagine it getting like wet and dirty. Like if I do the dishes, I spill water on myself and I just wonder if like, if it gets damp, it's going to stink. Is that the image I'm having? None of this matters because this is just a pump and dump scam. That's all it is. Like this thing, I just, if this, if this is it, if this is it, if this is the big breakthrough, this is the robot. I mean, I can't go on I wonder if for the people controlling it, you could have celebrities or Twitch stars doing your, I'll be your house husband for a day as your robot. And that's like a perk they can offer people. So it comes in three colors. They're interesting color choices. There's a white sort of tan one, which they used in all the demo videos. There's a black one. I don't know why that, like given slavery and having slaves in homes. And it'd be an interesting choice if that was the popular one. Yeah, like, who's ordering the black one? I just, whoa. And then there's, like, sort of a slightly sort of, like, I would say the middle shade between. It's not brown. It's not brown. It's, like, a gray, a gray color. So, I mean, I think the thing is, like, part of me, even though I recognize that it's probably in practicality not there yet, the idea that it might work is just so appealing. I just think that there'll be enough people out there willing to spend 20 grand, even if it can just do one or two things. Like if it could just fold the towels or it could just unload the dishwasher. Like if it could unload the dishwasher, that's pretty useful. This is similar to the thing Apple released, like the Vision Pro, where like you use it a few times, like, wow, it's a glimpse of the future. And then you put it in a drawer. Like I can imagine this thing being put in a body bag up in the garage. what else are you gonna do with it i guess you could talk to it the whole companionship for old people thing is pretty cool like if you could sit it with someone you know who's got dementia or something and it'll chat with them and remember their stories probably get sick of their stories as well. Scam them out of their money, tell the operator. You freaking told me this already. I know it better than you do. But yeah, like that kind of use case is actually kind of cool. Like even like you think about an elderly person who wants to live independently at home, right? One of the big risks is they have a fall and nobody finds them. This thing's going to have a fall, not the elderly person. Well, yeah, the robot. Well, vice versa, they can have each other's backs. But like, I like I reckon there'd be a segment of people out there who'd be like, look, I don't want to put them in a home. But if I chuck this robot in, it's going to, you know, make their eggs in the morning. It's going to have a chat with them and it's going to alert me if anything goes wrong. Like, that's quite a useful. I 100 percent am with you. I get the appeal. But I think what you're saying is 10 years away before, like before we get anything like that. Like, I don't see. I don't know. I just get the feeling that it'll happen. And when it does, we can't fathom the economic impact, like my experience in the Tesla. But I think this is peak AI bubble in terms of raise billies, hump and dump, get a bunch of orders in, raise more, cash out, and then the bubble bursts. People realize this stuff's a long way away. I'm very much looking forward to the fail videos from these things when people start to get them. It's not like Kickstarter, is it? It's not like the real funded business. No, I think it's reasonably legitimate. They've got a decently designed website, although that doesn't, you know. The one thing on their website is you're right about this old person companion. It's this old guy hugging the robot. I mean, what him and the robot are up to, I'm not sure. Yeah, it looks like he wants to kiss it. Yeah, it really does. There's nothing to kiss. yeah and nearly all the examples are like the elderly couple are playing cards and these robots carrying stuff around but you could actually see this going another way where they realise the AI is just simply not good enough which I think they kind of have already and instead it's just tele-operated you know slaves what if you bought like 10 of them and have them act out of play for you that'd be fun we've got to get our hands on one of these even if it is tele-operated We have to. It's worth it for content. And this is probably why it'll be successful. Did it take your piss in the toilet? No, I was pretending to claim it. I don't know why this has been the biggest lol ever. So this is when some dude is remote operating, is that right? Yeah, so all these ones are remote operated. Yeah, the laundry. Even the laundry example. That is so sad that someone's sitting there in a $1,000 VR headset and he's painstakingly putting your dishes away over the internet. Yeah, because you don't want them in your house, but it's fine if they're through the robot. But then it's like they're seeing everything. It's a real person. If you can throw away 20K on a toy, then maybe you could just afford a full-time maid. like i don't understand the benefit of this right now i just don't believe it'll ever ship or work or they'll ship it we'll get a lot more lulls and then it'll be like one of those like you know we're sorry videos maybe or is this like juicera you know where like the juice was already you know ready and then the the machine just squeezed it is this juicera in the sense of like they're just going to tele-operate these things and you pay 500 a month and they give 100 bucks a month to some company in India to operate them. Yeah, but I mean, I guess it still has the like dexterity and motion and stability to perform those tasks, which seems pretty good. Dexterity, I don't know. Watch the Jonas Stern video. It's okay. I think the examples I've seen from one other robot company and then obviously the Tesla robot seems just so far ahead. I mean, the Tesla robots, there's video of them being used on the production line to assemble things. I mean, they're trained on those tasks, obviously, but yeah, anyway, I want one. Don't get me wrong. And I want this to be true. And like, I want to believe. Yeah, it's sort of like, I guess that's what we always talk about on the podcast. Like when we used to make, remember in the early days, making the images and like people that have, you know, broken hands and weird stuff, but you just want it. It's like the same with this. It's like, I get it. I understand it's a scam and it probably won't work, but I still want one quite badly. All right. Any final thoughts on the big happenings in the AI market, OpenAI becoming the platform of choice, apparently, the state of MCP and the Neon robot? I think Sam Altman's just further diluting any sort of genius level sway that he has by doing these constant events that are not cohesive. I really think what they need to do is like Willy Wonka style, like shut down the factory, go hide for a while, don't say anything to anyone, and then come out with like the golden tickets or some massive event. But I'm talking about this from a pure PR play. I think that for me, what would sway me is these guys leaning into being like huge platforms for like MCP hosting and coming out with the best tools out there to build on and staying far ahead. But as we discussed, I just don't know if that is possible. Yeah, I think that's it, right, is like go away, focus on real use cases, and instead of just talking about it, deliver this future, like deliver a really good platform that developers like us and others want to build on, that are excited about it and get benefit out of it. And then if they can deliver on that vision, then, you know, maybe that can be the future of the business. One way with Unmassive Data Center is imagine if GPT-5 was the fastest model. Like, imagine if it was just so fast, like it was just blazingly Grok level, Grok with a Q level, fast and reliable, and it was the best model, and they made it like a tenth of the price. Like, to me, then, why would you use anything else? Like, you would only use them, and that's how they could possibly win. And then because presumably that comes from economies of scale, It's just so hard for the others to compete. Like that is how they could win on the front that he's talking about. Yeah, I think that and also just like figure out these real use cases that businesses, you know, want to do. The ChatGPT thing is so confusing to me. Like they don't mention that core product. They're just like it's like one of our products on top of this huge stack. But it is the product. and the reality is like also maybe just go all in on that like i don't understand like it's seemingly this threat to google and it makes me question like why don't they go all in on it and i'm like it's probably because the cogs aren't there right you can't make any money off it like even at 20 bucks a month or 200 a month i would argue they're probably still losing money definitely are i mean we i mean their cost base would be different but still like those plans are too cheap to make any real money. Yeah, and also I think they've in some ways done a disservice by just devaluing the benefit of these models. Like, I think I would be willing to pay a lot more to just retain access. I don't want them to obviously do that, but I think if they need to make a profit, that's okay. Yeah. You know. All right. We will see you next week. Stick around if you want to hear the Greg Brockman sad song for not being included in the YouTube video. We'll see if it's a banger. Who knows? All right. We'll see you next week. Goodbye. And I'm left with the night. I built cathedrals out of code, out of time. We dreamed of truth that could reason, could rhyme. But somewhere between the chips in the sky, They found a new climb that left me behind. Is this the price of building the dawn? To watch it arrive with your name withdrawn. A billion lines, a single cut. A door I open quietly shut. And I am not in the frame tonight. Not in the questions, not in the light I see the future dressed in gold But I feel the weight of what we hold If all our engines roar and rise Who keeps the compass for our skies I love the spark, I fear the blaze Don't let the world be set ablaze You said platform, said infrastructure trust Silos of power, silicon dust A trillion promises, long steel lines But who will guard the hearts and minds We taught a mind to think like fire To stretch a thought to climb it higher But will it learn to care to stay When all we do is pave the way You measure progress in gay watts and graphs But I hear echoes in the aftermath The model's cheaper, the cost is down What's precious now, that thought is found And I am not in the frame tonight Not in the credits, not in the light I see the future split in twos The things we make, the things we lose If every workflow wins the war Who tends the soul at the core We'll cross the sea, we always do But who will anchor something true? There's a kind of silence only founders know The clock that ticks behind applause The hands that shake when you let go We said we'd make a lantern for the dark But lanterns cast a shadow too If we forget to hold the heart The brightest path won't carry through Teach it to reason, teach it to wait Teach it to doubt when high the stakes Teach it to listen more than let to know The lines we'd want and need Not just alignment on a slide But vows we keep when markets rise A promise etched in human hands To build what serves, not just expanse And I am not in the frame tonight Not in the crest, it's not in the light I see the future split in twos The things we make, the things we lose If every workflow wins the war Who tends the soul at the core Will cross the sea, we always do But who will anchor something true? I wanted wonder, not a crown To lift us up, not press us down If I'm not there to say the rest Let every word still mean our best And I am not in the frame tonight But I am holding to the light A patient future carefully grown Where wisdom weighs, what power sows If we must build an endless sky Let mercy be the reason why And if I'm gone when credits roll Remember, guard the human soul Turn down the glare, let conscience in Not just to race, but learn to win I'll take my bow behind the screen And pray we earn the world we dream.
Related Episodes

GPT-5.2 Can't Identify a Serial Killer & Was The Year of Agents A Lie? EP99.28-5.2
This Day in AI
1h 3m

ChatGPT is Dying? OpenAI Code Red, DeepSeek V3.2 Threat & Why Meta Fires Non-AI Workers | EP99.27
This Day in AI
1h 3m

Claude 4.5 Opus Shocks, The State of AI in 2025, Fara-7B & MCP-UI | EP99.26
This Day in AI
1h 45m

Is Gemini 3 Really the Best Model? & Fun with Nano Banana Pro - EP99.25-GEMINI
This Day in AI
1h 44m

Are We In An AI Bubble? In Defense of Sam Altman & AI in The Enterprise | EP99.24
This Day in AI
1h 5m

Do We Need AI Browsers? What Are Claude Skills? - EP99.22
This Day in AI
1h 26m
No comments yet
Be the first to comment