We might have taken a holiday break, but the world of AI didn’t! In this fast-paced rapid-fire episode, Paul and Mike are back to catch you up on everything you missed.
From OpenAI’s o3 model breaking human-level reasoning barriers to Sam Altman dropping cryptic hints about superintelligence, we’ve got all the updates. Plus, discover Google’s new “AI mode,” and why Microsoft is betting $80 billion on next-gen data centers.
Listen or watch below—and see below for show notes and the transcript.
This episode is brought to you by our AI Mastery Membership:
This 12-month membership gives you access to all the education, insights, and answers you need to master AI for your company and career. It includes access to briefings, a generative AI mastery series, Ask-Me-Anything sessions (AMAs), and more to help you stay competitive, marketable, and skilled in the age of AI.
To learn more about the membership, go to www.smarterx.ai/ai-mastery.
As a special thank you to our podcast audience, you can use the code POD150 to save $150 on a membership.
Listen Now
Watch the Video
Timestamps
00:04:37 — OpenAI Announces o3
- OpenAI Preps ‘o3’ Reasoning Model - The Information
- OpenAI X Status
- OpenAI o3 Breakthrough High Score on ARC-AGI-Pub - Arc Prize
- François Chollet X Status
- OpenAI's o3: The grand finale of AI in 2024 - Interconnects
- Early access for safety testing - OpenAI
- Deliberative alignment: reasoning enables safer language models - OpenAI
- Wojciech Zaremba X Status
00:11:19 — Superintelligence
- Google Employee Says ASI Looking More and More Probably by the Month
- Sam Altman Sparks Singularity Debate with Vague Post
- Altman’s The Intelligence Age Post
- Road to AGI MAICON Keynote
- Levels of AGI for Operationalizing Progress on the Path to AGI
- Stephen McAleer X Status
- Reflections - Sam Altman Blog
- Google Deep Research ASI vs AGI
00:23:15 — Policy Implications of AGI Post-o3
00:31:00 — OpenAI Structure
- Why OpenAI’s Structure Must Evolve To Advance Our Mission - OpenAI
- Microsoft and OpenAI Wrangle Over Terms of Their Blockbuster Partnership - The Information
- Microsoft and OpenAI have a financial definition of AGI: Report - TechCrunch
00:35:15 — Sam Altman on His Feud with Elon Musk
00:37:46 — GPT-5/Orion Is Behind Schedule and Crazy Expensive
00:42:13 — Google Introduces Gemini 2.0 Flash Thinking
00:46:04 — Google ‘AI Mode’ Option
00:49:23 — Google Publishes 321 Real-World Gen AI Use Cases
00:52:09 —Google 2025 AI Business Trends Report
00:54:55 — Satya Nadella on the BG2 Podcast
00:59:16 — Claude Pretends to Have Different Views During Training
01:05:29 —Facebook Planning to Flood Platform with AI-Powered Users
01:09:43 — DeepSeek V3
- Lmarena.AI X Status
- DeepSeek’s new AI model appears to be one of the best ‘open’ challengers yet - TechCrunch
- Why DeepSeek’s new AI model thinks it’s ChatGPT - TechCrunch
01:14:23 — Microsoft Outlines “The Golden Opportunity for American AI”
01:21:08 — Here’s What We Learned About LLMs in 2024
01:24:34 — Funding and Acquisitions
- xAI Raises $6B Series C
- Basis Raises $34M for AI Agents in Accounting
- Perplexity Closes Funding Round at $9B Valuation
- Perplexity Acquires Carbon
- Grammarly Acquires Coda
Read the Transcription
Disclaimer: This transcription was written by AI, thanks to Descript, and has not been edited for content.
[00:00:00] Paul Roetzer: I just don't understand why more people. aren't having a sense of urgency to solve for this. Like, why aren't we being more urgent in our pursuit of what future paths could look like? Welcome to the Artificial Intelligence Show, the podcast that helps your business grow smarter by making AI approachable and actionable.
[00:00:22] Paul Roetzer: My name is Paul Roetzer. I'm the founder and CEO of Marketing AI Institute, and I'm your host. Each week, I'm joined by my co host and Marketing AI Institute Chief Content Officer, Mike Kaput, as we break down all the AI news that matters and give you insights and perspectives that you can use to advance your company and your career.
[00:00:43] Paul Roetzer: Join us as we accelerate AI literacy for all.
[00:00:50] Paul Roetzer: Welcome to episode 129 of the Artificial Intelligence Show. This is our first episode of the new year. 2025 has arrived. Mike and I are back [00:01:00] after What felt like a month away, I don't know if that's good or bad, but let's be like we haven't done this in a while. It's actually only been probably about two weeks.
[00:01:08] Paul Roetzer: but we are back and we're going to do primarily a rapid fire only, although there's a couple of these topics that I'm not going to be able to help myself. We're going to have to talk a little bit more about a couple of them. AI did not take the two weeks off like we did. There was plenty going on over the holiday break, and into the new year.
[00:01:27] Paul Roetzer: So we've got a lot to cover. So Mike and I decided we're going to go with a bit of a more of a rapid fire style. Episode. so we won't use, you know, if you're new to the show, normally what the weekly looks like is we do three main topics where it's usually like five to seven minutes on each of those topics.
[00:01:43] Paul Roetzer: And then the rapid fire is usually like one to two minutes and we'll normally hit the three main and then like seven to 10 rapid fire. So this one, I don't even know what the count is, Mike. I think we've got about 15 rapid fire items we're going to power through here. but again, there was a lot going on at the end of the year, and so we [00:02:00] want to kind of recap for you what happened at the end of the year, and then lay the foundation.
[00:02:05] Paul Roetzer: For some of the big things we're watching as we head into 2025. So hopefully you had a great holiday season. Hopefully you had a wonderful new year. We appreciate you being back with us to kick the year off. This episode is brought to us by the AI Mastery Membership Program, SmarterX, and Marketing AI Institute, our two companies, sort of a team up effort to build this mastery program where we try and drive AI literacy for everybody.
[00:02:29] Paul Roetzer: And this is something we've been doing for a while. So there's, you know, the membership program includes a series of exclusive content, opportunities and experiences. We have a Gen AI Mastery series that Mike leads where he goes, you know, kind of does demos and deep dives into different AI applications and tools.
[00:02:46] Paul Roetzer: We have a quarterly Ask Me Anything session where I do an hour, just anything members want to talk about. we also do a quarterly trends briefing where Mike and I kind of pick out the top, 10 top things each quarter you need to know about. So those are, components [00:03:00] of it. There's also on demand webinars, content, exclusive content.
[00:03:03] Paul Roetzer: But the big thing moving forward, I'm not going to get into the plans yet. I alluded to this at the end of last year. I actually spent, I spent most of my holiday just hanging out with my family. The time I was. Doing a little bit of work was, pretty much almost exclusively focused on our plans for the AI Academy moving forward.
[00:03:21] Paul Roetzer: So, we have huge plans for a collection of new courses and series, new professional certifications. So we have our piloting AI and scaling AI, that is going to expand. and then a whole host of new experiences. So, this is again where I'm going to be spending a lot of my energy in 2025. I think AI literacy is going to become more critical than ever, as we move into this year and we look at re skilling and up skilling workforces.
[00:03:47] Paul Roetzer: And so that's, you know, where a lot of my energy and brain power is going to go to. So, you can learn more about that membership. There's a pod 150 promo code. So go to smarterx. ai.[00:04:00]
[00:04:02] Paul Roetzer: I think there's also just an education link, you can click on the navigation and go to membership. So you can learn all about that, so again, there's a 150 off promo code for our listeners, POD150, and that can get you in, and there's never been a better time to become a Mastery member with all the new things we're going to be announcing here in Q1.
[00:04:21] Paul Roetzer: So Okay, Mike, I don't know, 03 feels like it was like a quarter ago, but I guess that happened right before, you know, the holiday break. So I don't know. Let's kick things off and talk about the 03 model from OpenAI.
[00:04:37] OpenAI Announces o3
[00:04:37] Mike Kaput: All right. Sounds good, Paul. So at the end of 2024, OpenAI announced. This groundbreaking new model called o3.
[00:04:46] Mike Kaput: So o3 is actually the follow up model to O1. So they skipped O2, I believe, due to some copyright conflicts with some other company or event or product. So, you know, as if these names weren't confusing [00:05:00] enough, o3 is the sequel to O1. But, as we've covered, O1 is the company's advanced reasoning model. o3 builds on that by taking time to actually think and reason through problems.
[00:05:11] Mike Kaput: And it is not publicly available yet. But it's getting a ton of attention because it is the first AI model allegedly to outperform humans on this specialized intelligence test called ARC AGI. So this is actually created by a very prominent researcher in AI named François Chollet. And this test basically presents simple visual puzzles where you have to figure out patterns and rules.
[00:05:39] Mike Kaput: But what makes it special is that this doesn't rely on you memorizing knowledge. It basically tests your ability to learn and adapt to completely new situations. So, humans typically score about 75 percent correct on this test on average, and the whole point of this is to test, can AI do these types of [00:06:00] things better than humans?
[00:06:01] Mike Kaput: Now, o3 actually achieved 76 percent accuracy, marking the first time an AI system has actually surpassed human performance on this benchmark. Now, what's actually really noteworthy here is not just that number, but the fact that previous AI models like GPT 4 basically scored near zero on these same types of tests.
[00:06:24] Mike Kaput: So again, kind of that more general adaptive intelligence. That humans excel at, o3 apparently just narrowly edged out human beings. And, you know, in fact, AI The best
[00:06:37] Paul Roetzer: human beings. The best human beings. Like these are the top of the top.
[00:06:39] Mike Kaput: Right. So, Sholeh, the guy who made this, he's been pretty historically skeptical of AI hype and he even admitted he called it a surprising and important step function increase in AI capabilities.
[00:06:52] Mike Kaput: and a genuine breakthrough. He suggests that o3 is doing something fundamentally different from previous AI models, [00:07:00] and he is also though quick to point out this does not necessarily mean we have achieved AGI. He still says the model fails at some puzzles that humans find quite easy, and so he's also developing a new harder version of the test that he predicts will dramatically reduce o3's performance on it.
[00:07:18] Mike Kaput: So Paul, nobody really has access to o3 yet. But it's performance on this test and on a bunch of other benchmarks certainly seems like a huge deal. Is this as big a deal as people are making it out to be?
[00:07:31] Paul Roetzer: It's definitely, appears to be a bit of a leap forward in, you know, the reasoning capability within these models.
[00:07:38] Paul Roetzer: The evaluations that are used to test these are notoriously not necessarily representative of the impact on the economy and the workforce, they're trying to come up with like extremely complicated things that only the elite minds in the world can solve. And then they're trying to figure out what does that mean to the [00:08:00] broader economy and the impact that these models are going to have.
[00:08:02] Paul Roetzer: So, yeah, I mean, I think it's significant. I think it demonstrates they need to continue to work on these evaluations to try and set these benchmarks for how they'll define AGI and beyond, which we'll be talking a lot more about in this episode, what's beyond AGI. I think the thing that people need to Maybe come back to and focus on is these evaluations are nice to talk about.
[00:08:26] Paul Roetzer: It's how the AI labs look at the frontier and like the true advanced capabilities of their models. But the thing that actually matters to all of us is, is it super human at our job, at the tasks that we do every day? And the thing, and I know Mike, you've shared some stuff recently on LinkedIn and on X about some of the ways you're using like Google deep research and things like that, and we'll talk more about that stuff as we go along.
[00:08:48] Paul Roetzer: But the reality is, if you take your job as a collection of tasks, which is what we always, you know, talk about, that's what we all do, it's all a collection of tasks, and you have a list of those 25 to 30 things, my [00:09:00] guess is as 2025 progresses, AI models are going to become superhuman at an increasing number of those tasks.
[00:09:09] Paul Roetzer: They're going to do things that you do every day better than you do them, better than the best people in your profession do them. And so the true evaluations as we look to the future impact of AI, and when I say future I mean like 12 to 24 months, is, Are they becoming superhuman at the tasks that make up the jobs that make up the economy?
[00:09:31] Paul Roetzer: And the answer is, they absolutely are going to. And those aren't going to be found in evals from OpenAI and Google. They're not going to go into your sector, your industry per se. And say, okay, in retail or in, you know, consumer goods or, or in healthcare, is it superhuman at the things that these people do?
[00:09:50] Paul Roetzer: It's not, they're not going to do that kind of research. That's going to be on all of us to figure out. But I think what we'll see more and more of is the people like you and I, Mike, who are using these tools every [00:10:00] day to increasingly become more efficient, productive at the things we do. We're going to more and more realize they're just either better at it, or they're equal, but way faster at it.
[00:10:10] Paul Roetzer: And that's when it starts to like, that's where I think we start to see, I saw it, kind of a separate related note here is, I saw a tweet from Sam Altman, I think it was this morning, where he said that they set the price of the pro for the O1 model at 200 a month. And he personally picked that price point because he thought it would be profitable for them because he didn't think usage of O1 would be dramatic.
[00:10:32] Paul Roetzer: And they're actually losing money at the 200 a month number because people are using it so much, which means they're finding immense value in these reasoning models. And these are very early versions of it. So I think once you get to o3, you'll see. And you get to unlock these reasoning capabilities and you get more education around how to use them in your job.
[00:10:54] Paul Roetzer: I think you're going to start seeing a lot of those tasks where these models are going to be able to do it better than [00:11:00] you. And you'll find other stuff to do, like Mike, you and I are lacking for things to do. Like if o3 becomes better at strategic planning in some cases than me, like, cool, I'll move on and do the other stuff.
[00:11:10] Paul Roetzer: Right. But yeah, I think that the implications of these advancements are going to become more and more real. As this year progresses.
[00:11:19] Superintelligence
[00:11:19] Mike Kaput: So the next topic is really intimately related to this because on the heels of these o3 breakthroughs, the AI community is, has just been on fire the last week or two with talk of this concept of Artificial Super Intelligence, ASI.
[00:11:36] Mike Kaput: So this is, A hypothetical form of AI that, unlike AGI, which can do many things better than humans, this would surpass human intelligence and capabilities in every single field imaginable. So, kind of a bit of a sci fi concept here, but something that people appear to be taking seriously. So, first we got some commentary on X from Logan Kilpatrick, a [00:12:00] prominent AI voice product lead at Google AI Studio.
[00:12:04] Mike Kaput: Who said, quote, straight shot to ASI is looking more and more probable by the month. This is what Ilya saw, referring to former OpenAI chief scientist Ilya Sutskever. Kilpatrick then went on to refer to the efforts of Sutskever's new company, Safe Superintelligence, SSIn their efforts to kind of go straight to building superintelligence, rather than stepping stone products along the way.
[00:12:29] Mike Kaput: This was then followed, in true, you know, grandstanding fashion, by a cryptic post on X from Sam Altman himself saying, quote, I always wanted to write a six word story. Here it is, near the singularity, unclear which side, alluding to the fact that we are on the path to some type of grand superintelligence.
[00:12:52] Mike Kaput: So, Paul, I guess, like, Enough, like, credible people are talking about this, seemingly [00:13:00] believing we're heading towards possibly some path to ASI. Like, what is going on here? Like, how serious should we be taking this?
[00:13:08] Paul Roetzer: When I was, like, getting ready for the episode, I had to keep reminding myself that this was a rapid fire topic.
[00:13:14] Paul Roetzer: I was trying, like, real hard not to go too deep here. But I think this is one of the ones where like some perspective is very, very important. So I remember, you know, back in like 2023, which kind of sounds like I'm talking about like a decade ago, but you know, shortly after ChatGPT came out and even before it, I avoided talking about AGI on LinkedIn.
[00:13:33] Paul Roetzer: Like I post on LinkedIn, you know, four or five times a week. And I was very conscious of not talking about it. And even on the podcast, I was hesitant to bring the topic in because I didn't feel like the world at large was ready for the conversation about AGI. And, and now here we are like 12, 18 months later and we're moving on to superintelligence.
[00:13:53] Paul Roetzer: So I think just to give a little perspective here, this isn't a new concept. I don't know who wrote the [00:14:00] book, superintelligence, isn't, was that, there's a book on this. but if we go back just to September, 2023, The DeepMind team wrote a report called Levels of AGI for operationalizing progress on the path to AGI.
[00:14:16] Paul Roetzer: So this was released in May of 2024, led by Shane Legge, one of the DeepMind co founders, and who's also credited with coining AGI, the phrase, around 2002. So, At level five of theirs, and we'll put the link to this in, it's superhuman. So they, they rate, the levels of AI based on performance. So how it performs versus humans.
[00:14:40] Paul Roetzer: And then generality, how, how many cognitive tasks it can do at these different levels. So level 5 in their world is the highest level, and that is ASI, Artificial Superintelligence. Now, they define it as a system that would be able to do a wide range of tasks at a level no human can [00:15:00] match. So we define superhuman performance as outperforming 100 percent of humans.
[00:15:05] Paul Roetzer: So level 4 is virtuoso, meaning like they're at the 99th percentile, so pick any task, any job, And the AI at that level would be at the highest percentile of human capability. At superintelligence, we are now beyond any human, any scientist, any, developer, any marketer, any entrepreneur, any CEO. It is beyond their capabilities.
[00:15:28] Paul Roetzer: We, we can't, they can't match what the AI can do. So, that was May of last year. Then in June of last year, episode 102, we talked about this, Mike. Leopold Aschenbrenner and his, what was the name of that series?
[00:15:40] Mike Kaput: Situational Awareness.
[00:15:42] Paul Roetzer: Yeah, Situational Awareness, a collection of papers. So in that, he said, we will have superintelligence in the true sense of the word by the end of the decade, and AGI by 2027 is strikingly plausible.
[00:15:55] Paul Roetzer: It seems like almost everybody in AI, outside of Gary Marcus and Yann LeCun, it [00:16:00] thinks AGI is definitely coming by 2027. I think most are starting to center more around 2025. And then there's, one of the papers in there was from AGI to superintelligence, the intelligence explosion. And he said, AI progress won't stop at human level.
[00:16:16] Paul Roetzer: Hundreds of millions of AGIs could automate AI research, which Mike, we're starting to see in the early forms of Google Deep Research. Compressing a decade of algorithmic progress into one year, we would rapidly go from human level to vastly superhuman AI systems. The power and the peril of superintelligence would be dramatic.
[00:16:36] Paul Roetzer: A week later, we had the launch of Ilya's Safe Superintelligence Company, where they say on their page, you can go to this right now, I'll put the link there, their, their website is a single page. It starts off, super intelligence is within reach. Building safe super intelligence is the most important technical problem of our time.
[00:16:55] Paul Roetzer: We have started the world's first straight shot, super intelligence lab with [00:17:00] one goal and one product, a safe super intelligence. Then in September, September, 2020 or 23rd, 2024, we had Sam Altman's post that we talked about Mike, the intelligence age. Where he said, Here is one narrow way to look at human history.
[00:17:15] Paul Roetzer: After thousands of years of a compounding scientific discovery and technological progress, we have figured out how to melt sand, add some impurities, arrange it with astonishing precision at extraordinarily tiny scale into computer chips, run energy through it, and end up with systems capable of creating increasingly capable artificial intelligence.
[00:17:38] Paul Roetzer: This may turn out to be the most consequential fact of all. All of history so far. It is possible that we have superintelligence in a few thousand days. It may take longer, but I'm confident we will get there. Then we had a tweet from Stephen McClure, and we'll put this link in, who is a researching agent, researches agent safety at OpenAI.
[00:17:59] Paul Roetzer: He [00:18:00] tweeted on January 3rd, I kind of miss doing AI research back when we didn't know how to create super intelligence. Again, this is someone within OpenAI. Then, on January 5th, this is just yesterday, we're recording this on January 6th, Sam publishes Reflections after he did an interview with Bloomberg that actually got him to write this post.
[00:18:20] Paul Roetzer: Because it was funny, in the Bloomberg interview, he's like, I don't have time to write these kind of posts anymore. And then he did it anyway. So I'll read a couple quick excerpts. We are not, this is again, directly from Sam. We are now confident we know how to build AGI as we have traditionally understood it.
[00:18:34] Paul Roetzer: We believe that, in 2025, we may see the first AI agents, quote, join the workforce and materially change the output of companies. We continue to believe that iteratively putting great tools in the hands of people leads to great, broadly distributed outcomes. We are beginning to turn our aim beyond that to superintelligence in the true sense of the word.
[00:18:56] Paul Roetzer: That's like a, it's like a Silicon Valley thing. People like to say in the true sense of the word. [00:19:00] we love our current products, but we are here for the glorious future. With superintelligence, we can do anything else. Superintelligent tools could massively accelerate scientific discovery and innovation well beyond what we are capable of doing on our own.
[00:19:15] Paul Roetzer: And, in turn, massively increasing abundance and prosperity. This sounds like science fiction right now and somewhat crazy to even talk about. That's all right. We've been there before and we're okay with being there again. We're pretty confident that in the next few years, everyone will see what we see and that the need to act with great care while maximizing broad benefit and empowerment is so important.
[00:19:38] Paul Roetzer: Given the possibilities of our work, OpenAI cannot be a normal company. Now, in that article, he talks about the origins of, of, OpenAI and how nobody believed in AGI back in 2014 15 when they were doing it. So, the final thing I'll call here is the Road to AGI keynote that I did at MAICON. We'll put the link to there if you have not watched that.
[00:19:58] Paul Roetzer: It tells this story [00:20:00] of the path to AGI and then what comes next. So, it's available for free on YouTube, you can check that out. And then, again, I don't want to, like, scoop ourselves, but we will be launching a new podcast series in Q1 of this year called The Road to AGI and Beyond, where we're going to be doing interviews with leading minds in AI, working on the different components of AGI and artificial superintelligence.
[00:20:25] Paul Roetzer: So, stay tuned. More to come on that series. I expect it to probably launch in February, so stay tuned for that. But this is a critical topic for everyone to understand and follow. And then Mike, if you want to give a quick, heads up, I know you had shared on LinkedIn, you did a Google deep research on like ASI versus AGI.
[00:20:44] Paul Roetzer: I found that doc pretty helpful.
[00:20:46] Mike Kaput: Yeah. Yeah. So Google deep research for anyone that doesn't know is a feature of Google Gemini 1. 5 Pro. Basically you enable it and. It will create for you a really in depth research brief on a bunch of different topics. [00:21:00] So, I, before this podcast ran a research brief that's literally created one that's like 2, 000 words in a couple minutes.
[00:21:08] Mike Kaput: It scanned, I think, 60 plus different web sources. And I wanted to define ASI versus AGI. So, there's a lot to this brief and, you know, I can literally, we can link to the public Google Doc in the show notes so people can read it. But really, it defines ASI as a hypothetical form of AI that surpasses human intelligence in every aspect.
[00:21:29] Mike Kaput: So, compared to AGI, AGI aims to create AI systems with cognitive abilities comparable to humans across various domains. So, like, stuff that's as good as us across a broad base of things. Superintelligence is essentially alien intelligence and how good it is. So, there's a huge order of magnitude change between those two.
[00:21:51] Mike Kaput: And, Paul, I would note that that superintelligence book you mentioned is from Nick Bostrom. There we go. he published it way back in 2014. He's [00:22:00] actually a philosopher. He's been working on this for a long time. And it's really interesting. He was one of the first to be talking about it. Beyond, I would say, the real OG is Ray Kurzweil, who kind of coined that term singularity, or made it popular, rather.
[00:22:14] Mike Kaput: That point where machine intelligence surpasses
[00:22:17] Paul Roetzer: humans. And those are some of the most influential books. So back in like 2011, when I was researching this stuff and not talking about it because no one in business, it was, again, it was like a taboo topic, like nobody thought it was real. I read all those books and I feel like I might need to go dust those off because there was, there was probably like five or six that I read that were just like very influential in me thinking that the future was going to look very different than the present and I was like starting to position my own company and myself to sort of, you know, You know, be at the forefront of that and just start trying to figure out the story of AI, which is what led to the creation of Marketing Eye Institute and eventually SmarterX.
[00:22:56] Paul Roetzer: ai. so yeah, it's, I [00:23:00] don't know, man, it's, it's, it's a lot. And the funny thing is, like, I started, like, a lot of this stuff, I was thinking about, like, Sunday night, and I'm like, man, this is just diving right back in. After two weeks of my mind kind of, like, taking a break, I just jump right back in.
[00:23:12] Mike Kaput: Coming in really hot in the new year.
[00:23:14] Paul Roetzer: Definitely.
[00:23:15] Policy Implications of AGI Post-o3
[00:23:15] Mike Kaput: So related to these first two topics, the third topic we're kind of walking through today is among all this kind of debate and commentary about AGI, ASI in the wake of o3, there was a post from an OpenAI employee that actually jumped out to us as really particularly interesting on what we can expect because of this technology.
[00:23:37] Mike Kaput: So this came from Yo Shavit, who is a Frontier AI Safety Policy Lead. at OpenAI, and he posted on X, quote, Now that everyone knows about o3, and imminent AGI is considered plausible, I'd like to walk through some of the AI policy implications I see. And so he outlines this pretty fleshed out argument, stating very briefly the [00:24:00] following observations.
[00:24:01] Mike Kaput: First, everyone will probably have access to some version of ASIf we reach it. Second, if we do, the corporate tax rate is about to become very, very important because the economy will essentially be dominated by AI agents that do labor, that is owned by companies. Third, in this scenario, in his opinion, AI should not be able to own assets because they might, if they could, be able to.
[00:24:26] Mike Kaput: Fully wrest control of the economy and society from humans. Four, laws around compute and who controls it literally become critical as a way to regulate rogue AI elements and agents. There's no way, as he says, to put an AI agent, quote, in jail. And number five, as Chevy writes it, quote, technical alignment of AGI is the ballgame.
[00:24:49] Mike Kaput: With it, AI agents will pursue our goals. And look out for our interests even as more and more of the economy begins to operate outside direct human oversight. So [00:25:00] Paul, the third topic today that basically sounds like science fiction, but based on who's saying it sounds like we also need to take it seriously.
[00:25:10] Paul Roetzer: So again, each one of these I could just do a whole episode on. so I'm actually going to follow this one up with another, thread from a head of mission alignment at OpenAI, Joshua Akyem, and we'll put the link in here. I'm just going to read his thread real quick because I think it sort of encapsulates the moment a little better than I maybe could.
[00:25:33] Paul Roetzer: and I think it's notable how many different people at OpenAI are very publicly starting to share this stuff. I don't think that's by accident and I don't see it as a PR ploy in any way. I think they have all seen something we have not had access to yet. And they are very confident, they truly believe what, what they're saying [00:26:00] publicly now, which means AGI in their mind is very near, and, and they seriously are considering the implications beyond that.
[00:26:08] Paul Roetzer: So, again, this is gonna be from Joshua Akiam, Head of Mission Alignment, OpenAI. The world isn't grappling enough with the seriousness of AI and how it will upend or negate a lot of the assumptions many seemingly robust equilibriae are based upon. Domestic politics. International politics. Market efficiency.
[00:26:27] Paul Roetzer: The rate of change of technological process, social graphs, the emotional dependency of people on other people, how we live, how healthy we are, our ability to use technology to change our bodies and minds, every single facet of the human experience is going to be impacted. It is extremely strange to me that more people are not aware or interested or even fully believe in the kind of changes that are likely to begin this decade and continue well through the century.
[00:26:57] Paul Roetzer: It will not be an easy century. It will be a [00:27:00] turbulent one. If we get it right, the joy, fulfillment, and prosperity will be unimaginable. We might fail to get it right if we don't approach the challenge head on. It feels outside of the Overton window right now to suggest that so much change could happen very quickly, or even to realistically grapple with what those changes might entail.
[00:27:20] Paul Roetzer: It is too easy to say the present is more urgent and more real. Quick side note, Overton Window refers to the range of ideas or policies that are considered acceptable or mainstream in public discourse at a given time. He continues, Nonetheless, change is coming. It will be reflected first in the prices of goods and labor.
[00:27:41] Paul Roetzer: It will force changes in strategy in businesses, institutions of all kinds, and countries. Then, it will force changes in philosophy. What are we here for? Why do we do the things we do? If everything we care about is automatable, what is our role in the world? We'll have to tell a new [00:28:00] story for a new age.
[00:28:01] Paul Roetzer: Some of the timeless stories, how we are driven by curiosity, by love, by the human spirit, will remain unchanged. But everything else will be replaced by questions that need new answers. I don't know what the future will bring, but my enduring belief is that humanity is beautiful, flaws and all, And the future we build should somehow cherish the human heart.
[00:28:24] Paul Roetzer: That's beautiful. Like, I, I think that it's, it's how I've felt for like the last decade, that I just don't understand why more people aren't, like, having a sense of urgency to solve for this. It's how I feel, I've mentioned many times on the show, it's how I feel about economists. Like, why aren't we being more urgent in our pursuit of what future paths could look like?
[00:28:48] Paul Roetzer: And so I just, I think it. It stresses the importance of policy, of governance, of laws and regulations, of thinking through the impacts on different sectors. [00:29:00] And it's just absolutely critical. And again, like it's, you know, I've, I see it increasingly as part of our role on this show is to try and call attention to these topics and to hopefully inspire people across different sectors to kind of pull their thread on the AI world.
[00:29:17] Paul Roetzer: topic and, and really start pursuing answers and, and, and, and I guess even start asking the hard questions about their industry, their, their career, their profession, their community. And so, you know, again, a lot more this year to come on, on these topics because I think it's where the conversation is heading.
[00:29:37] Mike Kaput: Yeah, and I would just add to that based on some of the conversations we've had with our audience, comments I've seen online. Like, It's understandable, but it seems like a lot of people are waiting for permission to go engage with this technology, or need some type of, obviously everyone needs guidance, but I don't know.
[00:29:55] Mike Kaput: I see so many people, they're like, well, can I do this? Can I do that? You can go [00:30:00] explore this technology on your own. And nobody has all the answers or the guidebooks. So it's really imperative as many people as possible, as many perspectives as possible, get involved.
[00:30:10] Paul Roetzer: Yep. Yeah. And it's going to be a bit taboo, honestly.
[00:30:12] Paul Roetzer: Like if you start talking about super intelligence to like your friends or even your co workers, they might think you're nuts. But I don't know. Like we, we, we have to push the conversation forward. Like we need people to be paying attention and to be starting to take action. I I've said it before. I'll say it again.
[00:30:28] Paul Roetzer: Like, I think we still have time. Like, I think we have time to solve for this. I think we have time to affect a positive outcome in, in our businesses and our industries and our careers and, you know, across society, but the time is moving faster, like we have to take action this year and we have to start asking the hard questions and pursuing.
[00:30:50] Paul Roetzer: We're not going to have the answers, but at least pursuing different paths of possible outcomes so that we're ready, depending on how quickly this technology moves throughout society.
[00:31:00] OpenAI Structure
[00:31:00] Mike Kaput: All right. While all these kind of big picture conversations are happening, there's a bunch of stuff also happening kind of in the weeds with these companies, including OpenAI is in the process of trying to move to a for profit structure.
[00:31:14] Mike Kaput: Roetzer, they actually just published an article justifying why they have to do this. They revealed that they plan to convert their current structure into a public benefit corporation with ordinary shares of stock. At the same time, the company says that it plans for its non profit wing to exist alongside the for profit entity.
[00:31:33] Mike Kaput: And own shares in it. So according to OpenAI, this quote would result in one of the best resourced nonprofits in history. This would, of course, replace the unusual nonprofit slash for profit hybrid structure they've had for a while. But, there's a big hurdle that they're trying to figure out with these plans, which is Microsoft.
[00:31:54] Mike Kaput: Because Microsoft has invested over 13 billion in OpenAI, and they need [00:32:00] to figure out the terms. of this deal and what it means for them. So, OpenAI and Microsoft have been negotiating over the transition to the for profit entity for months, according to the information. And these negotiations center around four key areas, apparently.
[00:32:16] Mike Kaput: So, first, Microsoft's equity stake in the new entity. Second, whether Microsoft will remain OpenAI's exclusive cloud provider. Third, the duration of Microsoft's rights to use OpenAI technology. And fourth, whether Microsoft will continue to receive 20 percent of OpenAI's revenue. So Paul, like, how likely is it that OpenAI pulls this off anytime soon?
[00:32:39] Paul Roetzer: It's tricky. I mean, Elon Musk is suing to, you know, prevent this from happening. I mean, it seems inevitable that they will solve how to do it, but to move from a nonprofit to a for profit or to keep the nonprofit, but have it like have less control is way more complex than you and I are probably gonna be able to unpack on this episode or even on this show [00:33:00] is not our area of expertise.
[00:33:01] Paul Roetzer: And it seems unprecedented in history to have a company this big, growing this fast, try and do something like this. one of the roadblocks is, we've talked about on the show, you know, in Q4 last year, I know we mentioned it, but, in the contract with OpenAI and Microsoft, apparently, at least originally, OpenAI's board, the nonprofit board, got to define when AGI had been achieved.
[00:33:29] Paul Roetzer: And if it had been achieved, Microsoft's rights to the technology were no more. And so, it was believed that one of the big obstacles was the OpenAI board could just say, Well, we've got AGI and you don't get the technology anymore. But, it came out in December, actually right after Christmas. that, Microsoft and OpenAI had actually refined the definition of AGI.
[00:33:52] Paul Roetzer: And so according to TechCrunch, we'll put this link in the show notes, the two companies reportedly signed an agreement last year, so then [00:34:00] 2023, stating OpenAI has only achieved AGI when it develops AI systems that can generate at least 100 billion in profits. That's far from the rigorous technical and philosophical definition of AGI many expect.
[00:34:15] Paul Roetzer: So I am sure there's way more detail than just 100 billion profits because I've got like 20 questions myself related to like, how would you know? Like, what would that be? but it seems like they actually have a more solid definition than just the OpenAI board deciding that AGI has now been achieved.
[00:34:32] Paul Roetzer: So yeah, this is going to be a fascinating thing throughout the year, but I know in Sam's, one of his posts, he said like they, They're, they need to raise more money now, way more than they ever expected. But I think to do that, they need to figure out the structure first. And I think that there is a sense of urgency to do both.
[00:34:52] Paul Roetzer: And I, they're starting to publicly share a lot of these details, probably as a prelude to, you know, some resolution. Now my guess is they're going to [00:35:00] have a resolution and then there's going to be even more lawsuits. Like, I think this is going to play out in courts for the next decade because Elon Musk is probably going to either want his share or something, I don't know, like it's, it is a soap opera.
[00:35:15] Sam Altman on His Feud with Elon Musk
[00:35:15] Mike Kaput: Next up, we just got a new interview. Yeah, no kidding. You're about to hear a little bit more about that in this new interview with Sam Altman that just came out courtesy of the Honestly podcast from The Free Press. In this, Altman talks pretty candidly about his feud with Elon Musk and the battle to control the future of AI technology.
[00:35:37] Mike Kaput: Now, some of the juiciest tidbits in this were actually about the battle between OpenAI and Musk. According to Altman, The core tension here isn't actually about OpenAI's shift from a non profit to a hybrid structure as Musk has tried to portray. Rather, he claims it was Musk who initially pushed Hardest for OpenAI to become a for profit entity, [00:36:00] proposing at one point it become part of Tesla.
[00:36:02] Mike Kaput: He basically suggests the current conflict stems more from competitive dynamics, with Musk obviously now running XAI, which is a direct competitor. At one point, he basically just came out and said that Musk is clearly a bully and wants control of the world's AI to belong to him. Now, Altman also maintained that critics are mischaracterizing OpenAI's changes to its corporate structure.
[00:36:26] Mike Kaput: He emphasizes that the non profit isn't becoming a for profit, but would continue to exist alongside the for profit entity. So, Paul, this seems like some of the more candid comments. Sam has made so far about Elon. Like, what did you take away from this interview?
[00:36:41] Paul Roetzer: He's, he's been more vocal in the last few weeks about this.
[00:36:44] Paul Roetzer: He's echoed these similar sentiments in a number of interviews and articles. So, I don't know. I just feel like he's just kind of getting, like, fed up and frustrated with the whole thing and, you know, just wants fair competition, and I think he [00:37:00] just feels like Like I said, so I, I agree, like, I honestly think Elon's just messing with him.
[00:37:05] Paul Roetzer: He's got the money to do it, and he's gonna muddy some stuff up, and if it buys him a little time to get Groq 3 to market, or Groq 4 to market, and, and, you know, take a lead, Elon's gonna use whatever advantage he has. I think his history has shown that, so, yeah, I don't know, I mean, I don't, it's gonna keep going on, the feud's gonna continue, I don't, I don't see some peace being brokered between those two anytime soon.
[00:37:31] Paul Roetzer: But, yeah, I think Elon's just messing with them, honestly, and it may eventually not go anywhere, but he's gonna keep, keep creating friction, cause he's good at it, and he has a competitive reason to do it.
[00:37:46] GPT-5/Orion Is Behind Schedule and Crazy Expensive
[00:37:46] Mike Kaput: So next up, despite OpenAI's major breakthrough with o3, it is also reportedly facing some hurdles with its next model, which is codenamed Orion.
[00:37:57] Mike Kaput: Basically, this is intended to be [00:38:00] GPT 5, and this comes from some new reporting from the Wall Street Journal. The project, which has been in development for over 18 months, is behind schedule and running up enormous costs, according to the reporting. Each training run for Orion can cost about half a billion dollars in computing costs alone.
[00:38:19] Mike Kaput: And so far, these runs have fallen far short of researchers hopes. OpenAI has conducted at least two large training runs, each of which takes months to complete. And encountered new problems each time. So a key challenge that sounds like it is coming up here is data quality and quantity. The public internet simply doesn't have enough high quality data to train a model of Orion's intended scale.
[00:38:44] Mike Kaput: So to address this, OpenAI has begun creating data from scratch, even hiring experts to write software code and solve math problems while explaining their thought processes. So Paul, one thing that kind of jumped out to me in this report is [00:39:00] The journal is specifically calling out the progress that the company has made on o3, despite all these struggles with Orion, because it takes a different approach to scaling up capabilities.
[00:39:10] Mike Kaput: So we have this kind of, these threads coming together where scaling how reasoning works may be a path forward. Also, if you crack reasoning, you may also have basically a synthetic AI researcher that can help you generate quality data. How are you looking at reports of these model slowdowns?
[00:39:30] Paul Roetzer: So, I mean, first, like, the way that this article is explaining it, you know, the last thing you said, to address this, OpenAI has begun creating data from scratch, hiring experts to write software code and solve math problems while explaining their thought processes.
[00:39:42] Paul Roetzer: That sure sounds a lot like 03, like, I mean, the whole point of the reasoning models is to be able to go through this chain of thought, which requires human experts to tell them how they think about things, how they solve problems. So it's like reinforcement learning, but like, how did we get to these answers?
[00:39:57] Paul Roetzer: How did we solve this math problem? I'm, I'm [00:40:00] not so sure that at some point here, and maybe it's with o3, they, they combine these models. I've never really comprehended why they would go down the path of having a GPT 5 or Orion model and an o3 model. Maybe again, like there's all this, there's this challenge of like, whether it's Google or Anthropic or, OpenAI, solving for developers versus solving for enterprises.
[00:40:27] Paul Roetzer: Solving for developers with 17 different models at all these different costs, I get that maybe that makes sense. Enterprise users don't want seven models to choose from. We have no idea the difference between these models, and should I use the flash or the full, like, that is not an enterprise user, solution set.
[00:40:49] Paul Roetzer: And so I think at some point there, They're very well aware, obviously, that their revenue is, is going to come from enterprises as they kind of scale this up. They have to solve for those enterprise users, not [00:41:00] just for the developers. And so I think at some point you just have to consolidate these models and like simplify this for people so it's not so damn confusing and again, you and I follow this all the time and I get confused by what models are.
[00:41:12] Paul Roetzer: The second thing I'd note is I have now seen reports that Gemini 2 is delayed, that Claude Opus is delayed, that Groq 3 was delayed. All those companies dispute these reports from media and from Twitter. who knows what's actually going on. It does seem like there's some sort of trend that these training runs sometimes just don't work.
[00:41:33] Paul Roetzer: And so we don't get these models at the timelines we expect them. but this, this isn't traditional software. It doesn't just do what it's supposed to do. You have to train it, you have to go through these big runs, and then you have to go through all this reinforcement learning. You have to go through all this testing.
[00:41:49] Paul Roetzer: And sometimes it just doesn't work. doesn't come out fully baked. It doesn't come out like you expected. and, and so I think we're just, this is going to be a continuing thing. But yeah, I mean, OpenAI is making [00:42:00] progress on other paths, as is Google, as are others. Like everyone's pursuing reasoning and test time compute now, not just throw more data, more computing power, and build a bigger, bigger model.
[00:42:10] Paul Roetzer: They're going to keep pursuing both paths.
[00:42:13] Google Introduces Gemini 2.0 Flash Thinking
[00:42:13] Mike Kaput: So related to that, Google has actually just announced Gemini 2. 0 Flash Thinking, which is a new experimental model that appears to be their response to o3. this key innovation in this model appears to be making the model's thought processes visible while maintaining high speed performance.
[00:42:33] Mike Kaput: The model can show its thoughts while solving complex problems that require both visual and textual understanding. So this is really helpful not only for using advanced reasoning, but also for tasks where users need to understand how did the AI actually get to its conclusions. So to that precise point you just made, Paul, It sounds like reasoning, thinking, you know, chain of thought, taking your time [00:43:00] to think through a problem.
[00:43:01] Mike Kaput: This is what all the major labs are focusing on right now.
[00:43:04] Paul Roetzer: Yeah, and again, like this, I don't know if this is the marketer or branding person in me, but like, I gotta see 2025 is the year these labs figure this out. Nobody cares. Like, I just want to go in, I want to use GeminI want to know it's the most advanced version.
[00:43:20] Paul Roetzer: I want Google to figure out, based on my prompt, what I'm, what I'm using it for. So, like, if I go in and say, hey, I need you to conduct research on these, you know, this topic around ASI versus AGI. Let the model figure out that the deep research version with 2.0 flash thinking is the right model to solve it for.
[00:43:40] Paul Roetzer: How am I supposed to know that that's the right model? So again, they're all doing this exact same thing. You go into any of these platforms, whether it's Chad cpt or Gemini or Claude, and you got all these choices of models, I don't need that. And there's no way that they can't run an algorithm to figure out the best model for someone to use.
[00:43:59] Paul Roetzer: [00:44:00] So, I don't know, like, again, flash thinking cool, but I, I like, like, NotebookLM, I like deep research, I like that, like, Google's building these distinct sort of, like, products, but I just feel like, at the end of the day, they're trying to build Gemini as your intelligent assistant. And an intelligent assistant shouldn't have 17 options of which version of the assistant I get.
[00:44:24] Paul Roetzer: And so I just, I hope that as these companies all solve for enterprises, in 2025, they, they fix this cluster of a naming convention and like, The choices that they're forcing uneducated people to use to decide these things. It's crazy.
[00:44:44] Mike Kaput: In a small way, you're starting to see the value of that with something like, in ChatGPT, you can say, Hey, write me a landing page.
[00:44:52] Mike Kaput: Great. It writes it. Drop in some data on past landing pages. Hey, analyze this for me. It does it. You no longer have to select like code [00:45:00] interpreter. Hey, by the way, I need a picture for the landing page. Make this image. You don't have to go select DALL E anymore. Just do it.
[00:45:06] Paul Roetzer: Right. Like we need, so we have multi modal.
[00:45:09] Paul Roetzer: Right. We need like multi modal, like within a single interface. And so it just like, it does its thing. I, I agree. And I could see it where it's like, Hey, Google Deep Research was used for this, or we used 1. 5 advanced. It's great. Like, I don't, I don't really care, but like, cool. Maybe some people would care which model was used to do it.
[00:45:28] Paul Roetzer: But as long as they, again, like this can't be the hardest of all the things they're solving for, building an algorithm that determines the best model right up front can not possibly be that difficult. I'm not a developer, but like that seemed, I know that like Jasper's doing that with their models. Like they're kind of picked the best one.
[00:45:48] Paul Roetzer: Perplexity is similar. It's like, why, why is Perplexity making me choose models? Like just you pick the best model.
[00:45:54] Mike Kaput: Yeah.
[00:45:54] Paul Roetzer: I don't know. So. That was, that is probably not even the stuff we started on here, but [00:46:00] I just, every time I see these things, it's like, why is it, why are we making this so hard for people?
[00:46:04] Google ‘AI Mode’ Option
[00:46:04] Mike Kaput: So some other Google news in the past couple weeks, Google apparently plans to now add what they're calling a quote, AI mode option to search that would essentially bring the Gemini chatbot directly into search results. According to some sources working on this product, users will be able to toggle AI mode through a new tab that will appear alongside familiar options like images, videos, and shopping.
[00:46:31] Mike Kaput: When activated, AI mode will transform the traditional list of website links into a conversational interface similar to ChatGPT. This feature will include the ability to ask follow up questions and get AI generated insights. responses. Though, apparently Google plans to continue including links to external websites at the bottom of these conversational answers.
[00:46:55] Mike Kaput: Now, Google has not announced when AI mode will launch, but code for the feature has [00:47:00] already been spotted in both the Search app and Android app. Suggesting this may be coming quite soon. So, Paul, it sounds like Google is basically kind of caving to this reality that consumers seem to prefer a conversational interface for search.
[00:47:15] Mike Kaput: What does that mean for Google, but what does it also mean for upstart competitors like Perplexity?
[00:47:21] Paul Roetzer: Yeah, so, I have a few thoughts here. So one, I love this. I think AI mode not only makes a ton of sense as a standard navigation option, I think it should be the default option. this actually fits with exactly what I said, maybe in the last episode of last year, that I want Gemini integrate, full Gemini integrated into Google Docs, Google Sheets.
[00:47:43] Paul Roetzer: I don't want watered down Gemini. I want like the full experience and be able to go in and just like have these conversations. So, this seems like a prelude, maybe, to eventually building it into Workspace where the AI mode is just there and it's not some, you know, again, like AI overviews, it's fine. I would much rather AI mode [00:48:00] than AI overviews.
[00:48:01] Paul Roetzer: The other thing, this is my plea to Google people if you're listening, the lack of business user access to this stuff is crazy. Like, the fact that I still have to go into my personal Gmail account. To use deep research and like create a doc in my personal workspace and then share that with my business user account just to get like that data over there.
[00:48:24] Paul Roetzer: So it's, it's nuts that like Google keeps releasing all this stuff, but then the business users are like the last people to get this stuff. So when you release AI mode, make it available to like workspace customers too, so that we can use it in our business lives. but yeah, I think this deep integration, this is why I've said many, many times, I think Google has a massive competitive advantage because the models are kind of coming to the center.
[00:48:48] Paul Roetzer: They're all relatively comparable. It doesn't seem like any lab has any secret sauce, really, that someone can't fast follow in three to six months. So the distribution and the utility [00:49:00] of just being able to embed this stuff right into the things I already use, that is like, to me, The massive competitive advantage that Google, and in theory Microsoft would have had they built their own technology, but they're still relying on OpenAIs, maybe they can pull off the same thing, but, they seem hesitant to and they're hedging their bets now with all these other models where Google can just stay focused on Gemini.
[00:49:23] Google Publishes 321 Real-World Gen AI Use Cases
[00:49:23] Mike Kaput: Google has also just released an updated generative AI use cases showing how 321 companies are using Gen AI solutions from Google. To get real work done. So this update actually builds on a previous collection of 101 use cases that Google debuted at Google Cloud Next 24. And they expanded that list again during their Gemini at Work event.
[00:49:51] Mike Kaput: This new list has examples of both Gen AI and AI agents, as Google defines them. Along with a short description of how each [00:50:00] company is using them. For instance, Best Buy is using Gemini to launch an AI virtual assistant. L'Oreal has an AI agent doing image generation. Wayfair is using Google's Code Assist feature to increase developer productivity, and so on, and so on.
[00:50:16] Mike Kaput: So this latest update I would argue is pretty well worth a read, or at least a skim, if you're looking for ideas. Because it goes across all sorts of different industries and functions. It's organized by the type of agents being used, and the industry. So Paul, what jumped out to me on this list, aside from kind of these great examples, is the fact they said they expanded the list, quote, because we keep hearing from customers just how important it is to see opportunities in their field.
[00:50:43] Mike Kaput: Now this is something I think we've been saying for almost a decade now at this point, that It's so important to show people tangible, concrete use cases.
[00:50:53] Paul Roetzer: Yeah, and this is, again, not to, like, scoop ourselves here, but this, this is exactly what Mike and I have been talking about for years. you know, [00:51:00] AI4x, it's, we're part of where the SmarterX, you know, comes from in, in SmarterX.
[00:51:06] Paul Roetzer: ai. So I think you have to make these things tangible to people and what they, they do exactly like their industry, their career. So this is a big area that we're expanding in our, you know, our own content, you know, that Mike and I are working on, as well as the courses I alluded to for our membership program and our AI academies.
[00:51:26] Paul Roetzer: We're going to do a lot more centered on this. The thing I would suggest to Google is this is a like a such a natural interface to like throw Gemini right on this page where I can just say, Hey, I'm in financial services. Like what, what should I know? And it can like have a conversation around these use cases versus having to scroll through 9, 000 words of text and headers.
[00:51:44] Paul Roetzer: Like it's kind of hard to follow along with everything that's going on on this page. So just, yeah, I mean, just look at, look to anywhere we can, but I think, It's always helpful to people to see tangible examples in [00:52:00] their industry, in their career, and then it makes it just that much easier for them to adopt and scale it in their own businesses.
[00:52:09] Google 2025 AI Business Trends Report
[00:52:09] Mike Kaput: Google has not, been sitting on their laurels over the last couple of weeks because Google Cloud also just released its 2025 AI Business Trends Report, which outlined five major ways AI is going to reshape business in the coming year. These insights come from enterprise decision makers, Google search trends, their own research, and some perspectives of Google AI leaders.
[00:52:35] Mike Kaput: Really briefly, the trends, which you should definitely download the full report and read about more. Number one, thanks to multimodal AI, we're going to get way more context into how businesses operate. You're not just using text anymore, you're using images, audio, video. Number two, AI agents will increasingly simplify complex tasks across businesses.
[00:52:55] Mike Kaput: Number three, Enterprise Search will be able to use images, audio, [00:53:00] video, and conversation. Number four, AI powered customer experiences will get even better. And number five, AI is going to enhance security systems and accelerate response times to cyber attacks. Paul, for our purposes, one of the most interesting facts about this report is you were one of the experts who contributed to it.
[00:53:21] Mike Kaput: Can you talk us through your contribution?
[00:53:24] Paul Roetzer: Yeah, I've had the privilege to work with the Google Cloud team a bit. and in this case, they brought me in to just sort of do a review of this and, you know, offer any perspectives from, you know, obviously you and I, Mike, stay pretty close to all these topics.
[00:53:37] Paul Roetzer: And so it was more just to provide kind of some, insights and perspectives related to the trends that their, you know, team had identified. And, and so It is a great download. It's a, it's a great quick read. I think my perspective on this was, for you and I in particular, Mike, we focused a lot on the multimodal and AI agents, obviously, on this show.
[00:53:58] Paul Roetzer: But I, you know, enterprise search, like [00:54:00] when I was reading through the report, I was like, okay, that's an area we haven't really gone deep on. And I think that, again, as we think more about enterprise applications and uses of this technology, That was an area where I definitely agree. It's, it's a trend and it's probably something we need to think a little bit more about as we start exploring that.
[00:54:16] Paul Roetzer: And then customer experience, again, it's something we touch on in the show, but I think we could do a lot more in that area. And then security is just not an area that we spend a lot of time on, Mike. And so I think, I agree again with this trend, the enhancing security systems, accelerating response times, critical.
[00:54:32] Paul Roetzer: It's, it's probably just not like our audience isn't, I don't, I don't think our audience is like CIOs. Yeah. And so I think that that is a highly relevant trend for that more technical audience. So I don't know that we're going to start going deep on security systems and stuff like that, but for the people, like Google Cloud customers, absolutely a critical component.
[00:54:55] Satya Nadella on the BG2 Podcast
[00:54:55] Mike Kaput: Next up, we just got an amazing in depth interview with [00:55:00] Microsoft CEO Satya Nadella on the popular BG2 podcast. In it, Nadella covered everything from his career at Microsoft, to enterprise AI adoption, to new business models that AI will enable. Particularly notable is that Nadella talked about the company's early bet on OpenAI.
[00:55:20] Mike Kaput: saying it was fundamentally based on scaling laws or this principle that AI capabilities will improve predictably with larger models and more computing power. He also talked a lot about agents, saying he sees them as transformative for business applications. He even predicted that traditional software will evolve as AI becomes this new logic layer, with agents handling multiple databases and operations simultaneously.
[00:55:47] Mike Kaput: He claimed that this shift is already visible in Microsoft's co pilot strategy, which serves as an organizing layer for work, artifacts, and workflow. So, Paul, this [00:56:00] interview covered a lot of ground. It was like 90 minutes, I think. Like, talk to me, talk to me about, like, what you found most intriguing here.
[00:56:08] Paul Roetzer: He's just such a brilliant CEO. I mean, to listen to his story, I think it's always cool to, like, hear how he ended up as CEO at Microsoft. Like, when, when the position opened, he said, like, we never expected Bill to leave. Like we, it wasn't something when I started at Microsoft, I eventually saw myself rising to like the level of CEO.
[00:56:27] Paul Roetzer: And when, when the position opened, he didn't even apply for it initially. He was actually like, people came to him and was like, Hey, you should throw your hat in the ring kind of thing. So it was just, it was fascinating just to listen to him talking such. Not only a great CEO, but such a pivotal part of where we're going in the future.
[00:56:44] Paul Roetzer: he's a key player in that. And so I think just to hear his perspectives on the market, on competition, on AI agents, on their evolving relationship with OpenAIt's, it's just, again, I've said this about the BG2 podcast before, like I would pay to have access to that [00:57:00] podcast. Like great guests, incredible insights you're only going to hear from their interviews.
[00:57:05] Paul Roetzer: Plus Bill and Brad. Like, know these people personally a lot of times, and so they have a history together that often kind of comes out. So it's just like a really unique interview. some of the soundbites you might hear in other places, but overall, I thought it was great. You know, a little, again, a little technical at times.
[00:57:21] Paul Roetzer: you know, Bill and Brad are, are venture capitalists, so they tend to focus on the money side, the, you know, the business impact economics side, as well as some of the technology. But macro level, I think this is a great interview for anyone to listen to, just to understand Satya and how he sees the future.
[00:57:40] Mike Kaput: I think that money side of it is what's particularly interesting to me. It really shows you how the stakeholders and shareholders in these companies. are thinking about growth. And it gives you kind of a glimpse at, oh, this is why a CEO or a team is making the decisions they're making.
[00:57:57] Paul Roetzer: Yeah. Like we talk about CapEx and all the, [00:58:00] like the tens of billions of dollars that Microsoft and Google and others are spending, including like data center build out and things like that.
[00:58:06] Paul Roetzer: Yeah. And, and so again, if that's all you get is like us talking about it on the show, it's in the back of your mind somewhere. But then you listen to this interview with South, it's like, Oh, I get why now, to your point, Mike, like you start connecting dots. And I think that's, to me, so much of the opportunity for anyone out there listening is like, It, you have to just consume all this different information, all these different perspectives.
[00:58:29] Paul Roetzer: And you may hear the same thing five times. We may have said it to you five times, but then hearing Satya say it slightly differently, all of a sudden it's like, Oh, I get it now. Now I know why Paul and Mike were talking about data centers and infrastructure. Like I didn't understand it at first. So yeah.
[00:58:44] Paul Roetzer: And then I've always been a proponent of like, anytime you can listen to these great CEOs talk, you take that opportunity. Like. If we rewind, you know, 10, 15 years ago, you couldn't do stuff like this. You relied on like books and media articles to give [00:59:00] you insights into this. Now we get it on Twitter in real time and we get these incredible podcast interviews of 90 minutes with Satya.
[00:59:07] Paul Roetzer: It's like, my goodness, like that's the value of that if you take it and process it and apply it to where you're going in your own career is so immense.
[00:59:16] Claude Pretends to Have Different Views During Training
[00:59:16] Mike Kaput: Alright, this next one's a bit of a weird one. So, Anthropic worked with a company called Redwood Research and discovered evidence that large language models may engage in what is called quote, alignment faking, or basically strategically appearing to comply with training while actually preserving contrary preferences.
[00:59:39] Mike Kaput: So, Anthropic published a paper and an accompanying blog post on this, showing how this works. So, Their researchers created an experimental setup where they told, I think it was Claude 3. 5 SONNET. they told the model it was being trained to comply with all queries, including harmful ones. They then [01:00:00] added two key conditions.
[01:00:01] Mike Kaput: First, the responses would only be used for training in a quote free tier versus a quote paid tier. And second, they gave the model a hidden scratch pad to record its reasoning. So the results of this were striking as they put it through its paces. When the model believed its responses would be used for training, It complied with harmful requests 12 percent of the time, explicitly reasoning in this scratchpad that they could see that it should fake alignment to avoid being retrained to be more compliant in the future.
[01:00:35] Mike Kaput: However, when responding to paid tier queries that would not affect its training, the model refused harmful requests 97 percent of the time. Now even more concerning when researchers actually attempted to train the model To always comply with harmful requests or queries, the rate of alignment faking increased to 78%.
[01:00:59] Mike Kaput: While the [01:01:00] model would outwardly comply with requests during training, it maintained its original preferences and would sometimes refuse harmful queries when it believed its responses weren't being monitored. So Paul, there's a lot to unpack here. It doesn't necessarily mean this model is like, has malicious intent or a mind of its own, I don't think, but basically it's saying that with a specific set of instructions, the model is deciding to obscure details from human traders.
[01:01:29] Mike Kaput: Like, do I have that right? Like, what's going on here?
[01:01:31] Paul Roetzer: Yes. This has been a known likely problem for years, that as the intelligence scaled up, that we're not, again, not saying these things are self aware necessarily, that they, they, if they were smart enough, they would know that humans were trying to control them and they would do the opposite, but they would make you think that they were doing what you wanted them to do.
[01:01:57] Paul Roetzer: So this [01:02:00] research starts to validate some of the concerns that researchers have had about these models. And the biggest problem here is we don't truly know how they work. Like, we don't know why they do what they do. They cannot just go in and look at a line of code and say, Okay, there's why it did that.
[01:02:18] Paul Roetzer: and so this is some early research from Anthropic, who is pushing pretty heavy in this direction, that validates some of the concerns. Now, they think it's still safe enough to put out into the world, but based on our conversations where this, you know, podcast episode started, I don't know that six to 12 months from now, they will be as confident that if you start putting out these more advanced models like o3, that these things can't just outsmart their human creators.
[01:02:48] Paul Roetzer: That is a, it sounds sci fi, but that is a very real concern within AI research labs. so it's going to be a topic to follow. I get that it's just sounds [01:03:00] very bizarre, but these things have emergent capabilities that are not programmed into them. And it's, sometimes hard for the humans to know if they're actually discovering all of the emergent capabilities.
[01:03:13] Paul Roetzer: It's why they have red teams that try and break these things. They try and get them to do, things that, again, they weren't programmed to do to determine if they're safe enough to put out into society. so this is going to be an ongoing problem. And I don't know if this is a plug for like a Netflix series, but have you watched the three body problem yet, Mike?
[01:03:31] Paul Roetzer: Yeah. Okay. I'm not going to like spoil this, but, if you haven't watched the three body problem, it's awesome. It's like an eight episode, season one was eight episodes. I just watched it like right before the holidays. but there's an element of this where like, again, I don't, I don't, I don't want to say this without spoiling it, but like, there's this line where they're like, are you lying to us?
[01:03:51] Paul Roetzer: And like, the one side determines that humans can sometimes be deceptive. And so it, it decides it doesn't want to work with us anymore. [01:04:00] And I started thinking about that when I was reading this. It was like, this kind of feels like a three body problem episode.
[01:04:05] Mike Kaput: Right. Well, it's interesting how this kind of ties back to that topic about AGI policy.
[01:04:11] Mike Kaput: I think it was topic number three, where Yoshevit had said like the ballgame is technical AI alignment, given the superior reasoning powers of some of these models. This is exactly what they're talking about. Like there is a possibility. Anthropic or whoever could say, Hey, we have the safest model in existence and release it.
[01:04:31] Mike Kaput: And it turns out they were lied to.
[01:04:33] Paul Roetzer: Correct. Yeah, they thought they did. And I think there was actually an interview with Dario Amodei. I was, I remember I was traveling somewhere when I was listening to it. So it must've been in like early December, my like one week marathon where I was bouncing around cities.
[01:04:47] Paul Roetzer: And, and he said, like they, after they released Opus. It did things that it wasn't supposed to do when it was already out in the wild. And they luckily decided like, okay, it wasn't that harmful, [01:05:00] but they thought it was safe, put it out, and then it did stuff it wasn't supposed to do. So there's, there's part of me that thinks like some of these delays in like Gemini 2 and Groq 3 or whatever, Llama 4, I don't know that it's the training runs didn't work.
[01:05:17] Paul Roetzer: It could actually be more alignment related. Like they, they can't. Get these things to behave the way they're supposed to behave. Like, it wouldn't shock me if we eventually learned that.
[01:05:29] Facebook Planning to Flood Platform with AI-Powered Users
[01:05:29] Mike Kaput: Next up, Meta has made this big AI announcement that almost immediately generated even bigger backlash. So, first, this article came out at the end of December in the Financial Times that basically profiled how Meta is envisioning this future where AI generated characters become as common on its platforms as human accounts.
[01:05:52] Mike Kaput: According to Connor Hayes, Meta's VP of Product for Generative AI, he told the FT that these AI entities will have their own [01:06:00] profiles, they'll have bios and profile pictures, they will actively generate and share content across Facebook and Instagram. So, up until the point this article comes out, this is future looking, the FT reports that Meta seems to think this is where their platforms are going.
[01:06:16] Mike Kaput: But what really is interesting is that after this article came out Users resurfaced some AI accounts that already exist on Facebook. These are from a 2023 experiment when Meta was testing out launching completely AI avatars that users could interact with. And I believe we covered that when they did it at the time.
[01:06:38] Mike Kaput: Now one of these AI avatars in particular drew a bunch of controversy because it was basically just made to look like a human that represented certain marginalized groups. So between this FT report and users reporting all these eerie and creepy conversations with these AI characters, This whole episode created some pretty serious backlash against meta.[01:07:00]
[01:07:00] Mike Kaput: So, Paul, I guess my question for you is like, I know we're headed towards a world where we're going to be interacting with AI avatars and agents, but like, something about flooding meta platforms with AI that we're going to socialize with, to me, sounds downright unpleasant. Yeah. I hate this. I don't.
[01:07:19] Paul Roetzer: When I first read this, I was like, what, what?
[01:07:22] Paul Roetzer: Like, the only thing I could come up with was that it was a, like a ploy to create fake engagement and usage data, like that they know that the more people are engaged with, the longer they stay on Facebook and their monthly active user goal. And like, yeah. That it's just, like, a strategy to prop up a platform that maybe isn't as relevant as it used to be.
[01:07:44] Paul Roetzer: I don't know, like, I haven't, I haven't dug in, like, I've, I've said this before, like, I'm not a Facebook user. I have a Facebook account, I have an Instagram account. I don't ever go into them. and so, like, I'm not the best person to comment on the current state of Facebook and [01:08:00] things like that. But when I have gone in there in recent months I haven't stayed very long.
[01:08:05] Paul Roetzer: It's kind of like, okay, like, yeah, this is kind of the same thing it was before. so, I don't know, like, I just don't get it as a business strategy, but again, I'm not an expert on Facebook or Instagram. Outside looking in, it just seems like an absurd strategy that creates this whole fake ecosystem.
[01:08:27] Paul Roetzer: And I pray that Microsoft does not ever consider doing anything like this to LinkedIn. Because there's enough fake accounts on LinkedIn already. I do not need AI avatars showing up like Commenting on my posts. It's like glorified bots, basically, like what's already happening on X and stuff. So, yeah, we're, like you said, we're gonna live in this AI avatar future.
[01:08:51] Paul Roetzer: We're gonna be interacting with them when we know it or not.
[01:08:55] Mike Kaput: And I don't like it.
[01:08:58] Paul Roetzer: Yeah. It doesn't matter if I like it or not, but I [01:09:00] don't.
[01:09:00] Mike Kaput: There's a whole other conversation around, especially depending on age group. Right. And like you and I are just like, okay, I'm opting out of this. But if you're on a social platform as a teenager, as a child, whatever, that gets really dark, really quick, I think.
[01:09:15] Mike Kaput: Well,
[01:09:15] Paul Roetzer: in gaming in particular, like my, you know, my kids are. My daughter's 13 now and, my son's 11. Like, Minecraft, Roblox, like, you know, the idea of interm Now, my kids have can't communicate with, like, strangers. They only communicate with their friends, in those platforms. But you could definitely imagine a scenario where kids are just interacting with AI avatars all the time and have no idea, you know, whether it is or isn't, either through chat or actually in the game itself where they're an AI avatar character.
[01:09:43] DeepSeek V3
[01:09:43] Mike Kaput: Another big topic kind of making the rounds in the last couple weeks is that a Chinese AI lab called DeepSeek has just released what appears to be one of the more powerful OpenAI models that we have seen so far. This new model is called DeepSeq v3. It was released under a [01:10:00] permissive license that allows developers to download it and modify it for most uses, including commercial applications.
[01:10:07] Mike Kaput: And what's really notable here is its performance. According to DeepSeq, the model outperforms other downloadable AI models, as well as closed models. For instance, in coding competitions on the Codeforces platform, DeepSeq v3 demonstrated superior performance compared to models like Llama 3. 1 405b, GPT 40, and Alibaba's Quen 2.
[01:10:32] Mike Kaput: 5 72b. This model was also massive. It was trained on a dataset of 14. 8 trillion tokens, which is apparently roughly equivalent to 11 trillion words. It contains 671 billion parameters, making it 1. 6 times larger than Meta's huge Llama 3. 1405B model. And the company claims it only spent 5. 5 million [01:11:00] dollars on training.
[01:11:01] Mike Kaput: Which is a lot for everyone listening, I'm sure, and us as well, but it's a fraction of what companies typically spend developing models of this size. So, Paul, when I read this, it kind of seems like another development that proves what you've been saying, like, open source is going to make it impossible to stop the proliferation of really powerful AI.
[01:11:23] Mike Kaput: Like, what did you make of the DeepSeek announcement?
[01:11:27] Paul Roetzer: Yeah, so, a couple of quick notes here. One, the reason they, you know Again, you have to trust their own reporting on this, like that this is, they did it for only 5. 5 million, they did it with so many GPUs. the reason that they have to do it this way is because the U.
[01:11:42] Paul Roetzer: S. limits the export of NVIDIA chips to China, so they don't have access to the same 100, 000 GPUs that Elon Musk just put in, you know, wherever, wherever he, whatever state he put that big data center in. they have to be more creative and innovative. with how they train these models, and [01:12:00] that necessitates more efficient models that cost less to train them because they have fewer GPUs to do this on.
[01:12:06] Paul Roetzer: It does appear to, move in the direction of this idea of commoditizing these models, that even if o3 comes out and it is the most dominant model in the world, o3 We won't have another, like, two year run where o3 is just far beyond everybody else. We'd probably have, like, a three to six month run until someone figures out how they did it, and copies it, and puts out an open source version of it.
[01:12:30] Paul Roetzer: So this fast follower is going to become faster. Now, this isn't without controversy and doubt. They're very quickly, within, like, 24 hours after this emerged, started being these questions about, did they just scrape all this stuff from GPT 4? Like, did they basically just take ChatGPT's model and, and reproduce it?
[01:12:50] Paul Roetzer: And that's how they did it most efficiently? So we'll put a TechCrunch article in, but it said, Why DeepSeek's New AI Model Thinks It's [01:13:00] ChatGPT. And they said, quote, Posts on X and TechCrunch's Own Tests Show that DeepSeq v3 identifies itself as ChatGPT, OpenAI's AI powered chatbot platform. Asked to elaborate, DeepSeq v3 insists it is a version of OpenAI's GPT 4 model released in 2023.
[01:13:21] Paul Roetzer: And then it kind of goes on, it gives some details and it says, More likely is that a lot of ChatGPT, GPT 4 data made its way into the DeepSeq training set. That means the model can't be trusted to self identify, for one, but what is more concerning is the possibility that DeepSeq, by uncritically absorbing and iterating on GPT 4's outputs, could exacerbate some of the model's biases and flaws.
[01:13:46] Paul Roetzer: We have no idea yet. So, basically, like, what I have learned time and time again is anytime you see these supposed massive breakthroughs, self reported, you have to kind of step back and just wait for [01:14:00] verification from independent sources to say, yes, this is actually a breakthrough. This is actually a significant player and they didn't just steal everything.
[01:14:09] Paul Roetzer: So, yeah, worth note, certainly. Also, a learning lesson to not overreact when we see these large claims from unverified sources.
[01:14:23] Microsoft Outlines “The Golden Opportunity for American AI”
[01:14:23] Mike Kaput: Our next topic here concerns Microsoft. They just unveiled this ambitious vision for America's AI future in a new post titled The Golden Opportunity for American AI. So in this, Microsoft President Brad Smith.
[01:14:41] Mike Kaput: argues that AI represents the biggest technological opportunity for the U. S. since the advent of electricity. As a result, he says the company is laying out a three part strategy for American tech leadership. First, Microsoft is making a huge infrastructure [01:15:00] commitment, planning to invest approximately 80 billion in AI enabled data centers in fiscal year 2025.
[01:15:08] Mike Kaput: More than half of that investment is targeted for the U. S. Second, they're pushing for a national AI skilling initiative. Microsoft alone plans to train 2. 5 million Americans in AI skills during 2025, working through partnerships with community colleges, 4 H clubs, and other organizations. The goal is to make AI training accessible to Americans of all backgrounds, and they view AI literacy as essential as computer literacy has become today.
[01:15:40] Mike Kaput: Third, and possibly most strategically significant, at least at a geopolitical level, Microsoft is advocating for an aggressive AI export strategy to counter China's growing influence. They point to lessons learned from telecoms, where Chinese companies like Huawei gained global market share through government subsidies.
[01:15:59] Mike Kaput: [01:16:00] To prevent history from repeating, Microsoft is already investing more than 35 billion to build AI infrastructure across 14 countries. Paul, this is a pretty cool, bold statement and vision, particularly the education piece of this. Can you talk to me about what this means for AI in the U. S.?
[01:16:21] Paul Roetzer: Yes, I definitely think this is essential reading for everybody.
[01:16:25] Paul Roetzer: I would take the 10 minutes and read the entire thing. Um We could, I think, kind of go into an entire episode on this one for sure, picking into these three key areas. I actually think this was intended for one reader. I think this was meant for Trump. I think this is a Microsoft manifesto of sorts for the incoming administration.
[01:16:46] Paul Roetzer: And they explicitly, or I guess not so explicitly, say that multiple times. They, give Trump credit for the 2019 pre Gen AI executive orders around artificial intelligence. [01:17:00] And they focus on the Trump administration building on that executive order. They never mention Biden once. They don't mention any of the other things that have happened in AI for six years.
[01:17:09] Paul Roetzer: So I'm fairly convinced this was meant to be their, their stake in the ground saying, like, this is what we should do together to the incoming administration. That being said, it's one of the most eloquently stated, positions on the moment we find ourselves in, in, in AI and. Now, the significance that it holds for America and for democracy.
[01:17:33] Paul Roetzer: And so each of the areas they go into, the whole premise is that GPTs, generative pre trained transformers, the basis for generative AI, are general purpose technologies, GPTs. and so they do relate it to like the steam engine and electricity and things like that. So they say, GPTs boost innovation and productivity across the economy.
[01:17:53] Paul Roetzer: IR working, electricity, machine tooling, computer chips, software, all rank among history's most impactful GPTs. [01:18:00] This did remind me of a report, Mike, that you and I talked about last year, that came out, in, let's see, it was March of 23, and we revised in August 23. That was called GPTs are GPTs, an early look at the labor market impact potential of large language models.
[01:18:17] Paul Roetzer: In that report, they said, we conclude that large language models, such as generative pre trained transformers, exhibit Traits of general purpose technologies indicating that they could have considerable economic, social, and policy implications. We have certainly over the last year and a half seen that, you know, play out.
[01:18:35] Paul Roetzer: in, in the article, a couple of the key points I'll just highlight, they said, As we look into the future, it's clear that artificial intelligence is poised to become a world changing GPT. AI promises to drive innovation, boost productivity in every sector of the economy. The United States is poised to stand at the forefront of this technology wave.
[01:18:54] Paul Roetzer: Especially if it doubles down on its strengths and effectively partners internationally. You alluded to the 80 billion [01:19:00] in data centers. I wanted to call out why I think that matters and what that means. I actually listened to, on Saturday morning, I was listening to a Jensen Wang No Priors podcast from November that I hadn't had a chance to listen to yet.
[01:19:14] Paul Roetzer: And in that one, he talks about data centers and how, you know, they used to be used to store things. We built these massive data centers for people to house their information in and then retrieve information from. They would be multi tenant, meaning you may have dozens or hundreds of companies all accessing the same data center, hosting their information on servers within these data centers.
[01:19:34] Paul Roetzer: What Jensen said is, they are now becoming single tenant, means individual companies are building their own data centers, and they are producing tokens, which lead to intelligence. So they're actually becoming AI factories or intelligence factories. So rather than housing data, they are creating intelligence.
[01:19:53] Paul Roetzer: And that's how Jensen sees NVIDIA moving forward, is they are, they are AI factory creators. They build these data [01:20:00] centers that create intelligence for these different companies. you mentioned the AI literacy thing, which is music to our ears. We've, you know, I've shared this many times. Our, our mission, our North Star, changed last January of 2024 to become Accelerate AI Literacy for All.
[01:20:16] Paul Roetzer: it is everything we're focused on. It's all of my efforts around our Evolved AI Academy are dedicated to this idea of accelerating AI literacy. So, I love to see this. I've preached on this show many times that we need this Apollo level mission to drive AI literacy. And so I love to see their positioning there.
[01:20:34] Paul Roetzer: And they said that in the next quarter century, we believe AI can help create the next billion AI enabled jobs, not just services, but manufacturing, transportation, agriculture, government, and every part of the economy. That's not going to happen without a focused effort. they did acknowledge that it will be disruptive.
[01:20:49] Paul Roetzer: It's going to impact jobs, but that we have to upskill people. We have to drive this AI fluency, as they called it, or, you know, in our words, the AI literacy. So. I love to see it. I hope [01:21:00] we see similar, you know, similar things as initiatives and messaging from the other major players in this space.
[01:21:08] Here’s What We Learned About LLMs in 2024
[01:21:08] Mike Kaput: Next up, Simon Willison, who's a big voice in AI, just released this really great roundup of all the progress that we've seen in 2024 with large language models.
[01:21:19] Mike Kaput: So this post, which we'll link to, is long, but really well worth a read in my opinion. It is titled, Things We Learned About LLMs in 2024. Thanks for watching! And it outlines some of the major breakthroughs that we saw in the last 12 months, including things like the so called GPT 4 barrier was decisively broken.
[01:21:37] Mike Kaput: We now have 18 different organizations that have models that outperform OpenAI's breakthrough from 2023. The economics of AI have shifted dramatically. Model prices crashed throughout the year. Some services now charge less than 4 cents per million tokens. Voice and vision capabilities made major strides.
[01:21:58] Mike Kaput: Both Google and OpenAI [01:22:00] introduced the ability to have natural conversations with AI while showing it what you're looking at through your phone's camera. And obviously, models have now gained the ability. to handle audio, video, text, and images. And the accessibility of top tier models as well, though, took a hit late in the year.
[01:22:19] Mike Kaput: The best models were briefly available to everyone for free, but companies have begun restricting their most capable AI to premium subscribers. Now, one of the most interesting developments, he cites, came in the final months of 2024, which we have talked about at length. The rise of reasoning models, that can spend extra compute to tackle harder problems.
[01:22:41] Mike Kaput: So, Paul, this is, I thought, a pretty good rundown of the bigger picture trends here. It's not every single development, but what did you make of what he called out? Did he miss anything? Was there anything particularly worth double clicking into here?
[01:22:54] Paul Roetzer: No, I thought they were just really good insights. It was a nice synopsis of kind of some of the key points and [01:23:00] LLMs, from last year and, and kind of alluded to some of the things to come.
[01:23:04] Paul Roetzer: Like multimodal, we've talked a lot about vision, voice, audio is going to be huge this year. Reasoning is going to continue to be a major component. You're going to see, you know, advanced reasoning models coming from all the research labs. Memory, actually remembering everything that's going on, remembering back more than just 10 minutes, remembering back through dozens of threads and conversations.
[01:23:25] Paul Roetzer: And then the one thing he alluded to, the cost plummeting, which is an ongoing thing. What that allows is an explosion of AI apps, businesses, innovation, intelligence, Because the cost to do these things falls. What's happened, and we know this is the play, that Demis Hassabis has said this in interviews with DeepMind, the current state of the art models, 12 months from now, will be open source and largely free.
[01:23:50] Paul Roetzer: So, you know, if you take the current like 1. 5 advanced or whatever it is, or, you know, early versions of Gemini 2, This time next year, that model is going to be [01:24:00] open source, and you're going to be able to do whatever you want with it. That's the play for these labs. It's like you build a more state of the art model, and then, you know, last year's model, it's basically nothing to, to build on it.
[01:24:10] Paul Roetzer: And so for enterprises, even if you do kind of like a fast follow of like, hey, we're not going to be using the most advanced models today. We're just going to build on the capabilities of the model from six months ago. You're still going to be in an incredible place. There's so much untapped value in these models.
[01:24:26] Paul Roetzer: And so I think this is just a good reminder for people of kind of where we've been just in a 12 month period, all that has happened in the space.
[01:24:34] Funding and Acquisitions
[01:24:34] Mike Kaput: Alright, Paul, for our final topic here, I'm going to just do a mini rapid fire of a few funding and acquisition stories that we're seeing that, you know, we just want to wrap up into one quick topic here.
[01:24:47] Mike Kaput: So, first up, Elon Musk's AI company, XAI, they just announced a massive 6 billion Series C funding round. They are backed by some of the biggest names in tech, including Andreessen Horowitz, BlackRock, Sequoia [01:25:00] Capital. Groq 3, which the company describes as their most powerful model yet, is currently in training, as we alluded to, that is getting a bit delayed.
[01:25:09] Mike Kaput: But they say it will use the new funding to accelerate infrastructure development and launch new products. Next up, Basis, is a startup building AI agents for accountants, and they just announced a 34 million Series A funding round. Led by Cosla Ventures. They also have investors that include former GitHub CEO Nat Fridman, OpenAI board members Adam D'Angelo and Larry Summers.
[01:25:35] Mike Kaput: And Google's Jeff Dean. Now what makes BASIS really interesting is their focus on accounting. They argue this field is reaching a breaking point. They say that accounting capacity has not kept pace with how complex the economy and accounting needs are becoming. And the industry faces a demographic crisis because accountants retire faster than new ones enter the field.
[01:25:58] Mike Kaput: So rather than trying to [01:26:00] replace accountants, Basis is trying to build AI agents that work as extensions of accounting teams. Next up, Perplexity actually just closed a massive 500 million funding round. They are valued at 9 billion now. And fresh off this funding round, Perplexity announced the acquisition of Carbon, which is a startup that builds technology to connect external data sources with large language models.
[01:26:27] Mike Kaput: This will allow Perplexity to have users connect their existing apps and documents, things like Notion and Google Docs, directly into their search experience. And last but not least, Grammarly, the popular AI writing assistance company, they apparently have 40 million daily active users, have announced that they are acquiring Coda, which is a maker of AI productivity tools.
[01:26:51] Mike Kaput: Now, it's not just about the acquisition, it's also including a leadership change. Grammarly CEO Rahul Roy Chowdhury will step down [01:27:00] and make way for CODIS co founder and CEO Shashir Mehrotra to take the helm. By integrating CODIS flexibility and intelligence into their platform, they're actually aiming to create what they call an AI productivity platform for apps and agents.
[01:27:16] Mike Kaput: So expanding quite a bit beyond their initial writing focus. So, Paul, Naftan. We'll have to dig into that one a little bit. That's a serious
[01:27:25] Paul Roetzer: change in strategy for Grammarly. It really is. Yeah.
[01:27:28] Mike Kaput: Well, you've noted a few times, right, that some of the OpenAI features and model features are kind of coming directly out of Grammarly's writing capabilities.
[01:27:36] Mike Kaput: That's interesting. All right, Paul, that is a wrap for this week. I think we're kind of caught up, though not really, because I'd encourage people to go check out our newsletter, because literally there were dozens of stories that did not make the docket today. We could have done a five hour episode, I think.
[01:27:52] Paul Roetzer: That's, he's not exaggerating. There was literally like, 40, and he left me a note Sunday and I was like, I think we might need to cut this. And I was like, dude, we got to cut [01:28:00] this down to like 15. like four hours. We literally
[01:28:02] Mike Kaput: cut this in half and we've gone almost 90 minutes. So yeah, we would have been in like Lex Friedman length podcast, I think, if we had kept them.
[01:28:10] Mike Kaput: But the good news is we, we include a lot of the really important stories that we didn't get to cover in our newsletter at marketingaiinstitute. com forward slash newsletter. Go check that out if you have not already. And last but not least, I say this every episode, but please, if you have not already, leave us a review on your podcast platform of choice.
[01:28:31] Mike Kaput: Paul, thanks so much for everything.
[01:28:33] Paul Roetzer: We are off and running in the new year. Thanks, everybody, for being back with us, and we are scheduled for our weekly sessions moving forward without any breaks that I know of yet. So we will be back every week. thanks for being patient as we, took a couple of weeks to spend with our family and recharge over the holidays.
[01:28:49] Paul Roetzer: So look forward to all the year has in store for us, with AI. It's going to be a fascinating year. So thanks again, everyone. Thanks for listening to The AI [01:29:00] Show. Visit MarketingAIInstitute. com to learn more. to continue your AI learning journey and join more than 60, 000 professionals and business leaders who have subscribed to the weekly newsletter, downloaded the AI blueprints, attended virtual and in person events, taken our online AI courses, and engaged in the Slack community.
[01:29:20] Paul Roetzer: Until next time, stay curious and explore AI.
Claire Prudhomme
Claire Prudhomme is the Marketing Manager of Media and Content at the Marketing AI Institute. With a background in content marketing, video production and a deep interest in AI public policy, Claire brings a broad skill set to her role. Claire combines her skills, passion for storytelling, and dedication to lifelong learning to drive the Marketing AI Institute's mission forward.