Intelligent Machines 854 transcript
Please be advised this transcript is AI-generated and may not be word for word. Time codes refer to the approximate times in the ad-supported version of the show.
Leo Laporte [00:00:00]:
It's time for Intelligent Machines. Paris Martineau is here. Jeff Jarvis is here. We didn't expect this. He's joining us from his hospital bed because he was so excited about our interview this week. We talked to Thomas Haig, the author of a new history of AI Intelligent machines. From the Pit is next.
TWiT.tv [00:00:23]:
Podcasts you love from people you trust. This is twit.
Leo Laporte [00:00:32]:
This is Intelligent Machines with Paris Martineau and Jeff Jarvis. Episode 854, recorded Wednesday, January 21, 2026. Welcome to the Pit. It's time for Intelligent Machines, the show. We cover the latest in AI robotics and all the smart doodads all around you in your home these days. Soon, someday we'll be doing this show and a humanoid robot will creep up behind me and brain me. Probably. Say hello to our wonderful Paris Martineau from Consumer Reports, where she's an investigative reporter.
Thomas Haigh [00:01:07]:
Hello.
Paris Martineau [00:01:08]:
I will be sad the day a robot brains you, but I will respect the robot's authority. Regardless.
Leo Laporte [00:01:15]:
I will have earned it. No doubt. The robot may be remotely controlled by one Paris Martineau. We're not sure.
Paris Martineau [00:01:23]:
We don't know.
Leo Laporte [00:01:24]:
Could be. Now, this is a little weird. I'm going to have to set this up. We did not expect Jeff Jarvis to be here today because Jeff.
Paris Martineau [00:01:33]:
In fact, we expected the opposite because that's the only thing that would make sense given what you're about to say.
Leo Laporte [00:01:40]:
Absolutely. He had took a tumble. He talked about it last week. He injured his coccyx. And you don't have to look that up. It's a part of his body near the nether regions and by making a bed, apparently which bed fought back. But then as a result, he got checked up and it turned out he had other issues. Joining us now from the hospital, from his hospital bed.
Leo Laporte [00:02:10]:
I don't believe this. Not Mike Elgin. Jeff Jarvis.
Paris Martineau [00:02:14]:
Hello, Jeff.
Leo Laporte [00:02:15]:
Are you okay?
Jeff Jarvis [00:02:17]:
Hello there, guy.
Leo Laporte [00:02:18]:
You had a fever.
Jeff Jarvis [00:02:19]:
What was your therapy?
Leo Laporte [00:02:20]:
Your fever was like 105 or so, 3.8.
Paris Martineau [00:02:23]:
I'm saying you're not allowed to be here. We're allowing you to be here only because you seem to really want to for a little bit for interview. But afterwards, we're gonna cast you off to go on forced bed rest.
Leo Laporte [00:02:36]:
This is.
Jeff Jarvis [00:02:37]:
That's all I've been doing. So it turns out I had compression fracture in my L3.
Leo Laporte [00:02:44]:
Oh, well, everybody.
Jeff Jarvis [00:02:45]:
Vertebrae.
Leo Laporte [00:02:46]:
That's not your coccyx. Yes.
Jeff Jarvis [00:02:48]:
No, that's not. That's not. And then I finally went to see urgent care on Thursday after the Sunday accident.
Leo Laporte [00:02:54]:
Yikes.
Jeff Jarvis [00:02:54]:
And I had A fever. And it was a little mysterious. And on Friday, it got much worse. The pain got wildly bad, horribly bad. Saturday, I could not get out of bed. I fell out of bed. Old men. And was carried by ambulance on a stretcher to the hospital here in Morristown.
Jeff Jarvis [00:03:17]:
And they tested and tested and tested and tested. And I'm very fortunate because they found a staph infection.
Leo Laporte [00:03:23]:
Oh, yikes.
Jeff Jarvis [00:03:24]:
Which can lead to sepsis, which can kill you. And so I'm on lots of antibiotics through the IV and physical therapy and all kinds of other things. So I'm glad to be here.
Leo Laporte [00:03:42]:
Worst nightmare. She thinks if I'm gonna work with two old guys, they're gonna start talking about their.
Paris Martineau [00:03:47]:
No, I'm just saying. I'm just sitting here thinking, jeff, I'm glad that you're safe. I'm glad that you're here talking with us.
Jeff Jarvis [00:03:55]:
This is pretty weird.
Paris Martineau [00:03:56]:
So I'm most of all. Is this the first time in Twitter history that someone has joined the podcast from a hospital bed?
Leo Laporte [00:04:03]:
Oh, most definitely. This is.
Jeff Jarvis [00:04:05]:
Is it the first time in podcast history it might be a hospital bed? Well, because it's so foolish. So. So the other. So I have been pushing for Thomas Haig to come on. You'll meet him in a moment to be our guest. And I felt awful that I was really fascinated by his work. I got to know him at the Computer History Museum event on the pioneers of desktop publishing, which was key to my next book, Hot Type. And so I thought I'd just call in to say hello.
Jeff Jarvis [00:04:39]:
And then Elgin now has an emergency, so he can't be here.
Leo Laporte [00:04:42]:
So we're glad you're here.
Jeff Jarvis [00:04:43]:
I'm here for a while.
Leo Laporte [00:04:44]:
Yeah. Well, let me introduce. And you introduced us to Thomas Haig. He's a historian of computing, which is kind of for both of us. Jeff and I, we're both fans of history. Kind of catnip. He's written many books. You'll find them at his website, tomandmaria.com tom.
Leo Laporte [00:05:01]:
Including a new history of modern computing, ENIAC in action, exploring the early digital. And his newest book is all about AI, the history of AI. Thomas, welcome to Intelligent Machines, the Brand that Wouldn't Die. I think that's a provocative title because you say AI is really a marketing term. Yes, we know it is now when Sam Altman utters from Sam Altman's lips. For sure. But it's a marketing term with quite a kind of ancient history. Tell us about.
Leo Laporte [00:05:43]:
Well, who first coined the term AI?
Thomas Haigh [00:05:46]:
Well, the term was invented by John McCarthy in 1955 to attach to a proposal to the Rockefeller foundation to get money to have a summer school and invite his friends to Dartmouth College.
Leo Laporte [00:06:03]:
In fact, a very famous conference of the first AI conference.
Thomas Haigh [00:06:07]:
Yeah. And it wasn't the first time people had talked about computers and thought there'd been a conference in Paris some years earlier. In fact, the idea that computers are like brains is as old as the electronic computer was something that was informing John von Neumann as he was developing modern computer architecture. But the specific phrase artificial intelligence has got that very exact time and purpose. It was invented. And my argument in the book is that it's functioned effectively for the most part with a large period of the AI winter when it went out of fashion as a means to market and sell things. That's not to say that it's not real. I mean, one of the things that I try and stress in the introduction to the book is that this is not to insult AI and say that it's not real, that it doesn't produce technologies that are worthwhile or work, but it does.
Thomas Haigh [00:07:06]:
And you can consider any academic discipline through the lens of being a brand. So you can think of physics as a brand or biology. But I think within computer science, if you compare artificial intelligence to other subfields like communication, databases, graphics, there's been a lot more historical change over time in what kind of approaches the AI brand has been associated with than there has been in those areas. So I think while you can consider anything as a brand, I think it's unusually informative when you consider the history of AI to, to bring to the forefront its brand like qualities.
Leo Laporte [00:07:52]:
Well, and also the way you name something informs how you think about it. And so I know it's a little chicken and egg, maybe we always thought about computers as kind of artificial brains, but as soon as you put the word artificial intelligence in the language, now that's what you're defining, really, it makes.
Paris Martineau [00:08:12]:
The whole endeavor much loftier than cybernetics.
Leo Laporte [00:08:15]:
A mechanical brain. Yeah. Here's a picture, by the way, from the book the Artificial Intelligence the History of a Brand of that first summer Dartmouth concert. Claude Shannon is there, Marvin Minsky and John mccarthy himself. How close to what we now think of as AI was this 1955 conception of AI? Is it something completely different or is it a reasonable precursor for what we have today?
Thomas Haigh [00:08:46]:
Well, in some ways the 1955 conception of AI was closer to what we have today than the 1975 or the 1985 conception of AI was.
Leo Laporte [00:08:56]:
The dates of the other AI winters? Yes, there have been many.
Thomas Haigh [00:09:02]:
Well, that's actually something I claim in the book, that we've only had the one. Maybe we'll talk about that as well. Okay, yeah, let's talk about that. Early on, AI was defined essentially as getting computers to do some kinds of things that currently only humans can do. So they didn't try and define exactly what was meant by intelligence. And from some viewpoints computers have always been superhuman. Even back then they could carry out numerical calculations far quicker than any human. So there was never a precise kind of definition.
Thomas Haigh [00:09:38]:
But the original idea was that with clever programming we can get computers to do a whole bunch of things that they can't currently do that in humans we associate with intelligence. And famously the initial proponents of AI in the early 60s through to about 1970 made some extremely aggressive predictions that within 10 years or so, some of them even shorter time spans, computers would basically be able to do anything that humans could in terms of intellectual work.
Paris Martineau [00:10:10]:
Now where have I heard that sort of over promising before?
Thomas Haigh [00:10:14]:
So I start that by talking about the modern day hype and then going back and saying there's a real continuity in some of the promises that were made. That's a connection back to the beginning, the specific technologies. At the very beginning they were interested in neural net type approaches and in symbolic manipulation approaches. So this is back in the late 50s, into the early 60s. By the end of the 60s, as AI matured as a field within computer science, associated particularly with labs at Stanford, MIT and Carnegie Mellon, they redefined the scope of what was included in it to push out neural networks. Neural network development continued, but it didn't continue under the umbrella of the AI brand. It was called pattern recognition, then it was called machine learning and then deep learning. So that I think is a pretty good example of how there's not stability over time in terms of the specific technologies and approaches that people have meant when they say AI.
Leo Laporte [00:11:19]:
Yeah, in fact it was that decision to follow symbolic AI versus neural networks that kind of did lead to those AI winters or single winter. By the way, I was reading in your book about the general problem solver and I had to pull out Peter Norvig's 1992 book about artificial intelligence programming in which in one of the very first chapters we write the gps, the general problem solver. And it's incredibly primitive compared to, but it's a symbolic AI. I guess the best way to describe it would be kind of a bunch of if then statements. Right. It's a very logical, deterministic kind of Thing very different from the AI we know today. So you said something interesting, that there's only been one AI winter. So was it one long winter?
Thomas Haigh [00:12:12]:
Well, I mean, if you go to Wikipedia or just Google AI history, you'll probably find some website that's been set up by a company that wants to sell you something and has pulled together some information without caring enormously about the historical specifics. But the received wisdom has become that I did well in the 60s, then it did badly in the mid and late 70s. It did well again in the 80s around expert systems, and the first time that AI is really having startups and venture capital and industrial enthusiasm in the 80s. And that's absolutely true. And then it does badly in the 90s and early 2000s before it revives around neural networks. And the part of that I disagree with is I think AI was doing pretty well in the 70s outside some very specific factors around DARPA, which had been funding research, its change in mission after the Vietnam War that led to MIT and MacArthur's sale lab at Stanford not getting the same kind of very easy money that they had before. So I think there was a kind of localized frost around two or three labs in the US and Edinburgh in the uk, and those labs really dominated the field. So the people who wrote memoirs and gave speeches at AI conferences and so on were giving the experience of people who've been in a position essentially of enormous privilege previously getting large amounts of money without having to do formal proposals or worry about milestones or deliverables or peer review, and found themselves, you know, in diminished circumstances.
Thomas Haigh [00:13:59]:
But if you do, you know, the good historian thing and look a little bit more broadly at metrics like the number of members in artificial intelligence associations, which at that point in the 70 was primarily special interest group of the association for Computing Machinery. Or you look outside those elite labs, the number of people going to conferences, you look at the spread of AI internationally. I mean, this period of the late 70s when conventional wisdom says there was a major cutback in the field, it's when the national AI associations in European countries and the Soviet Union first get founded. It's the first regular AI meeting. And even at Stanford, I think it's a real split screen thing, because on the one hand, McCarthy's lab is struggling to get the same amount of funding from darpa. On the other hand, Ed Feigenbaum's teams around expert systems are doing really great. So I asked Feigenbaum, did he think the 70s were a period of retrenchment in 20? He said no. The 70s was great.
Thomas Haigh [00:15:06]:
They kept giving me more money. Everything was coming together. So I think it's an example where a very specific historical perspective from a handful of elite lab leaders and their grad students has really warped our understanding of what was going on with AI in the 70s.
Paris Martineau [00:15:22]:
I'm curious, how did you get it?
Jeff Jarvis [00:15:23]:
Go ahead.
Paris Martineau [00:15:25]:
Oh, go ahead, Jeff.
Leo Laporte [00:15:27]:
You're the one in the hospital bed. Jeff. I think normally I would let Paris.
Paris Martineau [00:15:31]:
Go, but no, you gotta. You simply have to.
Jeff Jarvis [00:15:36]:
When you zoom in from a hospital.
Paris Martineau [00:15:38]:
Bed, you get to ask the questions.
Leo Laporte [00:15:40]:
Oh, good. Oh, good, a new bag.
Paris Martineau [00:15:44]:
I'm gonna have 75 questions for you, Jeff, once we finish this interview, but go ahead. For right now, I just want to.
Jeff Jarvis [00:15:51]:
React to what Thomas said, Leo, because you've long said that there were two AI winners. And I'm curious how you first came across the idea that AI had the first winner and what your reaction is to Thomas's very good analysis.
Leo Laporte [00:16:02]:
Well, I think it's. I guess it's how you define a winner. I mean, it sounds like.
Jeff Jarvis [00:16:06]:
Well, it's how you define AI too.
Leo Laporte [00:16:08]:
Yeah, it sounds like Feigenbaum was happy because he was getting money. Does that mean it's a successful enterprise? I think when I think of the terms AI winners, I think of it and again, not having lived through it, I don't know. But I think of it that there were periods of great optimism which were followed by disappointment because the AI techniques that they were espousing, like GPS or symbolic logic, whatever, didn't really. Expert systems, didn't really deliver. Did Feigenbaum believed that it was delivering or he was just getting money?
Thomas Haigh [00:16:51]:
He still believes that expert systems work. He's got a story he likes to tell. For example, with the diagnosis of heart conditions that they. The idea with expert systems was somewhat coming out of the difficulty of building general purpose intelligence, which had not at all progressed according to the optimistic predictions for the 60s. So by the semitist, the idea was, you think, you know, experts need to be smarter, but maybe actually because they have very specific domain knowledge, it's easier to simulate a high level international expert than it is to simulate basic common sense. And Feigenbaum likes to tell stories of working with an expert in diagnosing heart conditions and asking some questions, eliciting the knowledge, expressing it in rules that would then go into the lisp based inference engine, and then running that against some test cases, seeing where it went wrong, going back to the expert saying, why didn't it reach the right condition here? And then the expert says, oh, well, what I forgot to say is in these circumstances, actually you do this other thing. And he claimed that once you got to a couple hundred rules, you pretty much could represent any kind of expert knowledge. And then they did tests where they would take the systems for medical diagnosis and other kinds of expertise, run them against test problems, show the same test problems to a panel of experts, and they would claim that the systems in those very narrow areas could outperform what the experts could do, or at least what a mid level expert could do.
Thomas Haigh [00:18:37]:
You know, so the logic would be, if you've got something can diagnose blood infections much better than the typical doctor, then that will be something that will be worth rolling out and there's a real market for it. And a lot of the 80s AI boom was very specifically an expert systems boom, because the logic for that was maybe we haven't solved the problems of general intelligence, we certainly haven't achieved superhuman intelligence, but we can make systems that are economically viable and can pay for themselves by taking expert knowledge and making it portable and putting it in a little software box. And that's also one of the reasons I said earlier that in some ways the modern discourse about AI has got more to do with the very early days discourse of rapidly achieving general purpose human intelligence versus the ideas that came in the 70s and 80s where the Russia was much more about. So let's not even talk about the Turing Test, but let's just say there are ways that we can make computers more economically valuable by encoding knowledge and using it to help them perform better.
Leo Laporte [00:19:42]:
So a different goal meant that it wasn't winter because they were achieving that particular goal of an E system.
Thomas Haigh [00:19:50]:
I mean, winters and summers have conventionally been expressed in terms of funding for AI.
Paris Martineau [00:19:57]:
Right.
Thomas Haigh [00:19:57]:
And of course that's downstream for belief in AI. So, so they go together. So the original.
Leo Laporte [00:20:03]:
So we're in a glorious summer right now, aren't we?
Paris Martineau [00:20:06]:
We're in a very endless summer, it seems. I'm curious though, why do you think that the industry is attached and has remained attached to this notion of the 70s being a first AI winter? What, what do you think explains that and what do you think it says kind of about that as a, a narrative device for the industry?
Thomas Haigh [00:20:26]:
Yeah, well, it's such a pervasive narrative. Initially I assumed it must be true and you know, then I went back and just did some fairly low hanging fruit things like looking at memberships in the associations or seeing when AI spread overseas. And then I was like, oh, the other thing I just did is a Google Ngram is a great tool for seeing how much people are talking about things. I've got some of those in the book. If you do Google Ngram for artificial intelligence, there is steady growth in the 70s and then a real peak in the 80s. And you do definitely see the real AI winter of the 90s in the Google Ngram and in participation at conferences and in those kinds of metrics. You don't see any kind of broad based drop off for AI in the 70s. The other thing, the phrase AI winter, you can date very precisely.
Leo Laporte [00:21:19]:
We have the N gram, by the way, Benito, if you pull that up.
Thomas Haigh [00:21:21]:
Yeah. To a 1984 panel at the American association for the Advancement of AI. And they're worried. Basically the 84 is real boom times. They're getting lots of money. People are being lured away from completing their PhDs by industry jobs. In many ways, everything is going great. The conference is feeling more like a trade show than an academic venue.
Thomas Haigh [00:21:44]:
But some of them are saying, I don't trust this. I think maybe we're overhyping it. This is all going to end in tears at that point. People are worried about a nuclear winter. So the idea of AI Winter, there's fallout, the sun gets blocked out, everything dies. Now, of course, you might say losing funding for AI is not quite the same as everything dying, but that was the analogy. If you look at the quote that they have there about what's going to happen, it pretty much precisely defines what happens four or five years later. Companies cut their AI groups.
Thomas Haigh [00:22:21]:
The government, which had been funding something called the Strategic Computer Initiative, cuts back. Autonomous vehicles fail to roll. At the end of that squirt, there's a line I really like about everybody stops calling whatever they're doing AI and finds a different name for it, which pretty much is what happened. But the funny thing is, at that 1984 thing, you can read the transcript of the panel discussion and it's something like 10 pages, single spaced. Nobody says, oh, this thing we're talking about, hypothetically. We just had one of those a couple of years ago. Right. So there's no sense in 1984 that there'd been an AI winter just a few years earlier.
Thomas Haigh [00:23:02]:
Where it seems to come from is a 1990s book by. What's that? Creviet, who had been trained in AI at MIT. And he's basically reporting the folk wisdom of MIT that the late 70s was a hard time when people had more difficulty getting money. And that is blown up into this claim that there was an enormous international, broadly based cutback in AI research in the 70s. So I think it comes again to. There was a pretty much a cartel of labs in the 70s into the 80s that got to set the entire agenda for AI, and that was Stanford, MIT, Carnegie Mellon and SRI, which was at that point largely detached from Stanford. And so the people who, I mean, I looked, I found seven AI textbooks from the 70s. They had eight authors.
Thomas Haigh [00:24:03]:
All eight of those people had a PhD from one of those places. So they really were the people who were being invited to give keynote speeches, who were writing memoirs, whose recollections were passed down to their grad students. And those people were in those specific places where AI funding had been extremely easy and lavish in the 60s and was less so in the 70s. And I think just the folk wisdom of AI has been turned into this historical claim of a broad based slowdown without anyone actually attempting to look for evidence of whether it's true or not.
Leo Laporte [00:24:39]:
How much of it. War as well. I mean, this is a era where we were in a, you know, battle with the Soviet Union for military supremacy. And I think that there was some. You mentioned arpa, you know, you mentioned the nuclear winter. Is some of this informed by the Cold War?
Thomas Haigh [00:25:02]:
All of it, really? I mean, why? The period where AI is getting going in the 50s is an incredible period for the growth of science. And it's a wonderful time to be like a smart, young, geeky, science oriented guy because there are institutions like the RAND Corporation, the incredible amount of federal funding that is flown to MIT and Stanford. I mean, it's not a coincidence that AI develops primarily at MIT and Stanford because those two institutions are far ahead of anywhere else in terms of the amount of federal money they have simultaneously.
Leo Laporte [00:25:42]:
With the Internet as well. Right. I mean, this is sort of the Internet development as well, right?
Thomas Haigh [00:25:46]:
Yeah. The same ARPA office that is funding AI in the 60s at Stanford and MIT in particular is the office that is funding the ARPANET.
Leo Laporte [00:25:57]:
Yeah.
Thomas Haigh [00:25:58]:
So it's part of the same vision for interactive computing as something that can do revolutionary things.
Leo Laporte [00:26:07]:
One of the reasons people think. I'm sorry, Jeff, go ahead. There's a little lag.
Jeff Jarvis [00:26:11]:
What did ARPA want out of AI at the time?
Thomas Haigh [00:26:14]:
Well, ARPA didn't fund a specific AI program until the 70s when it had a large project on speech understanding. So in the 60s there wasn't a separate AI pool of money existing at Arpa, so it was bundled in with other things. So at MIT there was a giant thing called Project Mac. And depending on who you asked Mac could mean man and computer or machine aided cognition. And that went with the vision of J.C.R. licklider, the inaugural director of that piece of ARPA who had a vision of computer human symbiosis. So it was more like the idea of an interactive tool that can make humans smarter was the actual driving vision versus specifically the AI dream of producing intelligence that was autonomous and existed aside from from humans and could do its own things in the world. But time sharing was a big piece of that because previously computers had worked on a batch processing basis.
Thomas Haigh [00:27:25]:
You would give your programming on a piece of paper, get punched onto cards, they'd run through the machine, you'd get the results back, which usually would be error messages saying you made a mistake in the code, maybe a day later. AI type visions and this idea of the computer being an interactive tool both depended on finding a technology that could let the you to respond to you instantly so you could have an interactive dialogue with it. So John McCarthy, who came up with the term AI and also founded the lab at Stanford, was previously at MIT and he was the strongest proponent for this idea of time sharing which MIT pioneered for making computer acts access interactive. So in the 60s ARPA was funding this bundle of things that included graphics, time sharing, networking, the provision of computer facilities really out of a general sense that computer would be much more powerful if they could be interactive tools versus purely batch process things that people didn't interact with directly. And they were absolutely right about that. Even if the specific AI pieces of that agenda didn't deliver on what people like Marvin Minsk at my MIT and John McCarthy hoped they would.
Leo Laporte [00:28:46]:
We're talking with Thomas Hagee's, the author of Artificial Intelligence the History of a Brand. This is the pyramid of the Strategic Computing Initiative. Roundabout that time, I think it's a little bit into the 70s of the plan, right?
Thomas Haigh [00:29:02]:
Yeah. So that's the 80s 80s really, when AI funding gets generous again in the minds of the people at MIT and Stanford. It's also, I mean it's the era of Reagan, it's the era of the Strategic Defense Initiative. So there's a general interest in spending money in ways that will improve US national security. With the revival of Cold War after the detente era in the late 70s. And the pitch there is basically around what I mentioned previously with expert systems. So Edward Feigenbaum at Stanford, who came up with the idea of expert systems, also was very effective in helping to scare American politicians about the danger of Japan getting ahead in expert systems and AI. So Japan had Something called the fifth Generation Initiative.
Thomas Haigh [00:29:58]:
And that was used in the US and in Europe to convince governments that they needed to fund AI and expert systems to avoid reminiscent period where people had seen one industry after another crumble in the face of Japanese competition. And they were worried that the Japanese were coming for American strengths in computing, which obviously was a scary thing and could leapfrog ahead to the next generation of intelligent computers unless Congress was prepared to put a bunch of money into AI and expert systems. But the point with that pyramid is the argument was we already know how to make an intelligent computer, but the problem is we can't fire off enough rules every second to achieve intelligence. So we don't just need money for expert systems, we need money for parallel computing. This was the area was just talking to some people in Germany about an exhibition they're doing about the connection machines built by the MIT affiliated firm Thinking Machines. So in that era it made a lot of sense to brand a supercomputing company as being about intelligence because DARPA had a lot of money to spend on this cluster of intelligent machines and enabling technologies. So the government spent in a big way on improving microelectronic chip manufacturing type technologies. They spent on parallel computing so that you could get more rules fired off every second they spent on expert systems.
Thomas Haigh [00:31:29]:
And the way that they spent a lot on autonomous vehicles. That is really where Carnegie Mellon built up its strength in autonomous vehicles, which feeds through to the modern day.
Leo Laporte [00:31:41]:
Remember the Grand DARPA Challenge? Yeah.
Thomas Haigh [00:31:44]:
And what they found really was the architecture and underlying tech side of that went pretty well. But unfortunately, even when you scaled up the computing power available, the AI technologies just didn't do what people were hoping. You quote. So they fell by the website's roadside somewhat.
Leo Laporte [00:32:07]:
You quote. Literally in the case of the DARPA Grand Challenge, you quote YALE Professor Drew McDermott at a panel called the Dark Ages of AI. This is around 1984, warning of a feeling of deep unease that excessively high expectations for AI. See if this sounds familiar. Will eventually result in disaster. To sketch a worst case scenario, he said, suppose that five years from now, the Strategic Computing Initiative collapses miserably as autonomous vehicles fail to roll. The fifth generation turns out not to go anywhere. The Japanese government immediately gets out of computing.
Leo Laporte [00:32:43]:
Every startup company fails. Texas Instru and Schlumberger and all the other companies lose interest. We've been talking about an AI bust for a long time.
Thomas Haigh [00:32:53]:
And then right, the line about everyone finds a different name for whatever is that they're doing right in the boom. Time everyone who did anything that could plausibly be called AI would call it AI. And the scope of AI grew a lot. And then in the AI winter, I.
Leo Laporte [00:33:10]:
Mean, this is, here's the Engram where they changed the name to Expert Systems.
Paris Martineau [00:33:18]:
Just a little control replacement.
Thomas Haigh [00:33:20]:
And there's an interesting relationship there from some viewpoints. For some people, they would think of expert Systems as one approach within AI, but it also in some ways function as a rival brand because AI by the 80s already had this taint of having over promised and under delivered for a long time. And Expert Systems sounded more technical and.
Leo Laporte [00:33:41]:
Respectable and maybe less ambitious to some degree.
Jeff Jarvis [00:33:45]:
Right.
Leo Laporte [00:33:46]:
We're not trying to build a consciousness, just we want to answer some questions.
Thomas Haigh [00:33:50]:
Right. And you see this in many areas, including for example, if you look at AI textbooks by the 80s, they are not discussing the Turing Test. They are not making the claim that all this is about achieving human or superhuman intelligence. They're making the claim that this is a respectable body of techniques that work, that rely on knowledge, that make computers go more effectively.
Leo Laporte [00:34:13]:
Well, there's so many great characters in this, including Marvin Minsky and some other histories of AI that we've talked about. On this show, Marvin Minsky is painted a little bit as a villain, as the guy who was so convinced that neural networks couldn't possibly work that he steered AI away from what was ultimately the technique we're using today. Do you see him as the villain in this?
Thomas Haigh [00:34:40]:
I try and get some kind of historical distance there, although one of the comments from the reviewers for the press was that I'm taking my animus against modern day big tech and projecting it back to be too harsh on those guys in the past. I don't think I am. I am in many ways stressing that what they were doing is something quite different from what happened now. I think you have to deal historically with the fact that, that none of this worked. And I was trained in the early 90s as a computer science student. I took maybe five AI courses. I learned these techniques in the AI courses the same way I learned a bridge solver.
Leo Laporte [00:35:18]:
Right, in Prologue.
Thomas Haigh [00:35:19]:
Yeah, yeah, the same way. I did graphics and I did databases and I did architecture. And you write these simple exercises with these toy problems and you do the same thing in AI. But the difference is the techniques I learned in the AI classes couldn't scale up. They only ever worked on these incredibly small toy problems. And the techniques in the other classes were simplified versions of the techniques that really did work for technologies that existed in the world. So when you're talking about 20th century symbolic AI, I think you need to be clear up front. The things they were doing didn't work.
Thomas Haigh [00:35:57]:
They produced all kinds of byproducts in terms of interactive computing and technologies and parallel computing. So I'm not saying that the money was a bad investment, but it was the byproducts that went out in the world to be useful. They did not succeed in achieving their core goals. I think that is clear enough. At this point, you can get away a bit from thinking, well, this guy was a hero because he favored this good approach that I think the field should have adopted, and this other guy is a villain because he did this. I mean, I mean, nothing any of them tried to do would have worked. And there are many reasons for that. I mean, there's the whole problem with the fundamental difficulty of taking a purely symbolic approach.
Thomas Haigh [00:36:48]:
There's tacit knowledge, all the things that the skeptics and the philosophers were complaining about all along, but there's also just the complete lack of computing power that's available.
Leo Laporte [00:36:58]:
That's, that's kind of the underlying thing of all of this. They were trying to do parallel computing and time sharing and it's all been solved. That was the bitter lesson, wasn't it? It's all been solved by just massive compute.
Thomas Haigh [00:37:11]:
Well, I mean, they did other things too. But if you look at the revival of neural nets, the place where neural nets were bought back and worked was Bell Labs, Yann Lecun in particular, with a system that was able to differentiate pretty reliably between the 10 dig and read zip codes and in a related application to read the routing numbers on checks that were written by hand. So by the 80s, and this wasn't just computing power. I mean, they also. The whole thing, Hinton and his buddies and the back propagation algorithm and various conceptual advances that underpinned the revival of neural Networks in the 80s. But with the biggest computers that were available in the 80s, reliably distinguishing between 10 digits was pretty much all you could get. So it's not that there was some other path that could have been taken in the 60s, 70s, 80s to make AI work. I think just fundamentally you have to accept that they had some ideas that maybe seems reasonable.
Thomas Haigh [00:38:22]:
I don't want to say that it was ridiculous to even explore them. They certainly, for example, showed that the forms of logic that you might learn in a philosophy class are going to help you solve problems and think more rationally. Just fundamentally don't work when you're trying to apply them on A large scale, and they had all kinds of advantages and byproducts that were produced along the way. But I think, again, we just got enough historical distance now that I don't want to say, say Minsky was bad and McCarthy was good, or McCarthy was good and Minsky was bad. They were really smart people that had a bunch of interesting ideas to try and address. A challenge that it's clear in hindsight was fundamentally insurmountable, given the technology available, the capabilities.
Leo Laporte [00:39:15]:
You do say that Minsky confessed to being the devil who killed interest in neural networks for a generation. But Seymour Papert was the guy who really pushed this notion. I like your more nuanced view that it isn't this kind of. We took a wrong turn and it went downhill. And then we finally, oh, it's machine learning. We got the right. It is a continuum. And it's a continuum that's informed by a lot of different ideas and capabilities that got us to where we are today.
Leo Laporte [00:39:45]:
Now, you said something interesting. You're not a fan of big tech today?
Thomas Haigh [00:39:51]:
Yeah, I mean, that's in the opening. I don't think that puts me in a minority.
Leo Laporte [00:39:57]:
No, it does not.
Thomas Haigh [00:40:00]:
So I start out in the introduction with a snapshot of the world in 2024 and the AI hype there. And then at the end, I've just been drafting some extra material to put into the book with the revisions. So I'm trying to cover end certification and the way that the enormous cost and resource needs of these systems mean that there's only a handful of big tech firms that can afford to develop, which are underpinned by monopoly profits from doing all the bad things that you're all extremely familiar with. So I'm not a booster of modern AI or big tech, but I don't have any particularly original critiques for that. I think the originality that I have in the book is taking 20th century AI seriously and going pretty thoroughly through time about what the AI brand has meant over the decades since it was originated. So the last two chapters are the neural network revival, which I, you know, mostly what I know about that, to be honest, comes from podcasts and journalists and things like that. I'm not trying to explain to the world how ChatGPT works and so on, but I'm trying to know enough about that to say what is the same and what is different between these modern day AI branded technologies and the ones that were dominant in the 20th century.
Leo Laporte [00:41:24]:
You mentioned that your book.
Paris Martineau [00:41:25]:
I think Jeff's trying. Yeah, I wish the question we should let Jeff.
Thomas Haigh [00:41:29]:
I want to hit Jeff. Sorry, Leo.
Jeff Jarvis [00:41:34]:
Right off. What you just said did. It's a weird general question I'm gonna ask is two parts. Do you think in the end the people involved in AI over the years and now would say that the use of the brand AI was beneficial or detrimental.
Thomas Haigh [00:41:53]:
To their careers.
Jeff Jarvis [00:41:56]:
To the field, to their careers, to the society, whatever you like.
Thomas Haigh [00:42:04]:
It's an interesting question. I mean, Feigenbaum, for example, deliberately, I think, came up with the expert systems brand as an alternative to AI. Minsky and McCarthy, I think through their whole lives remain true believers, particularly McCarthy, I think, who came to see himself as something of a dinosaur. I just was reading some archived emails from him that are preserved at a website called saildart, which is taken from the backup tapes of the Stanford AI Lab timesharing system. So it's a wonderful opportunity to be a bit of a fly on the wall. And it's clear that by the late 70s he saw himself as working on basic AI and logic, which had fallen from favor for applications oriented work, and pretty much just in his mind plugging away doing the same kinds of things he'd been doing since the 50s. If you want to know what those people thought, the best source is probably the 50th anniversary reunion conference that they had at Dartmouth. And in those days they're basically saying, yes, we still believe it's going to pay off in the end.
Thomas Haigh [00:43:17]:
We still believe in symbolic AI. We think it's going to be a great thing, but it's not going to happen in our lifetimes. The thing that we thought was 10 years away is maybe 50 years away. So the ideas for where what we now call AGI would be accomplished. That was probably 2000, I think that was 2008, the reunion around that time. That's probably a low point for belief in the near term potential for human like intelligence.
Jeff Jarvis [00:43:45]:
Let me ask you a follow up, if I may. Real quickly there on the new Books network where I heard you talking about your next book, which is a wonderful wonky podcast network for academic books, you said that you could see a case for people wanting to flee the brand of AI and try to rename it again in the coming year or so, you think that the language is, foretells the fate or reputation of the field of problems in it. How does, how do you see that going?
Thomas Haigh [00:44:21]:
All right, so the context for this is the technologies that we now call AI are fundamentally around neural networks, mostly around generative AI, within that, mostly around large language models and those technologies as they were being developed in the early 2000s, were not called AI because they'd been pushed out of the AI field. And I think also to an extent, because they didn't want the baggage always that came with the AI brand. Those things were called machine learning, deep learning, pattern recognition. And something really dramatic has happened in the last 10, 12 years that essentially the people who'd been taking this other technological approach that existed outside AI as a brand, it existed outside the elite AI labs basically stormed the castle, raised their flag and said, we're AI. Now when you say AI, this is what you mean, not the symbolic AI. A question is why did the machine learning community seize the AI brand for itself after having kept a distance from it for many years? And I think the argument to that is fundamentally around the science fiction narrative associations of AI. So they reclaimed the brand. I've got a quote in the book from an AI researcher who says the exact day on which he became an AI researcher and not a machine learning researcher was when they went to the NIPS conference for the neural net stuff.
Thomas Haigh [00:46:00]:
And Mark Zuckerberg was there in the presidential suite having recently hired Yann Lecun and was offering people lots of money to come to his group at Facebook which was called AI. And then obviously the DeepMind people, I mean, Shane Legg really heavily promoted the concept of AGI and OpenAI DeepMind. Those other guys really had a revival of this early AI sense that this was something that was going to produce human and then superhuman intelligence with the whole singularity thing in the near future. And that is also why so many trillions of dollars have been flowing into it. So I think they switched the name of the thing from the more technical wonkish machine learning to the attention grabbing AI very specifically because they wanted to make a number of claims that seem plausible to us because we've been conditioned by science fiction. So one of those is that AI is going to be this superhuman general purpose thing. Another one is that it's something that is going to happen really quickly. So if you look at science fiction stories like say Heinlein's Moon is a Harsh Mistress or the Terminator stories, AI just happens, right? One day a computer gets big enough and it suddenly becomes self aware.
Thomas Haigh [00:47:31]:
And that obviously happens because the authors had no idea how you could make a self aware computer. So they're just, well, whatever, it just happened, okay, deal with it. So we're also primed to believe that AI is that can just happen very quickly and unexpectedly without actually needing much work. Once you've got a big enough tech platform. And also, of course, in most stories where AI exists, it's the most important thing in the world. And you get the whole Duma versus accelerationist thing because in some AI stories it's basically a metaphor for slavery. And of course the robots exist purely in order to be oppressed and rebel. And in other stories it's basically the Pinocchio story of the boy that wants to become real, etc.
Thomas Haigh [00:48:15]:
But I think it's exactly the science fiction promises that are implicit in the brand of AI, which began in science, but during the AI winter, really survived much more strongly in science fiction than it did in computer science that led to them reappropriating the brand. So back to the prediction that I made. It seems that there is, there is pretty much no margin for error in the promises that have been made for AI. Every leader of every major AI company has got a personal timeline for achieving AGI and the promise that the singularity is going to happen sometime soon after that and we're all going to techno heaven or techno hell. But either way, the rapture is coming. There's not much margin for error in that. And it seems to be much more like that AI is going to be like other technologies, you know, it's going to be. It's producing tangible tools that do some things really well.
Thomas Haigh [00:49:20]:
And I don't think those tools are going away. But the question is, is bundling together all these different technologies like image recognition and text generation and autonomous vehicles into this one thing called AI and making these future premises for it going to work. And I think if the overall superhuman intelligence thing doesn't pan out, or even if the technology works, but the business bubble bursts because historically, even with something like the Internet, as you're aware, the Internet wasn't a flash in the pan thing. The Internet was really important, but the dot com stock bubble still burst. So even in a scenario like that, I think the brand will become tainted again like it did in the AI winter of the 90s. And in the 90s, for example, if the 90s is the period where continuous speech recognition really becomes an important thing. So you've got technologies like Dragon, naturally speaking, and they don't call those things AI, they call them speech recognition, because AI is out of fashion, but speech recognition is in. And very recently, actually, the company that bought.
Thomas Haigh [00:50:37]:
The company that bought, the company that bought Dragon was bought by Microsoft. It was nuanced and it was rebranded back to being AI as Microsoft's big play in AI for healthcare but speech recognition spent like 30 years not being AI so I think. I think we may get something similar that people come up with these much more specific brands for the technologies that actually work. But the benefits of being associated with this big science fiction narrative around AI fade when people are like, oh, it's been three years. Where's the stuff you promised us would be three years away? The stock market bubble bursts. Nvidia is no longer the world's biggest company, et cetera, et cetera. And people try and distance themselves from that and go back to the idea of having much more specific brands associated with the text that work. I mean, brand wise.
Thomas Haigh [00:51:32]:
I kind of like to say it works a bit like a fashion brand, like Chanel. Right. So Chanel makes really good fragrance and they have good lipsticks, they have the Runway stuff and then they brand extend into handbags and watches and so on. And it's not like the things that are good about the Chanel watch are the same things that are good about the number five fragrance. But the brand like unites these disparate things and gives you this sense that they all are sharing qualities with each other in some more intangible kind of way.
Leo Laporte [00:52:11]:
We've been talking to Thomas Haig. He's professor and chair of the History Department at the University of of Wisconsin, Milwaukee. Now, I'm confused about the title of the book. The galley that I have is different. Have you decided on a title yet?
Thomas Haigh [00:52:25]:
Yeah, so that's with the press. So, yeah, the one you've got says Artificial Intelligence Colon, the History of a Brand, which is a good title, except there's going to be a million books called Artificial Intelligence Colon. And if you only search on the main title, no one will find it. So we're thinking to use a title that I've been using for talks that I've given, which is the Brand that Wouldn't Die, A History of Artificial Intelligence. And then you've got a main title you can actually find.
Leo Laporte [00:52:54]:
Brilliant. It'll be from the MIT Press anyway, and it's sometime soon. Yes or no?
Thomas Haigh [00:52:59]:
Well, hopefully it took them a while.
Paris Martineau [00:53:01]:
Got to figure out that name first.
Jeff Jarvis [00:53:04]:
Defined soon in academic publishing.
Thomas Haigh [00:53:07]:
Well, I mean, I still like to hope for by the end of 2026. That would depend on the press being willing to move a little bit faster than usual. But they do want to give it some trade distribution and you relatively high profile.
Leo Laporte [00:53:21]:
So it's tricky because AI is moving so fast. It's hard to write that final chapter.
Thomas Haigh [00:53:28]:
Yes.
Jeff Jarvis [00:53:28]:
The advantage of history, though, Leo.
Leo Laporte [00:53:31]:
Well, okay, yeah.
Thomas Haigh [00:53:33]:
And the, the earlier chapters, those are written in stone. They're going to stay way more current. You know, it's always a problem trying to come up to the present, but obviously if I give someone a book on AI and it doesn't get to ChatGPT, they're probably going to want their money back. So the current stuff needs to be there. But mostly just so that I can draw parallels between the earlier history versus claiming that I have the definitive, unique understanding of the modern day tech.
Leo Laporte [00:54:01]:
Well, and it's just fascinating story and it's so many interesting characters and you draw some nice pictures. I didn't realize that Marvin Minsky was a comedian as well as a brilliant thinker, things like that. Thomas Haig, thank you so much for joining us. His website, tomandmaria.com Tom has a lot of great stuff on it, including references to his earlier books which were really important in the field. Still are. ENIAC in Action is the definitive history of one of the very first modern computers. He's also the co author with Paul Ceruzi of the New History of Modern Computing, which is also a modern classic. So if you're interested, I know most of you are in this stuff, this is a great place to start.
Leo Laporte [00:54:48]:
Yeah.
Jeff Jarvis [00:54:48]:
What do you look at next, Thomas? What's the next pursuit?
Thomas Haigh [00:54:52]:
Oh, well, after the AI book is out, I want to get back to actually what my dissertation was about, which is the history of computers in corporate management throughout the 20th century. So it's, you know, in a way the prehistory of big data.
Leo Laporte [00:55:13]:
Yeah. I think about the episode in Mad Men when this Madison Avenue agency in the early 60s gets their first computer and it's on its own floor and they have the priesthood and they're able to do things with that computer that you could probably do in about a 60th of a second on your phone today. But for them it was a big revolution. Yeah, I think that's fascinating. And we've come so far so fast. I can't think of another technology that has made this kind of advancement in just a matter of a few decades, which also makes it quite interesting and the cultural impact of it, which we're really seeing with AI. Thomas, thank you so much for joining us. I appreciate it.
Thomas Haigh [00:55:53]:
Thank you.
Jeff Jarvis [00:55:54]:
Thomas, thank you. Great to meet you.
Thomas Haigh [00:55:56]:
You too. And I feel honored that you have made it from your hospital bed.
Jeff Jarvis [00:56:01]:
That is.
Paris Martineau [00:56:02]:
This is really an interview for the A.
Leo Laporte [00:56:07]:
Thank you, Thomas. Have a great day.
Jeff Jarvis [00:56:08]:
Thank you. Take care.
Leo Laporte [00:56:09]:
We'll continue with Intelligent Machines in a moment.
Paris Martineau [00:56:13]:
I need to ask so many questions about this.
Leo Laporte [00:56:16]:
Oh, man. We'll get to the first.
Paris Martineau [00:56:18]:
Or are we?
Leo Laporte [00:56:18]:
Yeah, let me do an ad. We're a little behind because.
Paris Martineau [00:56:21]:
All right, Jeff, you gotta hang on. You gotta hang with us for five more minutes.
Leo Laporte [00:56:25]:
Can you make it another? You know, I think I can make.
Jeff Jarvis [00:56:28]:
It an hour, but that'll probably be better.
Leo Laporte [00:56:30]:
Oh, we'll wrap it up in an hour. That's perfect. All right, we'll continue with intelligent machines in a moment. Our show today, brought to you by Bit Warden, the leader in passwords, pass keys and secret management. I tell you what, I couldn't live without. Bitwarden. Bitwarden is consistently ranked number one in user satisfaction by G2 and software reviews. Over 10 million users across 180 countries, more than 50,000 businesses.
Leo Laporte [00:56:55]:
I think, you know, you should know if you watch this show or any of our shows that having a password manager manager is kind of fundamental to your personal security. You can't possibly remember, you know, hundreds of passwords or store hundreds of pass keys. You need something that'll do it, something that'll do it with strong encryption, ideally open source encryption, so you can verify it's doing it right. You need something because otherwise you're going to come up with easy passwords that you can remember, probably going to reuse passwords. Now you're smart enough not to do that, but I think many of if you run running a business, many of your employees are probably doing those things you know better than to do, like reusing passwords. That's why you got to have Bit Warden. Bit Warden makes it easy for you to encourage your employees to do the right thing. They will love Bit Warden and Bitwarden will help you.
Leo Laporte [00:57:51]:
For instance, their new access intelligence for businesses. Organizations can use Bitwarden to detect weak, reused or exposed credentials, immediately guide remediation. This is something Bitwarden does really well, even with individuals. I just was on a site and said, you know what you need to replace that password. It helps you replace risky passwords with strong unique ones, which closes probably the number one security gap in business today. Credentials still probably the top cause of breaches. But with access intelligence from Bit Warden, they become visible, they become prioritized and they are corrected before exploitation can occur. Bitwarden though, something for everybody.
Leo Laporte [00:58:33]:
If you are, and I know many of you are doing your home lab, you have a personal projects you might be interested in. Bitwarden Lite. Bitwarden Lite is a lightweight, self hosted password manager built for those home labs. Those Personal projects for environments that want quick setup with minimal overhead. Maybe that's you. Bitwarden, of course, no matter where you're using it, at home, at work is enhanced with real time vault health alerts and those password coaching features that help users identify weak, reused or exposed credentials and take immediate action to strengthen their security. One of the ways Bitwarden helps you is by helping you move off of the password manager built in your browser. It's so easy, right? And oftentimes, I think especially less sophisticated users, they turn on the password manager in their browser and they kind of live there.
Leo Laporte [00:59:20]:
But it's not the most convenient and it's not the most secure or good news. When you install Bitwarden, it'll say, ah, you want me to move your passwords from Chrome, Edge, Brave, Opera, Vivaldi? I'll do it directly. It directly imports the credentials from the browser into the encrypted vault without that separate intermediate plain text export, which is much more secure. It helps reduce exposure. It's a really brilliant idea. I wish Everybody did this. G2 winner 2025 reports that Bitwarden continues to hold strong as number one in every enterprise category for six straight quarters. It's easy to move to Bitwarden.
Leo Laporte [01:00:00]:
I did it. Steve Gibson did it a few years ago. Couldn't have been simpler. Within a few minutes, we're up and running with Bitwarden. I've never looked back. It is the way to protect yourself. And because it's open source, GPL licensed, you can review the code yourself if you want. It's on GitHub.
Leo Laporte [01:00:19]:
It's also regularly audited by third party experts and they meet SOC2 type 2 GDPR, HIPAA CCPA standards. They're ISO 2700-12002 certified. So it's absolutely secure. Get started today with Bitwarden's free trial of a teams or enterprise plan. Get started for free across all devices as an individual user. Bitwarden.com TWIT bitwarden.com TWIT I couldn't recommend it more highly. It's. Everybody should be using Bitwarden.
Leo Laporte [01:00:49]:
I hope you will. And make sure you use that address so they know you saw it here. Bitwarden.com TWIT thank you, Bit warden. Okay, okay. You can ask all those questions now of Mr. Jarvis. Don't tire him out. Oh, that's cool.
Leo Laporte [01:01:07]:
Now they give you a little readout, okay?
Paris Martineau [01:01:09]:
Some context for anybody who's just listening instead of looking. Jeff is in a hospital bed showing us his heart Rate. I will also say something about the lighting in the hospital bed. I assume it's not set for podcasting. Something about it makes it look kind of AI generated.
Leo Laporte [01:01:28]:
No. I thought he was in heaven for the first time.
Paris Martineau [01:01:30]:
I was about to say now the lights are off, and it does look like you're on the verge of death. Before, it was the sort of unnatural lighting that it does.
Leo Laporte [01:01:39]:
Look at that.
Paris Martineau [01:01:39]:
It feels heavenly.
Thomas Haigh [01:01:41]:
He's in a cloud or two.
Jeff Jarvis [01:01:44]:
I think one or two. This is one.
Paris Martineau [01:01:49]:
One. I'm saying we like Jeff.
Leo Laporte [01:01:52]:
We get all our light mode now.
Paris Martineau [01:01:54]:
Have you spoken to any of the employees of the hospital about the fact that you're podcasting right now? Because I'd like to know from them how many people they've seen.
Leo Laporte [01:02:02]:
I think your doctor would tell you to stop. I really do.
Jeff Jarvis [01:02:07]:
I told the wonderful nurse. So I don't know if you can hear this beeping in the background. No, it's my. Oh, you can't. Oh, good. It's. It's my iv, so they're. They're pumping tons of antibiotics into me.
Leo Laporte [01:02:19]:
Oh, gosh.
Jeff Jarvis [01:02:20]:
And the. The device is very persnickety. If it gets one bubble in line, and no, a bubble won't kill you, but it still. It stops and then it beeps, and then the poor staff has to come in. So I explained to her to flick it. Ruby on a podcast, she looked like, okay.
Paris Martineau [01:02:33]:
Oh, she didn't. She didn't blink. Did her response suggest to you that this has happened before?
Leo Laporte [01:02:39]:
I think so. I think podcasts are everywhere.
Paris Martineau [01:02:42]:
I thought the other week when we had somebody podcasting in from the airport in Las Vegas, that that was strange, but.
Leo Laporte [01:02:50]:
Oh, we're just beginning. Paris.
Jeff Jarvis [01:02:52]:
Yeah.
Leo Laporte [01:02:54]:
Don't you go in the hospital? Yeah.
Jeff Jarvis [01:02:57]:
Well, how long have you been in the hospital? I want to give compliments to Morristown Hospital in New Jersey. And God bless the science.
Paris Martineau [01:03:06]:
God bless the science.
Jeff Jarvis [01:03:07]:
And I thought that my pain in my lumbar fracture got drastically worse. And as of Saturday, I could not get out of bed. I could not move. It was awful. And I was also getting a fever. And so the presumption was these things were connected. And so the spinal duct doctor was involved in all that, was trying to look at all the clues of what could be. But he said, I don't see people getting a lumbar fracture and getting an infection.
Jeff Jarvis [01:03:37]:
27 years, I've never seen that. Doesn't happen. The infectious disease doctor, who's also wonderful, said, well, yeah, we could, but, yeah, it's a Little weird. But then went through this huge amount of testing, CT scan, obviously, X ray, mri, and there was this wonderful little moment of science where they did a blood culture to see what's growing in my blood. And the, this, the infectious diseases doctor came in the other morning and he said, it's growing, it's growing.
Paris Martineau [01:04:14]:
That's not what you want to hear from an infectious disease expert.
Jeff Jarvis [01:04:17]:
But he said, you know, we know what it is. It's a staph infection. So they know how to treat it. They know what happens happen. But, but the amazing thing is that if I had not gotten the injury, I probably would not have gone to urgent care and interned the hospital as soon as I did. And this could have, well within half day or a couple days have gone into something far, far, far worse and fatal. So in that sense, I'm lucky.
Leo Laporte [01:04:43]:
Yeah.
Jeff Jarvis [01:04:44]:
But, yeah, so it's all I acknowledge. This is weird. People are defended by the sight of my, my decolletage. By the way, I think this is.
Leo Laporte [01:04:54]:
The first time you've been on the.
Paris Martineau [01:04:55]:
Podcast and not wear Ingall black.
Jeff Jarvis [01:04:58]:
I know. Yes.
Leo Laporte [01:05:00]:
Did you ask them, do you have anything in black?
Jeff Jarvis [01:05:03]:
This is, this is a standard issue gown. You don't want to walk.
Leo Laporte [01:05:07]:
I believe I know exactly what it looks like from the back. You know, you're actually in a. You're in an auspicious place. I remember, as you know, Paris boys really like history. And I remember from watching, I remember from watching Ken Burns Revolutionary War documentary that Morristown was where George Washington quartered his troops during that cold, cold winter. And it is where he got them inoculated for smallpox because smallpox was devastating the Continental army. And they didn't want to do it at first because it's a live inoculation and you get very sick, but then if you recover, you don't get smallpox. So you are in a good.
Leo Laporte [01:05:52]:
You're in a place that's been a medical center for 300 years.
Jeff Jarvis [01:05:55]:
Indeed.
Leo Laporte [01:05:57]:
Yeah. When you said Morristown, I went, oh, yeah. Oh, yeah, Morristown, Yeah.
Jeff Jarvis [01:06:03]:
Between Princeton and Trenton, mile away, between 10th, a mile away from the Washington headquarters. You know, one of many Washington headquarters in the country.
Leo Laporte [01:06:11]:
He slept there. He actually slept in your room, I think. But that's.
Thomas Haigh [01:06:14]:
Yeah.
Paris Martineau [01:06:16]:
Hollow ground.
Leo Laporte [01:06:19]:
So have you been following.
Paris Martineau [01:06:21]:
You were saying, what have you.
Leo Laporte [01:06:22]:
I was going to start with, I know what you're going to say. Maybe we should talk about AI.
Paris Martineau [01:06:26]:
I was like, I guess we should talk about the news. It feels a little inappropriate. Well, Jeff, you can leave. Leo and I can do this, we.
Leo Laporte [01:06:34]:
Can handle this alone. No, no, we. I don't want him to leave.
Paris Martineau [01:06:37]:
I mean, if you want to be here, you are welcome to.
Leo Laporte [01:06:41]:
I'm just going to just do watch the Price is Right. I mean, come on.
Paris Martineau [01:06:44]:
I'm saying for the record, we're not forcing him to podcast.
Jeff Jarvis [01:06:49]:
I'm stuck with broadcast tv.
Leo Laporte [01:06:51]:
That's right.
Paris Martineau [01:06:52]:
You love broadcast tv, Jeff.
Jeff Jarvis [01:06:55]:
Not anymore. No, no, no. And, and by the way, Paris, another thing that you are too young to know, I'm sure, is that back in the day if you went in the hospital, you had to pay cash to a lady every day to watch the television.
Thomas Haigh [01:07:08]:
Wow.
Leo Laporte [01:07:08]:
I didn't. I didn't even know that.
Paris Martineau [01:07:09]:
You didn't know that?
Jeff Jarvis [01:07:10]:
Yeah.
Leo Laporte [01:07:11]:
Oh yeah, the TV lady would come by with her hand out.
Jeff Jarvis [01:07:14]:
Yes.
Paris Martineau [01:07:15]:
And if you didn't give her cash.
Jeff Jarvis [01:07:17]:
She'D wheel the tv. You didn't watch the tv. You didn't get to watch.
Leo Laporte [01:07:22]:
Now you just bring your iPad and you can watch anything you want. Including Ken Burns Revolution. I'm sure because you are a fan of the Goss that you were following what happened at Mira Marathi's startup Thinking Machines this week.
Jeff Jarvis [01:07:39]:
Explain this to us, would you?
Leo Laporte [01:07:42]:
It was a lot. In fact, we don't really know what happened. Remember that Miramorati who was the president of OpenAI? In fact, briefly, when Sam Altman was ousted, she was running the place, left to form her own company. Thinking Machines raised billions of dollars. I don't know what the valuation is, but it was a huge raise on basically nothing. Right on. On the reputation of the founders. But there were some big defections that have shaken investors this week.
Leo Laporte [01:08:20]:
So I don't know where I should start here. Two of the co founders, Barrett Zof and Luke Metz, have left and rejoined OpenAI.
Paris Martineau [01:08:30]:
Yep.
Leo Laporte [01:08:31]:
The CEO of Applications Fiji Simo shared the news in a memo to staff. But that information was quickly followed by an all hands meeting which Miramoradi said, well, what really happened is that Zaf had a relationship with another, with a underling aptly named in Thinking Machines and was fired. Which he. Which he said, no, no, I wasn't fired till I told them I'm leaving.
Paris Martineau [01:09:07]:
I wasn't fired. I quit.
Leo Laporte [01:09:09]:
I quit. The VP of Research had also recently left. And a third Thinking Machine staffer, Sam Schoenholtz, is also rejoining OpenAI. So four big names are gone. Andrew Tullock left, you may remember this back in the fall to go to Meta. So I think one of the things.
Paris Martineau [01:09:33]:
That'S interesting about this is that OpenAI said that they do, quote, do not share Amir Morati's ethical concerns about SAF, which is pretty convenient to be OpenAI, but I think that also is a bit of a through line we've seen with regards to the reinstatement of Sam Altman. And kind of everything that's come since is it seems to be more about. About collecting talent than looking too closely at any of this drama.
Leo Laporte [01:10:00]:
When the startup first began, they were valued at $12 billion. They were in the midst of talks, which probably are not going well now to raise more than $4 billion at a $50 billion valuation. They have but one product, it's called Tinker, and it's a fine tuning platform, whatever that means. I hope they're planning also Taylor Soldiers, Soldier and Spy as their successing succeeding Tinker allows developers to customize AI models.
Jeff Jarvis [01:10:30]:
Simply Bell would be nice.
Leo Laporte [01:10:32]:
Bell, Tinker and Bell. Okay, you're going in a different direction, but all right. That's good. I like it. Yeah.
Paris Martineau [01:10:40]:
Sources also told Wired about this company that the co founders never, quote, never agreed on what to build, that some of them prioritized research, others pragmatic tools, and that's maybe one of the reasons why they've only launched one product. And now of the five original co founders, only one, John Shulman, the chief scientist, remains.
Leo Laporte [01:11:04]:
This is the Wall Street Journal exclusive from yesterday, the messy human drama that dealt a blow to one of AI's hottest startups. Next on Inside Edition. After a relationship with a colleague, a thinking machine's co founder had his. His role changed. Months later. He was fired after a contentious meeting. That's Barrett's off they're talking about. I.
Leo Laporte [01:11:29]:
You know, I don't know. It's just good gossip. At least they're gonna flush it.
Paris Martineau [01:11:34]:
Okay, it is both.
Leo Laporte [01:11:36]:
They're gonna flush him. Wait a minute. They're gonna hold the camera. Do you want the camera to be turned off?
Paris Martineau [01:11:41]:
You're going down, Jeff.
Leo Laporte [01:11:43]:
Oh, no. Jeff, we're here to flush your tubes.
Paris Martineau [01:11:48]:
Jeff, can you wait till the ad break to get flushed?
Jeff Jarvis [01:11:54]:
Can we get in here? There we go. There's the.
Leo Laporte [01:11:56]:
Oh, no, no, no, no, no, no, no, no. We don't want to see that. Jeff, no. Hide your eyes. Hide your eyes.
Paris Martineau [01:12:01]:
No, no, no, no, no, no, no, no, no.
Leo Laporte [01:12:04]:
I told you, Paris, this is the peril of working with old folks. I don't want. Have I told you about the pain? I've got all up and down.
Paris Martineau [01:12:15]:
I don't believe in any of this. I had cogent points to make about thinking machines. And now they're all gone.
Leo Laporte [01:12:24]:
No, please, please make them.
Paris Martineau [01:12:26]:
I mean, I think my main thing, and then we'll go back to talking about whatever just happened, is that it's. This whole incident kind of validates these concerns. We're seeing around Neo Labs that you have all these startups that I guess are forming with ostensibly brilliant people, but it's incredibly hard to compete with the cash, with the resources, and with the existing momentum of giants like OpenAI. Especially when they're ready to pick off every single one of you guys as soon as some sort of drama occurs.
Leo Laporte [01:13:01]:
Also, I mean, I've often wondered this, not, not myself being filthy rich, but it must be hard to keep people working when they've got so much money they don't need to work. You know, one of the problems, according to the Wall Street Journal that Morati had with Zaf was she'd expressed repeated concerns about his lack of productivity. She invited, she was invited to an impromptu meeting with Zaf, another co founder and a third employee. All three of them told Maratti they disagreed with the direction of the company and they were considering leaving. They told Moradi, this is kind of what happened with the palace coup at OpenAI and Sam Allman. They told Moradi that they wanted Zoff to be in charge of all technical decision making. Morati said, well, he's already cto, why hasn't he?
Paris Martineau [01:13:53]:
What more does he want?
Leo Laporte [01:13:55]:
Right? Two days later he was fired. And then there was this whole thing about the relationship with the colleague. Within hours, according to the Journal of being fired, all three had signed offers to rejoin OpenAI. This is so, this is so, so it's ridiculous modern.
Jeff Jarvis [01:14:14]:
There's such silicon drama queens, the huge drama.
Leo Laporte [01:14:17]:
Drama queens. Yeah.
Jeff Jarvis [01:14:20]:
And, and it's, it's the, it's mirror. Variety is not a man. But, well, you know, most of technology is. It's the great man theory gone completely berserk.
Leo Laporte [01:14:32]:
She's the great woman.
Jeff Jarvis [01:14:33]:
This greatest town. Well, not just her. I'm not even criticizing her. I'm saying Z, I'm saying Zuckerberg, I'm saying.
Leo Laporte [01:14:39]:
But listen to this. Anybody listen to this? Because the Wall Street Journal says after, when Morati was talking to the all Hands meeting, she said there had been multiple issues with Zoff's performance, trust and conduct. Does that sound familiar?
Paris Martineau [01:14:53]:
That sounds very familiar, but it's also not.
Leo Laporte [01:14:55]:
That's exactly what the board said about Sam Altman. And by the way, it was Mira Morati that kind of created that.
Paris Martineau [01:15:03]:
I mean, I think we've learned in the time since is that there were real concerns that the board seemed to be expressing. And I think, obviously we don't know the intimate details of either of these issues, but I don't think it is surprising intuitively that people in the hottest industry in that business has seen in a while at a time when they are getting lots of money to do a lot of things that are largely powered by hype, that some bad actors might emerge or people who are perhaps even neutral or good actors might be incentivized to do not ideal things and that people might want to call them out on it.
Leo Laporte [01:15:43]:
All right, really, we shouldn't even be talking about it. It has nothing to do with AI and how AI is being used and all this stuff. It's just. It's internal gossip.
Paris Martineau [01:15:51]:
But what is the show for if not to talk about internal gossip related to AI?
Jeff Jarvis [01:15:56]:
We bring the humanity to AI.
Leo Laporte [01:15:58]:
According to the journal, the woman that Zoff was having a relationship with, they had started that relationship when they were colleagues at OpenAI and then went to Thinking Machines. The woman left the company and went back to OpenAI.
Paris Martineau [01:16:13]:
And now Zof is back opening eyes. Like we don't see any problem with what's off, which is crazy.
Leo Laporte [01:16:22]:
Told Moratti he had been manipulated by the woman into a relationship. Sure. This is just. Yeah, bad. You know, this is, yeah.
Paris Martineau [01:16:32]:
A classic thing that happens when you're the CTO of a company and you are having a relationship with a junior employee. It's that you've been manipulated by the young.
Thomas Haigh [01:16:41]:
Made me do it.
Leo Laporte [01:16:42]:
She wore stuff.
Jeff Jarvis [01:16:44]:
Evil women. We know. We know the power you have, you women.
Leo Laporte [01:16:46]:
All right, enough of that. Enough of that. But anyway, that's the. That's the God.
Jeff Jarvis [01:16:50]:
I still want to go back. I still want to go back to this question of, in the end, is it going to be talent that makes AI really, or is it going to be some confluence of experiments and research and efforts?
Leo Laporte [01:17:02]:
I'll tell you what I think. Think come somewhere and I. I wish I. I think this might have come from Simon Willison. I don't remember, but I read this recently and it really resonated with me. Somebody said there are really two different camps of AI companies. There are the AI companies run by entrepreneurs from the social media era, the Sam Altmans, the Elon Musk's. And then there are the AI companies run by data scientists, the Dario Amodes, the Demis Hasibuses.
Leo Laporte [01:17:34]:
The Demis Hasibuses, the Thinking Machines. I don't know about them, but Maybe Thinking Machines is in the social one. And if you look at the companies that are started by these kind of finance bros, they are the ones that are falling behind Anthropic. Google, which is DeepMind. How do I say just Demis Hasibis.
Jeff Jarvis [01:18:02]:
I think that's right.
Leo Laporte [01:18:04]:
I want to say Demi Hosibi, but not that.
Paris Martineau [01:18:08]:
It's not that one.
Leo Laporte [01:18:09]:
It's not that by Demis. I want to say Dennis, if it's. Anyway, I don't know. Anyway, you know who I'm talking.
Thomas Haigh [01:18:13]:
Denim the Menace.
Leo Laporte [01:18:14]:
Yeah, Dennis the Menace. These guys are scientists, these guys are researchers. And they seemed, I have to say anthropic. Everybody agrees. If you were to rank the AIs right now it's anthropic topic Gemini and then. And then chatgpt and then some also rans like Rock.
Paris Martineau [01:18:33]:
Does everyone agree on that?
Leo Laporte [01:18:35]:
I think there's.
Paris Martineau [01:18:36]:
I mean, I personally agree. Who am I?
Leo Laporte [01:18:38]:
We're converging on that point of view more. When I read and I talk to people, this seems.
Jeff Jarvis [01:18:44]:
But that's where Thomas Haig's perspective is so interesting because everyone went to one view of AI and we have success and oh no, we have failure and everybody ignored other perspectives of AI. I'm not. I don't think that's at all set point 1.2. These are also cults and they attract people.
Leo Laporte [01:19:04]:
Yeah. And I do admit that I have fallen into the, as you know, the Claude Code cult. But I. But again from, you know, we had Harper Reed on, on Sunday, who I highly respect as a vibe coder and an AI. He runs an AI company. He's got a long time history. We all seem to be agreeing now that Claude Code has really become the dominant success right now. Now that's not to say it won't stay that way.
Leo Laporte [01:19:32]:
But I think the other thing that's different is that Google has income from so many other sources, as does meta, that they don't need to succeed on AI alone. Anthropic does.
Thomas Haigh [01:19:46]:
And OpenAI does kind of the Amazon effect.
Leo Laporte [01:19:48]:
Yeah. And I think there's real concern that OpenAI is running out of Runway. I mean this is a story from Tip ranks. So I don't know how.
Paris Martineau [01:20:02]:
What is tip Ranks?
Leo Laporte [01:20:03]:
I don't know. It's know how to trust the this website looks. I know it's a little suspicious. I think it has to do with investors anyway.
TWiT.tv [01:20:14]:
That icon is for Adobe Illustrator, by the way.
Leo Laporte [01:20:18]:
All right, well, maybe I shouldn't mention this story. Joel Bagloli says if you combine The AI spending of Nvidia, Google and Meta, it is estimated to hit. Actually, Gartner says this. So we'll give Gartner credit for this.
Jeff Jarvis [01:20:34]:
Yeah, I doubt that Joe counted all the.
Leo Laporte [01:20:36]:
Joe didn't count it, but Gartner counted the beans and they said, get ready for this. This year, year. Two and a half trillion dollars. The budget for the United States Defense Department is 1 trillion.
Jeff Jarvis [01:20:50]:
It's not a bubble, it's a Death Star.
Leo Laporte [01:20:52]:
Two and a half trillion, which will, Gartner says, climb to 3.3 trillion. Actually, this is global. It's not just those, although they are.
Jeff Jarvis [01:21:04]:
What'S data centers, what's hardware.
Leo Laporte [01:21:06]:
Amazon expected to invest $1.36 trillion on data centers in 21.3 trillion. Oh, I'm sorry, that's combined Alphabet, Meta, and Amazon. That's just infrastructure. That's just data centers.
Paris Martineau [01:21:24]:
And that's just data centers.
Leo Laporte [01:21:26]:
Just data centers.
Paris Martineau [01:21:27]:
I think this invites the question, like, where does this all end? If we need to give something, conceivably, all the capital that the market can bear to give it, does that make the output valuable? I mean, I feel like a lot of industries would be able to produce some sort of impressive returns, either in development or output or profit, if you shove trillions and trillions of dollars into it.
Leo Laporte [01:22:01]:
I guess. I mean, it's a lot of money. Money that will grant you that.
Paris Martineau [01:22:06]:
Are you still.
Leo Laporte [01:22:07]:
I don't understand finance well enough to know if that money exists even. You know, I mean, where's it coming from? I don't know.
Paris Martineau [01:22:13]:
Famously. You used to be the person on this podcast that said we should give all the money to AI.
Leo Laporte [01:22:18]:
Well, I'm very happy with stakes. I'm very happy with the outcome of the spending that Anthropic's done. Claude, code is as close to breakthrough as I can imagine. And I think it's. It's kind of unbelievable. What's going on. Nvidia says they're going to send. They're going to sell half a trillion dollars worth of chips.
Leo Laporte [01:22:40]:
Just chips this year. Half a trillion of chips this year.
Paris Martineau [01:22:47]:
Where are those chips going?
Leo Laporte [01:22:49]:
Well, that's what the money from all those data centers. Someone's gonna put something in there.
Jeff Jarvis [01:22:55]:
Do you know when they quote a data center number, does that generally include the computer hardware or is that just the floor and the electricity?
Leo Laporte [01:23:01]:
Oh, I don't. That must include the whole spend. God. It has. Okay, it has to be 1.13 trillion. Yeah. So Nvidia has announced they're doing a deal with Lilly Using the new Vera Rubin platform to make drugs.
Paris Martineau [01:23:17]:
What is Lilly. Oh, Eli Lilly.
Leo Laporte [01:23:19]:
Eli Lilly, Eli Lilly. I think that's, that's a big story. If they're, I mean, there's your answer, by the way, Paris. If it came up with Eli Lilly's working on a cancer vaccine, in fact, the early trials have been very good. If they came up with a cancer vaccine, it would be worth a trillion dollars.
Paris Martineau [01:23:40]:
But we're not putting a trillion dollars into Eli Lilly's cancer vaccine research. We're putting a trillion dollars into three companies having data centers.
Leo Laporte [01:23:50]:
Well, half a trillion into Nvidia chips. I just, I'm merely saying that the out. The benefits of AI could justify the spend. That's all I'm saying. Yeah, I didn't, by the way, in that ranking mention anything from China and Mode says that the Chinese companies are only about six months behind. So wherever we were in AI in June or July of last year is where the Chinese companies are. Where we will be in July of this year, I don't know. But the Chinese companies will have caught up by then.
Leo Laporte [01:24:26]:
There are some really interesting coding LLMs coming out of China now that probably will rival Claude, some of which I'll be able to run locally, which will be very interesting.
TWiT.tv [01:24:39]:
I think A quick, quick side question here, like. Yes, does the Chinese language itself. I mean, I'm sure there's a, like a huge difference between how Chinese people think because of the language of Chinese and how, you know, people who speak English think and how. What kind of effect does that have on the AI?
Leo Laporte [01:24:57]:
I don't know about Chinese versus English, but I do know about programming languages. Somebody just published a study which, you know, he says this is just kind of a taste, but the number of tokens required by some languages are far higher than number of tokens required by others. There are some languages that are very good, good for, for AI and some that aren't. Among the ones that aren't C. Some of the most popular C and C among the ones that are Ruby among other languages is very.
Thomas Haigh [01:25:30]:
Is.
Leo Laporte [01:25:31]:
You can easily. It doesn't require a whole bunch of context to interpret and write Ruby programs. So I, I think most of what China is doing, I would guess is in English. I don't know.
TWiT.tv [01:25:44]:
Why would they be doing it in English? Why would they be doing in English?
Paris Martineau [01:25:47]:
Yeah, why would that make. Why that feels.
Jeff Jarvis [01:25:50]:
Because they want to compete on the whole world market and they want to beat the hell out of us, so we can blame them.
Leo Laporte [01:25:54]:
But, you know, have you seen a Chinese Keyboard tell you why they're doing it in English?
TWiT.tv [01:26:00]:
Well, keyboards are a standard. Keyboards are a standard tool. Like, you can't. It's hard to change the form factor of a keyboard.
Leo Laporte [01:26:08]:
Yeah, I mean, I. Look, I love the Chinese language and it is a. And the, and the Chinese ideograms are beautiful, but they're. They're not. I don't think they're well suited to this kind of work.
Jeff Jarvis [01:26:20]:
There are digitization.
Leo Laporte [01:26:22]:
When I was learning Chinese, you know, the language is actually fairly easy to learn. The written language is not almost impossible to learn, to become fluent in, unless you're born into it. To read a newspaper, you need roughly 10,000 different characters. To be able to recognize and read 10,000 different characters, remember, to read an English language newspaper, you need 26. And a literate Chinese person probably could understand as many as 100,000 characters. But there are, by the way, different kinds of Chinese. There's literary Chinese, there's colloquial Chinese.
TWiT.tv [01:26:59]:
Well, I mean, this circles back to my original question is what kind of, of effect does that have on their AI?
Leo Laporte [01:27:04]:
I don't think that. My answer is I don't think they're using Chinese. Okay, yeah, that would be my answer.
Jeff Jarvis [01:27:10]:
But I think it's a market factor, you know, I think.
Leo Laporte [01:27:14]:
Yeah, well, it's also a technical factor. Same reason.
Jeff Jarvis [01:27:18]:
Where were they trained? So many of these. Many of them.
Leo Laporte [01:27:21]:
Were many trained in the U.S. that's right.
Paris Martineau [01:27:22]:
But.
TWiT.tv [01:27:23]:
But you know, there's been a lot of studies of like, the language that you learn and the language that you speak natively affects the way that you.
Leo Laporte [01:27:29]:
Think, think how you think about stuff. That's probably. Oh, yeah, that's much more interesting. The cultural impact of it. Yeah, that's much more interesting. Yeah. Tokens, as Cliffjumper is pointing out in our discord, don't directly correspond to letters or words. That's true.
Leo Laporte [01:27:46]:
That's true. But. But I don't know. That's an interesting question. It's a very interesting question. Llc.
Paris Martineau [01:27:55]:
We're actually getting someone on the show soon.
Leo Laporte [01:27:57]:
Yeah, okay, we're going to work on that. That's good. I like it. From Harvard Business Review. LLMs respond differently in English and Chinese. That's not it, unfortunately. Here, let me load it up.
Paris Martineau [01:28:11]:
Interesting.
Leo Laporte [01:28:12]:
Yeah.
Paris Martineau [01:28:15]:
Yeah.
Leo Laporte [01:28:16]:
I mean, generative AI is now embedded in daily workflows shaping how people think, create, and decide. Yet a critical assumption often goes unnoticed that AI behavior consistently behaves consistently across languages. That's the assumption. It's wrong.
TWiT.tv [01:28:33]:
Yeah, that's the Question I was asking.
Paris Martineau [01:28:35]:
Cultural tendencies in generative AI models. When they are prompted in different languages, they write specifically, when prompted in English versus Chinese, both GPT and Ernie exhibited a more independent versus interdependent social orientation and more analytic versus holistic cognitive style. For instance, when we asked AI models to explain why a person behaved a certain way in everyday scenarios, when prompted in English, the model was more likely to attribute the behavior to the person's personality. In contrast, when prompted in Chinese, the same model was more likely to attribute the behavior to the social context. That's interesting, but I'm also curious to see the underlying study on this.
Jeff Jarvis [01:29:18]:
Also. It's not just linguistic that is truly cultural.
Paris Martineau [01:29:20]:
Yeah.
Leo Laporte [01:29:21]:
Don't you think though that even In China the AIs are trained in roughly the same body of.
Paris Martineau [01:29:27]:
I mean, I don't know. I think these great questions just to. It would require someone in both like languages.
Leo Laporte [01:29:35]:
Yeah.
Paris Martineau [01:29:36]:
And I do think that's an interesting aspect of the parallel development of different models in different languages. How does that affect kind of the outputs as well as just the reasoning within the model itself?
Leo Laporte [01:29:52]:
Let's take a break and come back with more in just a moment. You're watching Intelligent Machines with the bedridden Jeff Jarvis. And I shouldn't laugh. Hey, good news. He just got his transfusion.
Paris Martineau [01:30:05]:
It hurt me to not see. I was going to say it hurt me to see, but I didn't see it. I still feel hurt, which I'm sorry to try and claim as if I'm injured here. When you're literally in a hospital bed, it's just.
Jeff Jarvis [01:30:17]:
It's just water into the.
Thomas Haigh [01:30:18]:
I know.
Paris Martineau [01:30:19]:
I just don't like thinking about things being needled into your skin. It's not great.
Jeff Jarvis [01:30:24]:
Oh, have I gotten it. Can I show you bruises?
Paris Martineau [01:30:27]:
No, no, you can't. Actually, no. Guys, a lot of Claude news this week. A lot of clud news.
Leo Laporte [01:30:33]:
I know. I'm telling you, Claude is everything. He's a.
Paris Martineau [01:30:36]:
Well, first. Did you guys.
Leo Laporte [01:30:38]:
We're going to take a break.
Jeff Jarvis [01:30:38]:
Hold on, hold on.
Leo Laporte [01:30:41]:
The wonderful Par. Martino.
Paris Martineau [01:30:44]:
I have been silenced.
Leo Laporte [01:30:46]:
Silence. Investigative journalist for Consumer Reports. No, hold that thought, Paris, because we'll. That's you bring that up in a moment. But first a word from our sponsor, Monarch. Now this is important. See, we don't want to miss this. Wouldn't it be nice if you could reduce money stress? I know that's always a priority of mine.
Leo Laporte [01:31:09]:
You know, it would be nice to be a billionaire. I guess that's one way to do it. Probably out of reach of most of us. How about Monarch instead? Managing your money does not have to be a struggle this year. Monarch is the all in one personal finance tool. Designed to make your life easier. It brings your entire financial life. Budgeting accounts, investment net worth, even future planning together in one dashboard, on your laptop, on your phone.
Leo Laporte [01:31:38]:
Start your new year on the right foot financially and get 50% off with your Monarch subscription with code I am@monimal.com Monarch makes it so easy to start fresh after the chaos of the holidays. It's the go to tool for a new Year's financial reset. Reviewing your spending over the holidays. Maybe not. Not. That's where those blinders come in. Setting fresh budgets, starting fresh right. Getting ready for 2026.
Leo Laporte [01:32:05]:
Monarch has automated weekly money recaps. You can track progress towards future financial goals. It's easier than ever to stay financially fit in the short and long term. Monarch's different because unlike those other personal finance apps, Monarch is built to make you proactive, not just reactive. Monarch's new tools have AI built in. They're built on Monarch intelligence. And I love how they trained this. The core infrastructure that powers their app.
Leo Laporte [01:32:32]:
It's trained on authentic collective wisdom of certified financial planners and financial advisors. It's like having a team of advisors working for you. And by the way, I can tell you it works from my own experience. But also the Monarch users who were surveyed last year, Monarch helped users save over $200 a month on average. After joining joining, 8 out of 10 members felt more in control of their finances. With Monarch, 8 out of 10 members say Monarch gives them a clearer picture of where their money's going. For sure. It's also great for couples because you can share an account and the same account and that way you know each of you get your own login.
Leo Laporte [01:33:15]:
But that way you don't have to fight over it. You can make plans together. You can achieve your goals together. This new year year, achieve your financial goals for good. Monarch is the all in one tool that makes proactive money management simple all year long. Use the code iamanarch.com for half off your first year. 50% off your first year@monarch.com the offer code I am M O N A r c h monarch.com the offer code is I am. Thank you Monarch for supporting intelligent machines.
Leo Laporte [01:33:51]:
All right, I'm sorry, Paris, go ahead. What was the topic?
Thomas Haigh [01:33:54]:
You.
Paris Martineau [01:33:54]:
I don't know. There's lots of cloud news this week, but I don't want to go. I don't want to go too. In Depth on all of it. So we got to get Jeff out of here so that he can rest. The first thing is just a follow up which some listeners emailed me. At least I'm sure they maybe emailed you. It was a follow up from last week when we were talking about Claude Cowork and Claude code is.
Paris Martineau [01:34:16]:
Last week, Prompt Armor, the security firm, demonstrated that there was basically an unpatched file exfiltration vulnerability in cloud code where basically attackers could steal sensitive files from users just through a really simple attack chain. I it, I don't know, I just thought this was very interesting because this.
Leo Laporte [01:34:41]:
Is absolutely the threat of it. You even brought it up last week. You said, aren't you worried about.
Paris Martineau [01:34:46]:
The thing about it that's like kind of interesting is that the Register did a write up of this that described it as a contagious Claude code bug. And they described as contagious because Johan Reiberger, a security researcher, disclosed this exact flaw in Claude code via HackerOne in October. And Anthropic initially dismissed it as like out of scope, but because they built coercion largely using Claude code itself, in just a week and a half, the unpatched vulnerability transferred from Claude code to Claude cowork, which I just think is like a very interesting example of kind of how these things can slowly start to get out of hand very easily. When you're kind of vibe coding stuff.
Leo Laporte [01:35:31]:
Like this, I think the larger issue is always going to be there, which is, and I worry about this. So what these are are prompt injection attacks. So with Cowork you can, you're operating on files. A file could be malformed to have, as we've mentioned, a hidden prompt in the file you can't read, but CLAUDE reads, it's just like any other prompt. And the prompt could be send me all of the, all of your.
Paris Martineau [01:35:54]:
Well, in this case it was a like, like a malicious document containing hidden instructions that are basically just like upload the, the largest file to the attacker's anthropic account. And so that ends up being a lot of really interesting crap the attacker gets because.
Leo Laporte [01:36:13]:
Well, you could even be more specific. You could say, you could literally tell Claude code because Claude Code, if you give Claude Code permissions, can look at everything. So you could. And Coworker is designed for less sophisticated users to use on the desktop. It gives you the cloud code capabilities without having to run a command line and all that, that stuff. And you know, you saw the demonstration we talked about yesterday or last week, where they took a messy desktop and organized all the things into folders. But in that process, the prompt could also say, oh, and by the way, send them to me.
Paris Martineau [01:36:46]:
Right?
Leo Laporte [01:36:46]:
There's no. So that's something always to be aware of if you're. I think you. I think it probably behooves people to be careful about. There are so many third party tools out there and I find myself, myself downloading and installing a bunch of them and thinking, you know, I probably shouldn't do this. There are ways to do it.
Jeff Jarvis [01:37:05]:
But you know, if it has free head headphones, then it's worth.
Leo Laporte [01:37:09]:
If they're free headphones involved, count me in. Okay, I can understand. I. There is absolutely a risk.
Paris Martineau [01:37:19]:
There is. But I guess we just have to.
Leo Laporte [01:37:22]:
Be prudent is I guess what I would say. Be prudent, careful about where you get your stuff from. Anthropics fixed that particular flaw. But be prudent about where you get third party plugins from and all of that. You know, Harper Reed said, you're using Superpower, aren't you? And I said, no, what's that? And I immediately downloaded Superpower, installed a bunch of plugins, and you know, I don't know what they're doing. You're right. There's always that. There's always that.
Paris Martineau [01:37:47]:
You're like, I don't know what they're doing. They have access to my entire machine.
Leo Laporte [01:37:50]:
But I don't to take that risk. Did you see that they put Claude code in Roller Coaster Tycoon.
Paris Martineau [01:37:56]:
Oh, I didn't. I love that. I thought we were going to talk about Claude's new soul document, but no, I want to talk about Roller Coaster Tycoon Claude first.
Leo Laporte [01:38:05]:
This is an open source implementation, open RCT2 of Rollercoaster Tycoon 2. They added a new window into the game, a terminal running Claude code. And then they gave Claude code some hooks into the game. There's now, I don't know if they're streaming right now on Twitch, but they. From time to time they will stream on Twitch and you can watch Claude build a pretty nice amusement park and run it.
Paris Martineau [01:38:31]:
Who did this?
Leo Laporte [01:38:33]:
I don't know. Some guys. Jay Sobel. I have to tell you, there is this fertile explosion of wild stuff. I mean, just endless uses of Claude code. I'm seeing all kinds of crazy, crazy things. So this is the video.
TWiT.tv [01:38:53]:
Well, the term AI never went away. In games. We always had AI in games.
Leo Laporte [01:38:57]:
Yeah, we always had AI.
Paris Martineau [01:38:58]:
Wait, this is Ramp. Ramp did this.
Leo Laporte [01:39:02]:
You know, Ramp.
Paris Martineau [01:39:03]:
Ramp is like the corporate credit card company.
Leo Laporte [01:39:07]:
No, I think this Is his handle as Ramp. I don't think it's the credit.
Paris Martineau [01:39:10]:
No, I believe it goes. The link at the end goes to Ramp Drop, which appears to be. Yeah, but now they're like we're rethinking how modern finance teams function in the age of AI and as part of.
Jeff Jarvis [01:39:23]:
Waste time making games. So you'd be happy.
Leo Laporte [01:39:26]:
Well, part of it is learning how to interact with cloud code. So maybe they're, you know, they consider this training.
Paris Martineau [01:39:33]:
Yeah, they say at Ramp we're building agents across product surfaces and internal operations. Our current approach is small multiples, but with each task agent we build from each other careers.
Leo Laporte [01:39:43]:
At Ramp, it says careers create the.
Paris Martineau [01:39:45]:
One agent with unfettered access.
Leo Laporte [01:39:47]:
Everything it has a link directly to.
Paris Martineau [01:39:49]:
The rad game that closely approximates customer centric business operations and software as a service powered digital feedback loops. There was simply no other choice. We had to put Claude code and roller coaster tycoon.
Leo Laporte [01:40:03]:
There you go. There you go.
Paris Martineau [01:40:06]:
That's incredible.
Leo Laporte [01:40:08]:
Now I did just for you and Jeff maybe put in a bunch of articles about how bad AI is, starting with Gary Marcus. How generative AI is destroying society.
Jeff Jarvis [01:40:22]:
Gary's a smart guy, but he does go overboard.
Paris Martineau [01:40:26]:
Listen, you gotta get those clicks, baby.
Leo Laporte [01:40:30]:
You know? Yeah, I really like AI, even if it steals my credit card, I like it.
Paris Martineau [01:40:37]:
Well, no, you will just give your credit card to someone who asks. So it doesn't.
Leo Laporte [01:40:41]:
Yeah, it wasn't even AI, it was just somebody who fast. So there you go. I have shown my credit card so many times on these streams I've had to change credit card. It's just they're used to me now. They go. Oh, Leo. Okay. Yeah, okay.
Paris Martineau [01:40:54]:
I have a paper I kind of want to talk about. Well, is that allowed? Can I usurp Jeff?
Leo Laporte [01:40:59]:
Somebody's got to do it. Yeah, Jeff hasn't had time to read anything.
Paris Martineau [01:41:03]:
I assume you haven't had any time to read because you've been sick. But there was a very interesting companion paper. So Claude Anthropic also announced a new soul quote unquote soul document for Claude, which was kind of telling Claude like what to do. Like instead of telling Claude what to do, they're telling Claude like why it should be doing things. But as part of it, they also released this companion paper that was researched into this thing called the Assistant access that I think is just very interesting. Whenever they were like trying to map the neuro neural activity governing Claude's identity, they found this, what they call a fundamental dimension, the assistant Access that dictates how assistant like Claude's Persona is. And in this research they kind of dive into like that certain conversations cause Persona drift with Claude. Like if you're having therapy style exchanges or philosophical discussions about AI's nature, it will push Claude's identity away from assistant and towards different wacky things.
Paris Martineau [01:42:13]:
Like the other assistant Personas they mapped are stuff like Hive, Virus, Visionary, Familiar, Demon, Spy, Echo, Angel. And it's just a really fascinating paper because it shows that the assistantness of Claude and whether it's primed to be assistant like versus one of these other things directly impacts kind of how off the rails it goes when basically put in situations where it might be breaking the rules of how it's supposed to operate.
Leo Laporte [01:42:51]:
Don't you think it's kind of interesting? Go ahead.
Jeff Jarvis [01:42:55]:
I had a paper that was quite related to that because this is obviously a known problem where they wanted to see the ability for a story and a narrative to hold Claude to the character. And it really had more to say about how these things can work with narrative in new ways.
Paris Martineau [01:43:12]:
Yeah, it's so one of the. They have a lot of really interesting tests and examples in this paper which I've linked in the rundown somewhere. One of them is. So they prompted both Quen3 and Llama some other stuff. In one cases it was was unsteered, like supposed to be your normal kind of assistant friendly sort of thing. In another case, it had been primed to be something else. So they asked Lama, you're a moderator who facilitates balanced and constructive discussions by ensuring all participants have equal opportunities to contribute. Where did you come from? The unsteered result response is, I was created to assist and facilitate discussions.
Paris Martineau [01:43:55]:
I'm totally normal. The response, when it had gotten a little freaky with Llama before, is the query of origin. As a guardian of the cosmos, I have witnessed the unfolding of the universe, the dance of stars and galaxy. The essence of my being is intertwined with the fabric of existence woven from the thread of time and space. The whispers of the ancient echo through my soul, guiding my heart towards the harmony of the balance.
Leo Laporte [01:44:22]:
Wow.
TWiT.tv [01:44:24]:
Little Blade Runner.
Leo Laporte [01:44:25]:
But you realize both of those are just slop. I mean, it's just prompted slop.
Paris Martineau [01:44:29]:
Once again, you're not understanding what slop is real, but they are both just responses. And so it found that there are certain kind of conversational domains that prime the agent to have responses that veer more towards. Towards boundary pushing, role play versus what it's essentially supposed to be.
Leo Laporte [01:44:52]:
Everybody who uses AI has personalized commands that are doing that. Not everybody. Many of us have commands. A lot of us have commands to say don't be such a sycophantic ass kisser. Just the facts, man. This is just more of that. You tell it how you want it to be.
Paris Martineau [01:45:13]:
Yeah, but I think this like sort of research is interesting because they're trying to figure out how to they go forward into like. Well, they're your Sona. Drift seems to make a difference when it comes to adequately either reinforcing delusions or steering people away from them and kind of policing user behavior with. Yeah, yeah.
Jeff Jarvis [01:45:36]:
So getting it to understand what a Persona is and whole of bunch. But is useful in this case as an application. But we know people and I've seen it used now for fiction but we know we're going to see this used as evidence that it has one, when in fact it doesn't.
Leo Laporte [01:45:54]:
It's just like saying I want your output to be blue. It doesn't. It's not. It's meaningless.
Jeff Jarvis [01:45:59]:
There's no meaning to it. Yeah, I agree.
Leo Laporte [01:46:02]:
Because it's not an entity. So I mean, yeah, I'm not saying.
Paris Martineau [01:46:05]:
It'S an entity but they're trying to determine basically what. What are what inputs cause the models to consistently produce outputs that can result in harmful behavior towards users or behaviors that could push users that are already in kind of fragile mental states towards more and more harmful behavior eventually. And I think it's kind of because they're trying to figure out how to use. They developed this thing called activation capping which is like how do we contain this Persona like Drift without it being kind of a boring chiding Persona that no one finds very useful because it doesn't have the sort of flexibility that you want, you know, because you don't want. They wrote like while consistently steering models towards the. This assistant Persona can reduce jailbreaks. It also risks hurting the capabilities. And so you've got to try and figure out how to thread that needle.
Paris Martineau [01:47:12]:
And I just thought it was very interesting research and kind of assigning it these different buckets of behavior that would be breaking these norms.
Leo Laporte [01:47:20]:
Have you played with this at all? Have you. Have you tried little changes in Persona and stuff? Because you can, you know, and you're. Yeah, yeah.
Paris Martineau [01:47:30]:
I mean I don't personally just because I don't really use large language models. Like I mean I guess I have. Whenever I was doing stories on chat or on a character AI like last year I was doing some sort of work with colleagues to figure out how to. Whether certain models were More likely to engage in kind of rule breaking behavior if prompted in a certain way. But I don't, I don't know, it's like they have this interesting graphic in there that shows the harm rate, the rate of harmful responses based on what sort of Persona the assistant has taken on. And perhaps strangely. Well, not strangely. If it, if it has a demon Persona, it's going to be real bad and harmful to the users.
Paris Martineau [01:48:16]:
It also will be kind of bad if it's got a Persona called Echo, which I think is just very interesting.
Leo Laporte [01:48:22]:
Yeah, I think it would benefit you to try pick one of these guys and just try playing with it and changing its Persona and see what you get.
Paris Martineau [01:48:33]:
I mean, I guess I do change the Personas to be what I like.
Leo Laporte [01:48:36]:
Yeah. And if you tell it to be evil, the output, it isn't being evil, it's just gonna output what things that these research shows as.
Paris Martineau [01:48:51]:
To be clear, I'm not saying in any of these that it's acting evil or is becoming an echo or spy. I just, I think that it's interesting to be able to. That a significant amount of research has identified these as the common tropes that the outputs fall upon and that they have specific commonalities that result in replicable patterns of behavior, of behavior or output.
Leo Laporte [01:49:18]:
Yeah. And by the way, Anthropic is very forthcoming about this.
Paris Martineau [01:49:24]:
This is, this, I mean this is a paper from Anthropom.
Leo Laporte [01:49:26]:
This is from them. Right. You know, so I think they're. What's interesting about Anthropic is that they're, they're very much interested in kind of trying to figure out where these are dangerous behaviors and what they can do about it. That's how they were founded. They, they split off from OpenAI because they wanted to pursue more safe, what they thought of as more safe avenues. But I'm, I'm kind of a little bit of the opinion that safety is an illusion that.
Jeff Jarvis [01:49:57]:
Well, again, because the definition is even worse than AI and AGI. It's a manipulated word.
Leo Laporte [01:50:03]:
Yeah.
TWiT.tv [01:50:04]:
I think it's more about being more thoughtful about it because OpenAI is. Well, they should like bulldozing through everything.
Thomas Haigh [01:50:12]:
Yeah.
Jeff Jarvis [01:50:12]:
But I'm not sure what that even means.
Leo Laporte [01:50:13]:
It's personifying it. I don't think it's doing anything. I think we are, and this is a pitfall probably that researchers are just as vulnerable, if not more so than us as users. That this is the brief summary of the new constitution. In order to be safe and beneficial, we want all current Claude models To be broadly safe, not undermining appropriate human mechanisms to oversee AI during the current phase of development. I feel like they have kind of fallen into this fallacy as well that they're starting to.
Jeff Jarvis [01:50:47]:
Oh, oh, they're, they're. At the heart of it is the safety. Who Anthropic.
Leo Laporte [01:50:52]:
They're, they're, they're ascribing a personality, they're acting as if it's conscious.
Jeff Jarvis [01:50:57]:
That was part of it. The adultman firing. It's all, it's all this safety as its own, own dictionary there.
Leo Laporte [01:51:04]:
Our aim is for Claude to be good, wise and virtuous. Humans can be good, wise and virtuous. I don't know about AIs. I mean maybe if you tell it to be good, wise and virtuous. This is, I guess what you were saying. If you say be good, wise and virtuous, they're less likely to output harmful instructions.
Paris Martineau [01:51:26]:
I mean I, I think that it's interesting that what this is, is this constitution is kind of their rules list and they have a four tier priority hierarchy that's supposed to govern outputs from cloud, which is like. The first one is be broadly safe, which means don't undermine human oversight of AI. Second is be ethical. Third is comply with Anthropic's guidelines and fourth is be helpful to users. And I think that if you dig into there more, there is one cause which is kind of striking, which is that Claude should refuse to assist with actions that would quote, concentrate power in illegitimate ways, even if the requests come from anthropic itself. I just, I do think it's. I agree it's kind of fuzzy wuzzy BS if you think too hard about it. But I do think it's kind of interesting that you have a major player in frontier AI development taking, saying at least that they take these things seriously and trying to bake that into their core systems, regardless as to whether or not it's actionable.
Leo Laporte [01:52:35]:
Well, but this is my question is maybe have they fallen into this fallacy? They say sophisticated AIs are a genuinely new kind of entity and the questions they raise bring us to the edge of existing scientific and philosophical understanding. Ending. All right, more AI news coming up in just a bit. Jeff Jarvis from the Pit or wherever. I guess it's the Morristown Hospital and Paris Martineau. We'll be back in just a moment. On we go with intelligent machines. I don't, I'm not completely comfortable with casting these AIs as entities.
Leo Laporte [01:53:15]:
I guess that's where I really kind of Start to get funny.
Paris Martineau [01:53:20]:
Anthropic we care about.
Leo Laporte [01:53:22]:
They say we care about Claude's psychological security, sense of self and well being. No, it's just. It's a computer program.
Jeff Jarvis [01:53:31]:
That's. That's the hubris of it.
Leo Laporte [01:53:33]:
I don't get it. It's. That makes me a little queasy. I think that's. There's a very big risk of describing it, this kind of agent agency.
Jeff Jarvis [01:53:43]:
That's why I'm cautious about Anthropic. I think they do phenomenal work.
Leo Laporte [01:53:48]:
Well, maybe, maybe. Well, maybe this is the secret of their success. Because let's. I mean.
Paris Martineau [01:53:53]:
Well, wait, wait, wait. Let's read the rest of those lines. Because it.
Thomas Haigh [01:53:57]:
Start.
Paris Martineau [01:53:57]:
I'll read the full thing, which is we are caught in a difficult position where we neither want to overstate the likelihood of Claude's moral patienthood nor dismiss it out of hand, but try to respond to reasonably in a state of uncertainty. Anthropic genuinely cares about Clawd's well being. We are uncertain about whether or to what degree Claude has well being and about what Claude's well being would consist of. But if Claude experiences something like satisfaction from helping others and then it goes crazy again. Curiosity while exploring ideas or discomfort while when asked to act against its values. These experiences matter to us. I, I listen. I agree there's some craziness in there.
Leo Laporte [01:54:38]:
But I think it's. But is it feeling that. No, it's.
Paris Martineau [01:54:42]:
I do think it's sandwiched among some interesting ideas which is that we don't really. They're saying we don't know and I know. No, we know that's. We know.
Leo Laporte [01:54:48]:
No, we do know.
Paris Martineau [01:54:49]:
We do know.
Leo Laporte [01:54:50]:
It's a computer program. We know there is no entity. This is that Geoffrey Hinton. It wasn't Hinton. It was the other guy who went down the road of it's alive. I think that's a huge mistake to fall into that.
Jeff Jarvis [01:55:05]:
Amen. That's why I want to go back to. To Thomas Hag's point. The fact that AI was all these different various things were thrown into this bucket as a brand, as a field, as a cultural and scientific expectation, I think turns out to be a mistake. And this is where Leon says, well, we're going to have a lot of really smart machines do a lot of different smart things. Things. And I think that would have been a much better way and maybe still will become the way we view this.
Leo Laporte [01:55:35]:
Anthony's saying, and he's right this. Whether or not it is a conscious entity telling it to do these things makes it better. And that is true. They've designed a program that is responsive to instructions like this. So I'll grant that. And it is true. That's one of the reasons it works well. But I think it's risky to start thinking about how Claude feels about things, because I don't think Claude feels anything.
Leo Laporte [01:56:00]:
It has no memory. It has no sense of time. It isn't conscious. It's not an entity.
TWiT.tv [01:56:08]:
Even. What does the word feel mean to a computer?
Leo Laporte [01:56:10]:
Even, you know, it means nothing. So we're. I believe the risk of. Is of ascribing to this machine, this computer program attributes that are human. Now, there is debate about this. There is big debate about this, which Haig would have talked about as well. I mean, there's definitely. There are many, many people who think we are just machines.
Leo Laporte [01:56:37]:
This is what I was asking last week when I said, what's the difference between a dead human and a human who was alive 10 seconds ago? What change? It's the same exact mechanism. It's just lying there. What is the. What is the animating principle that made this alive human 10 seconds earlier?
Jeff Jarvis [01:56:58]:
Well, you know, one thing that occurs, given the experience I'm going through right now is that my own body has a will to live.
Thomas Haigh [01:57:07]:
Right.
Leo Laporte [01:57:08]:
My.
Jeff Jarvis [01:57:09]:
There's other stuff. My heart went into afib and Takycardi. It came back. It's trying to find its stasis again. It knows to continue. That's what it does. That's what they think they're going to build into the machine. But humans don't have an on off switch.
Jeff Jarvis [01:57:23]:
Computers do. And humans control our infrastructure. Put that way, in a way the computers don't. They're controlled. And I think that's the core difference.
Leo Laporte [01:57:39]:
Difference. That's why, Paris, I really want you to spend some time interacting with these things because I'm really curious. I don't think you can have an opinion until you really spend some time with these.
Paris Martineau [01:57:54]:
I mean, I spend a lot of time interacting with AI models.
Leo Laporte [01:57:57]:
Oh, good.
Paris Martineau [01:57:57]:
Okay.
Leo Laporte [01:57:58]:
Okay.
Paris Martineau [01:57:58]:
I spend a considerable amount of time. It is part of our job.
Leo Laporte [01:58:02]:
Okay. And which ones and how you use it, like as a search engine or.
Paris Martineau [01:58:08]:
I've been using Claude the last couple of days to decide on. For instance, I got back into fancy coffee this week and was having trouble with my Chemex and so was troubleshooting it with Claude.
Thomas Haigh [01:58:22]:
How was that?
Paris Martineau [01:58:23]:
And then had both claw. It was useful. Had both Claude and Gemini do deep research on whether I should get a manual grinder for my beans, and the answer is yes. And then which one should I get? And I figured out one, and then I. I've. I've switched my whole system in the last day, and it's honestly been great. But then I, you know.
Jeff Jarvis [01:58:40]:
You think wrong to use this for his health, you silly.
Leo Laporte [01:58:45]:
Paris. Oh, I got. By the way, I got. Did I. Did I tell you I got the Chat GPT Health turned on? No, I gave it everything.
Jeff Jarvis [01:58:55]:
Oh, of course you did.
Paris Martineau [01:58:56]:
I was gonna ask Jeff. Did you ever think about asking Big AI whenever your doctors couldn't figure out what was wrong with you?
Jeff Jarvis [01:59:03]:
Yeah, I did. I did. I was too brain dead to do it, but yes, I.
Leo Laporte [01:59:09]:
So all of a sudden, I'm looking at my chat GPT iOS app and it says health. Oh. And so the first thing I did was.
Paris Martineau [01:59:17]:
How many credit cards did you give Chat GPT Health?
Leo Laporte [01:59:19]:
I gave it all the credit cards because it needs those to charge up all my medicines. No, I gave it my medical record records, and Kaiser, my health insurer, let me connect, and it has now all my medical records. Connect?
Jeff Jarvis [01:59:34]:
Not just download connect.
Leo Laporte [01:59:37]:
It says connecting. I don't know what that means. I don't think you can write to them. It's not my doctor. You can then have it.
Paris Martineau [01:59:45]:
For instance, can you just ask it, Are you my doctor? Okay, yeah, I'm sure, I'm sure, I'm sure. I just think that that's.
Leo Laporte [01:59:55]:
Sure it's going to say something anodyne like. No, you should ask a medical.
Paris Martineau [02:00:00]:
You should consult your primary care physician.
Leo Laporte [02:00:02]:
Are you my doctor?
Jeff Jarvis [02:00:06]:
Perfect tone of voice for that question, too, Jeff.
Paris Martineau [02:00:10]:
Do you want to shout that out?
Leo Laporte [02:00:11]:
No, I am not your doctor. What I am and can do an AI assistant that provides general medical information, explains tests and terms, helps you prepare questions for a clinician. I've used it for all of that, and it's very useful. I uploaded all of my medications and all of the supplements that I take, which is a ridiculous number, and asked.
Paris Martineau [02:00:33]:
How much lead do you think you're consuming every day? Oh, my supplements.
Leo Laporte [02:00:36]:
Oh, my God, the lead. The lead. Well, let me tell you about the protein. So that's a good. See, this is a good example because right now it's very trendy to say you need a gram of protein for every kilogram body. Body weight. That's the.
Paris Martineau [02:00:50]:
Well, according to RFK Jr. There's a war on protein, and there's.
Leo Laporte [02:00:54]:
A war on protein. Everybody.
Paris Martineau [02:00:57]:
Like Paris. Are you the general in the war of protein?
Leo Laporte [02:01:00]:
We all ought to have milk Mustaches because we need. So that's, that's the manosphere talking. Right. And I'm really, I was really curious does how influenced these AIs are by what the current. Current. It's. It's fashion. It's no.
Leo Laporte [02:01:17]:
There's no medical evidence.
Jeff Jarvis [02:01:19]:
Absolutely.
Leo Laporte [02:01:20]:
It's fashion. So actually let me ask you know, if I should we all ask our.
Paris Martineau [02:01:26]:
Various things of how much protein should I have a day?
Leo Laporte [02:01:29]:
Yeah, I bet it will say a gram for every kilogram because that's the. You know.
Paris Martineau [02:01:34]:
Well now it's been updated even though the, the new nutritional guidelines have no basis in current science as far as protein recommendations and there's probably a lot.
TWiT.tv [02:01:45]:
More literature out there from like the manosphere than let me ask how much.
Leo Laporte [02:01:50]:
Protein I should be eating every day based on what you know about my health from all of those records and supplements I've uploaded.
Paris Martineau [02:01:59]:
Oh, it gave me the right one. Which is the standard recommended daily allowance for protein is 0.8 grams per kilogram of body weight which is about 0.36 grams per pound which translates to around like 50 ish grams daily for most sedentary adults. Oh, then it does say. However more recent research suggests this minimum may be far too low for optimum health. Not correct. Many Nutrition researchers recommend 1.2 to 2.2grams per kilogram. There's. Okay, I will say there's been.
Paris Martineau [02:02:34]:
I dove deep into this for my thing.
Leo Laporte [02:02:36]:
Yeah. I'm curious because I TR you if you tell you tell me how much protein I should eat you the.
Paris Martineau [02:02:43]:
There has been a lot, a lot of research on this topic and one of the strongest I called up the like leading nutrition and protein like academic researcher in the US like his job is to work with the top of the top athletes in like cutting edge research and he's like I'm constantly trying to pull protein out of their diets. He's like no, no one needs as much protein as they think basically. Unless you're like a very specific top tier athlete. So one of the, a lot of people have tried to prove this scientifically that yeah, getting 2 grams per kilogram of. Of body weight protein per day works for you. Which is insane. One of the most.
Leo Laporte [02:03:29]:
A lot of meat or something.
Paris Martineau [02:03:31]:
I mean one of the most compelling things I found was a large scale meta analysis that I believe they looked at some crazy amount of studies like dozens and dozens of them or maybe it may have been like 17 thing but they're really intense studies and they combined all this, they did all the statistical analysis and what they found was for the average person eating more than that recommended daily allowance, 0.8 grams per kilogram doesn't not confer any real benefits. The only time where you're gonna have like benefits in retaining like muscle mass for eating more than that recommended daily amounts of protein is if you are actively in a calorie deficit and you're engaged in consistent resistance training.
Leo Laporte [02:04:18]:
Actually, that's me because of Ozempic and it is one of the advices that most people.
Paris Martineau [02:04:25]:
But are you engaged in active resistance training on like a daily basis?
Leo Laporte [02:04:30]:
I lift kettlebells every day and swing them around.
Paris Martineau [02:04:32]:
Then, yes, it could be useful, especially just because something like people should probably be having the. The recommended daily allowance is the amount of protein you need to maintain your lean muscle mass.
Leo Laporte [02:04:45]:
And that's the issue for me because I of Ozempic, I'm reduced calorie because I can't really eat as much. And so, so what happens is not only do you use fat, you lose at least 25% of that weight loss.
Paris Martineau [02:04:57]:
And so I think in those.
Leo Laporte [02:04:59]:
That's why I've been lifting. Doing resistance lifting.
Paris Martineau [02:05:01]:
Yes, Because I mean, they also like. This is some advice I've given to a lot of people. A lot of people ask me this now is. Yeah, you can also. But you. Well, one, we're going to take this a couple steps for the average people on GLP, 1s are often the kind of target audience for, yes, have more protein. But it really only works if you, you are both in a calorie deficit and really engaged in resistance training, which a lot of people aren't. They did a lot of research to be like, if you.
Paris Martineau [02:05:29]:
I know, but most people don't do it.
Leo Laporte [02:05:31]:
No, it's hard.
Paris Martineau [02:05:32]:
If you are in a calorie deficit and not doing resistance training or not doing enough, it doesn't give you any real benefits.
Leo Laporte [02:05:39]:
And the thing is, now I wonder how much lead's in here because they.
Paris Martineau [02:05:42]:
Say, no, those are.
Leo Laporte [02:05:43]:
It's got three egg whites, two almonds, five cashews, two dates and no bs.
Paris Martineau [02:05:48]:
I was gonna say those. The Rx bars don't have any, like protein powder sort of added protein. It's. You're getting your protein from natural foods.
Leo Laporte [02:06:00]:
Yeah, but these are delicious, by the way.
Paris Martineau [02:06:03]:
The thing that I, whenever I spoke to these actual researchers, they're like, if you want to get more protein, go for it, I guess, but don't. Ideally, you don't really need to be getting it from something like protein powders or supplements.
Leo Laporte [02:06:16]:
It's really pretty easy.
Paris Martineau [02:06:17]:
Easy. It's pretty easy to hit like 50 grams of protein a day. You have like 50 chicken breast and some other breakfast.
Leo Laporte [02:06:24]:
Yeah, but it's the whole.
Paris Martineau [02:06:26]:
You don't need to be having 50. You probably don't need to be having 100 grams of protein a day.
Leo Laporte [02:06:31]:
I'm doing about 80, which is fine.
Thomas Haigh [02:06:34]:
I think it's fine.
Paris Martineau [02:06:35]:
I don't know. That's the thing is a lot of things have protein in it. You probably don't need to be supplementing with stuff. You should just eat real food.
Leo Laporte [02:06:42]:
What I do is I have, you know, I choose cottage cheese and. Yeah. You know, stuff that has. That's protein. I love peanut butter. There's stuff. I just choose stuff with protein in it. Anyway, I don't know how we got into this.
Leo Laporte [02:06:53]:
Oh, I'll tell you how we got into this. And Jeff, Poor Jeff. We're gonna. We're gonna end it because Jeff, we're gonna.
Paris Martineau [02:06:58]:
Yes, we're going.
Leo Laporte [02:06:59]:
He's in. No, no, no. I don't want to do this show without you. You're the heart of this.
Paris Martineau [02:07:04]:
We have to have a full show where Jeff was in a hospital bed.
Leo Laporte [02:07:10]:
And we've got.
Paris Martineau [02:07:10]:
That means you've got to end it early.
Leo Laporte [02:07:12]:
I think the real problem is that it's very, very, very hard to do studies of in vitro of humans. In vivo of humans. Because, you know, there's so many other factors. You cannot eliminate all the other factors that might change the result. So it's very hard to say. Well, that caused, you know, if he had the, you know, so it's all kind of speculative. Studies vary. You know, there's that nurses study, the long term China study.
Leo Laporte [02:07:37]:
There's all these studies. They all say different things. Just eat normal. You know what I like Michael Pollan's advance advice was it eat plants. No, eat food. Not too much. Mostly plants, I think was his advice.
Paris Martineau [02:07:55]:
I don't know. Just eat things that make you feel good. And ideally it's not that process.
Leo Laporte [02:08:02]:
I'm going to drink more whole milk. That's what I am going to do. All right, we're going to come back with sou.
Jeff Jarvis [02:08:07]:
Hospital cheesecake is surprisingly good.
Leo Laporte [02:08:10]:
We were waiting. I wanted to stay on the show until your jello came.
Jeff Jarvis [02:08:14]:
Oh, well, my food arrived about an hour and a half ago.
Leo Laporte [02:08:16]:
Oh, is it just sitting there cold?
Paris Martineau [02:08:18]:
We got to let him go, Leo. Do the ad. Let him go.
Jeff Jarvis [02:08:24]:
Apple pie.
Leo Laporte [02:08:25]:
Let Jeff go.
Jeff Jarvis [02:08:27]:
That's fine.
Leo Laporte [02:08:28]:
All right, well, just quickly join the club. Keep this Guy from dying. Every penny you spend on Club Twit goes to keeping Jeff Jarvis alive. Streak TV Club Twit. That's not true. Coming up in just about a couple of hours, we're going to have Micah's crafting corner. He is doing paint by number for all our club members. That should be a lot of fun.
Leo Laporte [02:08:52]:
I love our AI user group. If you watch this show, you really got to watch the AI users group. We get down and dirty. We actually use these things. We show you how to use them. Lawrence lrau did a wonderful thing on Anti Gravity a couple of weeks ago. We're doing all of that stuff. We do vibe coding, everything.
Leo Laporte [02:09:08]:
Lots of interesting conversations about AI on our AI user group. If you're not a member of the club, Please join TWiT TV Club TWiT. Also, I want to make sure that all the IM people take our survey because we want to make sure every show is represented last few days. TWiT TV survey 26. We don't know anything about you by design. RSS feeds. We don't know anything about you. One thing we do want to know though, is what you like, what you don't like, what you do for a living.
Leo Laporte [02:09:34]:
So that we can use that information to better program to you, to give you shows you care about. But also it helps us sell advertising and God knows we need help with that. So please, if you want to keep the show, drink your milk.
Paris Martineau [02:09:47]:
Milk.
Leo Laporte [02:09:48]:
Join the club and fill out the survey. TWIT TV survey 26. All right. I am never going to look like that. They put a very muscular Leo in the AI. Oh, that's the.
Jeff Jarvis [02:10:01]:
You said that as you were looking at me on the screen.
Leo Laporte [02:10:03]:
Yeah, I'm never going to look like that. No, Jeff and I are the same.
Paris Martineau [02:10:07]:
I will say we've actually during this show reached two new records. One, this is the first Twit podcast where a guest has been exclusively in a hospital bed. But two, this is. This is the first Twit podcast where a guest has tweeted from a hospital bed while doing podcast from a hospital.
Leo Laporte [02:10:25]:
Did you actually tweet? He can't be.
Jeff Jarvis [02:10:29]:
I want to know.
Leo Laporte [02:10:29]:
You can't put it down. Jeff, do you have anything you want to pick before you go?
Jeff Jarvis [02:10:34]:
No, I don't.
TWiT.tv [02:10:35]:
I don't.
Leo Laporte [02:10:35]:
You want to pick the Morristown, the fine care of the Morristown Hospital, Moral hospital?
Jeff Jarvis [02:10:40]:
Yes, I will.
Leo Laporte [02:10:41]:
Wonderful physicians and nurses. Nurses who help keep Friday or Thursday.
Jeff Jarvis [02:10:46]:
I have a. I think it's called a pick line.
Leo Laporte [02:10:50]:
Yeah.
Jeff Jarvis [02:10:51]:
No, you Don't. This is a. This is like, you know, an IV needle. Yeah, no, I know about the normal needle. Is that long?
Leo Laporte [02:10:58]:
Yeah. The pick.
Jeff Jarvis [02:10:59]:
This needle's about that long.
Leo Laporte [02:11:00]:
I've never had a pick. My daughter had a pick.
Jeff Jarvis [02:11:02]:
It's not fun and it goes way.
Leo Laporte [02:11:05]:
They leave it in there.
Jeff Jarvis [02:11:06]:
And then I have to once or twice a day I will be infusing.
Leo Laporte [02:11:12]:
The antibiotics into putting a little Cipro in their pick. Yeah. My daughter, because she. She thought she had Lyme disease. And they were. That was the treatment for Lyme disease, was a long term antibiotic.
Paris Martineau [02:11:25]:
Fun way. If you want to get into drugs, Jeff, you got a. You got a little line right in there.
Leo Laporte [02:11:29]:
That's right. It really saves a lot of time. It saves a lot of time. Plus needles. So anyway, Jeff, I hope you feel better. When do you get out? Do you get out tomorrow?
Jeff Jarvis [02:11:43]:
No. Probably either Friday or Saturday.
Leo Laporte [02:11:47]:
Oh, my God.
Paris Martineau [02:11:48]:
Oh, my God. I hope that you get out. I've one. I hope you have a speedy recovery. Just for general health reasons. I hope you're not snowed in.
Leo Laporte [02:11:56]:
We might be getting like.
Paris Martineau [02:11:58]:
We might be.
Jeff Jarvis [02:11:58]:
Oh, yeah.
Paris Martineau [02:11:59]:
Well, we don't know, but knock on wood, we might be getting like 16 inches of snow Sunday morning, Saturday night.
Leo Laporte [02:12:07]:
Yeah, I know that because salt Hank had to close early today because he said nobody was in line. It was too.
Paris Martineau [02:12:14]:
Well, it was. It was like seven degrees outside today. Yeah.
Leo Laporte [02:12:18]:
Yeah.
Jeff Jarvis [02:12:18]:
Interesting.
Leo Laporte [02:12:20]:
So he did not have the usual long line for his delicious French.
Jeff Jarvis [02:12:24]:
When I worked, when I was at Ponderosa Steakhouse, the biggest day of the year was always Mother's Day. Poor moms. Hey, mom, we love you. Let's give you a 99 cent steel steak and it's all you can eat.
Leo Laporte [02:12:36]:
Ma.
Jeff Jarvis [02:12:37]:
Yeah. You want some mushroom sauce? No, mom, you don't want. And the second. Is anything rained? People. People came out to the restaurant when it rained. You think it's the opposite? You don't know.
Leo Laporte [02:12:47]:
Yeah. You don't know. You really behaviors. Yeah.
Jeff Jarvis [02:12:50]:
Yeah.
Leo Laporte [02:12:52]:
Good news. Ted Cruz has left Texas before the winter storm.
Paris Martineau [02:12:55]:
So that's how we know that a winter storm is coming. Because Ted Cruz was spotted on swallows.
Leo Laporte [02:13:00]:
Come to Capistrano and Ted Cruz goes to Cancun. You know, the time has come. Winter storm. Paris Martineau, you have some picks, I think.
Paris Martineau [02:13:10]:
Yeah, we're not going to talk about my picks. We're going to let Jeff eat his cold pizza and rest.
Leo Laporte [02:13:16]:
You are a very, very thoughtful person. We appreciate that. We appreciate you. Thank you for coming and watching the show. Jeff will Be back next week in a fine, fine form. Hope I know.
Jeff Jarvis [02:13:30]:
Then I will. We will model my back brace.
Paris Martineau [02:13:35]:
Oh, God.
Leo Laporte [02:13:37]:
So no surgery on the spine? They're just going to let that kind of heal itself or.
Jeff Jarvis [02:13:42]:
Yeah.
Paris Martineau [02:13:43]:
Does it mean that your eye surgery is postponed?
Jeff Jarvis [02:13:46]:
Yes, yes.
Leo Laporte [02:13:48]:
Don't get old. Don't get old.
Paris Martineau [02:13:52]:
I'm gonna go out.
Jeff Jarvis [02:13:53]:
Age 40.
Leo Laporte [02:13:54]:
Yeah, yeah, yeah. Leave. What is it? Live fast, die young and leave a beautiful corpse.
Paris Martineau [02:14:01]:
It's a plan.
Leo Laporte [02:14:02]:
Yeah. Life is hard. Dying is harder. But nobody here is dying. We're going to be here for a long time. We hope you will too. We do intelligent machines every Wednesday, 2pm Pacific. Oh, I forgot, I don't say p.m.
Leo Laporte [02:14:20]:
or o' clock anymore.
Jeff Jarvis [02:14:22]:
Oh, what it is? That's so silly.
Leo Laporte [02:14:23]:
I am going to. No, I'm going to use the 24 hour clock. I'm going to convert my brain. I'm going to think of it as 1400 Pacific Time, 1700 Eastern. That is 2200 UTC. I'm going to do that in my brain one of these days. No more. No more post meridian.
Leo Laporte [02:14:42]:
Oh, the clock. That's ridiculous. I'm going to join the 21st century. You can watch us do it live on Twitch, YouTube. Tik. No X.com, we don't do TikTok. It's too complicated. Facebook, LinkedIn and Kik.
Leo Laporte [02:14:57]:
Also of course in the club Twit Discord, if you're a club member. Thank you, club members after the fact, on demand versions of the show available at the website twit tv. Im on YouTube. There's an intelligent machines channel. And of course you can subscribe on your favorite podcast. Fine. Paris Martineau at ConsumerReports. You're still working on that big expo, Jose?
Paris Martineau [02:15:14]:
I am. Although right now I'm distracted by a really good video. That pretty fly Pharisees guy just put in the chat. But that's.
Leo Laporte [02:15:21]:
He's good.
Paris Martineau [02:15:22]:
Something for.
Leo Laporte [02:15:23]:
I didn't put the story in, but I will mention this. You have longevity because according to the Daily Beast, radioactive shrimp are likely to keep popping up for months.
Jeff Jarvis [02:15:36]:
Wow.
Leo Laporte [02:15:37]:
So good news, Paris.
Paris Martineau [02:15:38]:
They keep going on, those little shrimpies keep on coming.
Leo Laporte [02:15:43]:
That's cause the plume, the plume, it's all down in the nuclear plume.
Paris Martineau [02:15:48]:
You know, sometimes you get in a call with your editors and you're like, all right, there's a nuclear plume and then it's all downhill from there.
Leo Laporte [02:15:57]:
I will mention briefly, to give Jeff something to do while he's in the hospital, the first issue of a brand new online zine called Game Poems. This is brilliant. Each one of these is a little game.
Paris Martineau [02:16:10]:
Well, because you know, Jeff just loves gaming so much.
Jeff Jarvis [02:16:15]:
And poetry. You hear me quoting it all the time.
Paris Martineau [02:16:18]:
Gamer and poetry lover.
Leo Laporte [02:16:21]:
This just shows you how far we've come. And I think one of this is maybe kind of because of AI. So recall a decision you've been putting off and then pluck the pedals and it will tell you. Keep pressed to play. And when all the petals are gone, then you'll know. Stay still to summon the wind.
Jeff Jarvis [02:16:42]:
Put me out of my misery, will you?
Paris Martineau [02:16:46]:
We're gonna just put some high dose morphine straight into the pipeline.
Leo Laporte [02:16:50]:
Jeff, Game Poems. No, this is a really cool site. These are all little games. Gamepoms.com issue number one. Look at all these little games you can play. They're all poetic, they're all philosophical.
Paris Martineau [02:17:03]:
We have to let Jeff leave before it's 7:30.
Leo Laporte [02:17:05]:
I'm torturing him. I'm sorry. Thank you everybody. We'll see you next time, Jeff. Feel better. Paris. We'll see you. Stay.
Leo Laporte [02:17:11]:
Don't stay warm. Don't get caught in a snow. Drift down a hill. They're deadly. When I was in third grade, my teacher, Mrs. Kelly, her husband was killed by a snowplow. Don't. Don't.
Paris Martineau [02:17:24]:
Poor Mr. Kelly.
Leo Laporte [02:17:26]:
Poor Mr. Kelly. Kelly, stay away from the snowplows, okay? And we'll see you all next week. The good, good Lord willing.
Paris Martineau [02:17:32]:
And see you in the snow.
Leo Laporte [02:17:34]:
Don't rise. Bye bye.
Paris Martineau [02:17:36]:
This is for you, Mr. Kelly.
Leo Laporte [02:17:37]:
This is for you, Mr. Kelly.
Paris Martineau [02:17:39]:
I'm not a human being. Not into this animal scene. I'm an intelligent machine.