Why ChatGPT Sucks at Poker - podcast episode cover

Why ChatGPT Sucks at Poker

May 29, 202551 minSeason 1Ep. 56
--:--
--:--
Listen in podcast apps:
Metacast
Spotify
Youtube
RSS

Episode description

ChatGPT is amazing at many things…but it can’t seem to figure out poker. Nate and Maria explain why, and talk about the implications for artificial general intelligence. Plus, they discuss Trump vs. Harvard, round two.

Further Reading:

Silver Bulletin: ChatGPT Is Shockingly Bad at Poker

Harvard Derangement Syndrome, by Steven Pinker

For more from Nate and Maria, subscribe to their newsletters:

The Leap from Maria Konnikova

Silver Bulletin from Nate Silver 

Get ad-free episodes, and get your questions answered in an exclusive weekly bonus episode, of Risky Business by subscribing to Pushkin+ on Apple Podcasts or Pushkin.fm. Pushkin+ subscribers can access ad-free episodes, full audiobooks, exclusive binges, and bonus content for all Pushkin shows. 

Subscribe on Apple: apple.co/pushkin
Subscribe on Pushkin: pushkin.fm/plus

See omnystudio.com/listener for privacy information.

Transcript

Speaker 1

Pushkin. Welcome back to Risky Business, a show about making better decisions. I'm Maria Kanakova.

Speaker 2

And I'm Nate Silver.

Speaker 3

Today on the show, we're gonna be talking about how chat GPT plays poker, i e.

Speaker 2

Poorly.

Speaker 3

If chat gipt is getting ready for the World Series of Poker like us, then it's got a lot of homework to do.

Speaker 1

Yes. And after we talk about that, we'll talk about Harvard two point zero aka the fact that foreign students are now being potentially barred from attending Harvard University. Harvard has sued the administration. Anyway, we'll be talking about what's going on with that and what the implications are for the future of the United States and kind of research development,

brain power and the US competitive edge. So for people who don't subscribe to Silver Bulletin, which you absolutely should, Nate and I both had poker posts this last week. Mine was about cheating, was about chat GPT, but his post this week was one of the funniest things I've read in a while about his attempts to get chat GPT to simulate a hand of cash game poker. And one of the reasons I mean, I found this amusing because I mean, first of all, it fused cards together

for an image. From the beginning to the end, it was pretty spot on in terms of being spot on wrong. But Nay, You're also someone who is constantly writing about how good AI is at so many things, and so it was funny to me to see it fall so short of something that actually tests intelligence as opposed to being able to kind of pull together a lot of things and spit out an answer.

Speaker 3

There are some reasons why I'm just sid in testing chatchipt on poker, but they kind of fall into two big buckets. Right, What is it like poker played in the context of a real hand or in this case, a real text based simulation I guess of a hand. It really does require quite a few skills, right. There is the pure math part of it. What is the equilibrium, the Nash equilibrium, the gto strategy that you're solving for.

There's also making adjustments for how other players play. There's the conversation you're having, the physical reads that you're getting their stuff, like knowing what the rules are right, which might seem trivial, but we've yeah, chatchpt also.

Speaker 1

Does not know.

Speaker 2

It seems to be no, it is fun to look.

Speaker 3

I mean, I'm sure we've all had hands where we like misread a board or I you know, I had a big hand in the main event at the when I play back in Florida back in April or whatever, right where like a guy had like a twenty five thousand dollars chip that was like almost the same color as the felt, which is not good chipkit maintenance, by the way, Seminal hard Rock. I like the WPT, but that chip kit it's got to be improved, right, But yeah,

you're tired, you make mistakes like that. But and you also have to have like a lot of short term memory, which sounds trivial, but you you know, you want to remember how you arrived at this current spot in the hand, right, you have to keep crack of all these stack sizes, which again seems trivial, but like it's a good test of general intelligence of a certain type, especially a live poker hand, like I'm asking.

Speaker 2

Chat ept to simulate.

Speaker 3

Right. The other reason is that like I don't think engineers in Silicon Valley are trying to optimize it for poker. And the reason that's important is because like, look, there are benchmarks like math Olympiad problems that these LM's large language models like CHATCHPT, Claude, et cetera compete on, right, And they're like, well, we be bragging that we can now, you know, beat all but the Nobel Prize mathematicians on x percentage of math problems and things like that. And

there are a couple of issues with that, right. Like one is that, like, clearly, if you train a transformer model, a machine learning model on a particular type of problem, like it can do fairly well or very well often, right, Like poker is to a first approximation, and it's an important approximation. Right, But if you have poker dedicated tools, I wouldn't consider a solver and AI that's a technical distinction I think might not be that important for our listeners.

But like, you know, if you want to use computers to play very good poker, then they can play very very good poker. Right. The question is like, can you take a text based model and have this organic property where intelligence emerges and converts from the data set towards superintelligence without training it on poker specifically. And the answer, at least as of last week when I did this, is no, it fails miserably.

Speaker 1

Yeah, and I think it's actually important. One of the things that came to my mind when I was reading your piece was that poker has been a benchmark for AI development well before LMS, Right, It has been kind of the gold standard that teams all over the world have been working on for decades because it is a much more complex game in many senses than chess, then even go, and it's a game where, you know, if computers can actually manage to outthink human players on a

broader scale, that would mean something in terms of kind of broader intelligence capabilities. And before LMS were developed, that had not happened. Right, There had been computer programs, kind of AI algorithms that have been able to beat heads

up one on one opponents in poker. But when it came to full ring games, So for people who don't know poker, that means basically, when you put it in a rich environment with you know, six players, eight players and you have a computer there, it was still not quite there. And one of the you know, one of the reasons why poker so interesting is it's not just math, right,

especially when you have multiple players. There's so many dynamics and there's actually a very thorny problems for AI, which is that if you program something to follow an algorithm right to be GTO game theory optimal, then it can be exploited by a human if the human figures out the GTO strategy. Now, if it were another human, then you'd adjust immediately right the moment the human adjusts, the other human would adjust and kind of figure it out.

But if you're an a if you're a computer, and you build in that adjustment parameter, what ends up happening is that the model adjusts too much and too quickly, and so that lacks kind of some of the nuance and flexibility that marks that's kind of the hallmark of the great poker players, right, kind of that ability to

into it when it's time to slightly deviate. And I mean, it would be terrifying but amazing if a model could spontaneously, like an LLM, could actually spontaneously figure that out and so far, I mean, instead instead of doing that, Nate, you know, your model decided that the person who won the hand actually lost the hand, right, It just got basic things completely wrong, and you just you know, obviously we've talked about pe doom a lot on the show. I think you and I are both worried about that.

But when I see this poker stuff and like, oh man.

Speaker 3

Yeah, I don't know how much to update. But look my say in the post, I mean, like my appreciation and expectations for large language models have increased a lot over the past year. I had a pretty good experience working with them when I was building the NCAA tournament model that we host silver bulletin and just helpful kind of Swiss army knives or like annoying data related tasks.

It's late at night, you forget the Stata command for something, You're like, right, give me the right command and write five lines of code, and like usually it works, and it kind of now shows you like the chain of reasoning and like what it's thinking. Right, And but you know, people are saying that we're going to have like AI revolutionizing all programming or you know, well, first of all, the big claim is going to have AGI, meaning that a machine. It's a little ambiguous when people say, does

this mean like chat GPT itself is AGI? Right, Well, clearly not. It's very far from human being in some things, right, Does that mean that a combination of transformer or machine learning based technologies is or zooming out the definition of AI. It's really AI still, right, so that you have chechiptiza seventy percent of things well, and then you specialized applications for twenty percent, and then there are ten percent where you're not doing it very well. Like that would still

be a big achievement, very disruptive achievement. I'm not sure it would kind of quite count as AGI. Like one thing you could do is, like some of these new models will say, okay, I can detect that you ask for something that requires mathematical precisions. So let's take a simpler example. Right, Let's say I want to take the

distance between pick two cities, Boise, Idaho and Tampa, Florida. Right, that's pretty easy to calculate mathematically if you look up the latlong coordinates for Boise and Tampa, and then there are various formulas that.

Speaker 2

You can use.

Speaker 3

Right, So a pure data set where it's just crunching texts might not find that answer because there's not necessarily a lot of examples of like what's the distance between Boise and Tampa? Right, Whereas you would have that for like La and new York or something. However, you can set up kind of agentic meaning agent ais where they're like, okay, well this person asked a question that triggered a routine

of mine. Where now I'm going to go and I know to search the internet for the coordinates or maybe it's stored somewhere right of Boise and Tampa. And then the formula. I know the formula and I can apply it, right, so like you have the scaffolding. It's sometimes called a different processes on top of one another. And I guess that kind of I guess that counts as I mean, it's kind of how humans solve problems, right, They're like, oh, and now I have to go look something up, right,

It's still I think subtly different. So one implication I think this has is artificial general intelligence versus super intelligence ASI, sometimes called where there's like an explosion of intelligence just from kind of mining text or other data sources. I have actually become a little bit more skeptical of that, or I think it's presumptive to assume it can find that.

I mean, like, let me give you another example, right, like one I had an idea for either a silver bulletin post or maybe even a risky business segment, right where like I asked, chatchipt deep research, design a meal with foods that have no clear precedent in other countries, right, huh? And I thought, and the idea was that, like, you know, at some point somebody invented pizza or spaghetti or whatever else, right, or or I.

Speaker 1

Mean at some point somebody invented bread and figured out that you can take flour and like bake shit with it. It's amazing.

Speaker 3

Yes, you have to be able to buy these ingredientstead of Trader Joe's or a Whole Foods. Right. I was just going to like hire professional chef and have a dinner part and ask how these creations went, right, And it's like and all they recipes were like here's salmon with miso paste, right, really creative and things like that, right, and like it didn't have any ability to like extrapolate beyond the data set. I'm sure these preparations are fine.

It was like their little chefy, meaning like things that like have one too many ingredients, right, and like and I don't, yeah, look so bad chefe.

Speaker 2

Yeah.

Speaker 3

And meanwhile, I've been like I've been interviewing candidates for like a sports assistant position at Silver Bulletin, and one of the questions is, how are you using AI? These are very bright I want to say kids. A lot of them are are recent college grads, you know, your early twenties to mid thirties, basically right, and very technically proficient people. And asked them how they're using a GI and their experiences like kind of similar to mine. They're like, Yeah,

for a certain task, it's very helpful. For certain task, it's somewhat helpful for certain tasks, not helpful at all. I can get you in trouble, right, which again contrast with the experience of people at the AI l apps to say, yeah, it's going to like displace all programming within a year or two.

Speaker 2

I mean, I know you should weigh in here.

Speaker 1

So I have an interesting kind of anecdote to add to this that's not poker related, but that talks about kind of some of this nuance and complexity and the fact that you know that all of these lms still are wanting in certain major respects. So I had the chance last week to speak to the CEO of one of the major AI companies. I won't name which company

it was, but one of the big ones. And someone had asked him earlier, what was something that was surprising about the way that this AI functioned, and he said, you know, one of the mistakes that people is kind of they trust it, And what we know is that it's remarkably accurate ninety percent of the time, right, and then ten percent of the time it isn't. But the ten percent is getting harder and harder for humans to spot because it doesn't make errors in a way that's

intuitive for people, right. It makes errors in different ways. It's like computer errors AI errs, that's not the way that the human mind makes errors. He's like, but like, it's fine because we have really smart people who can figure out where it's making errors and can work with

it and can kind of calibrate it and help. And so my question, my follow up question, was, Okay, you know what happens when the new generation that's being brought up with this AI comes up and they've been educated with it and they've been using it the whole time, and they didn't necessarily get the same education that we got, right, because they aren't the incentives are different, and they're not

getting the deep knowledge. They're not actually able to figure out what are the programming errors because this person was talking that, you know, some of the biggest potential of AIS and like programming and biology and kind of those those types of things, and I said, well, what if that person doesn't have the neuroscience and the biology background

or the programming background. And what the CEO said was, this is a major problem, and what we're hoping is that we can develop even more advanced AI to help and to help fix the problems that AI is causing. Because he was like, yeah, this is a big problem and we're not quite sure, like we're hoping for the best, but we don't know. And the fact that this was our conversation didn't exactly inspire me to heights of thinking this is going to replace intelligence or become AGI or

what did you call, Nate the other ASI artificial super intelligent? Yeah, artificial superintelligence. So it didn't inspire me that that was going to happen, and instead it was like an uh oh moment where like, what happened when the humans who are kind of helping push it along and make it better?

When that generation, like when they age out and they're new people out there who who lack that sort of skill and expertise and who grew up with AI that is ninety percent accurate or even ninety five percent accurate, but they can't. It's harder and harder to spot that five and ten percent. And that to me was actually, like, it's a worrisome thing, and it's clearly worrisome to the people who are developing these technologies as well. And so that I think is something that dovetails with what you

were talking about here. What happens Nate if all of a sudden, all poker players are training with chatchipt right, and that's that's actually great for us, but not great for the future of poker.

Speaker 3

And we'll be right back after this break. I use large language models for tasks that I have strong domain knowledge over and could do myself. But they're a labor

saving device. Often they say substantial labor, right, Like I use them to copy edit articles I posted the newsletter, and I you know, I'm a fairly proficient user of the English language and like for programming, right Like you know, if you're like using AI to like design a website and some of the functionality breaks, right, that might be okay.

Speaker 2

You can patch it and fix it. Right, if I'm doing things.

Speaker 3

Involving modeling, I'm trying to project the value of basketball teams or football teams, or basketball players or football players. Right, and there's a bug where all of a sudden, like one of the worst players in the league is rated as being very good. Right. I mean, and you see this in poker, where like you know, AI is relying. You get punished in poker for sloppy application of imprecise

application of heuristics. Like one of the things like the AI did so at first, I had it simulate just one hand, and it had like a backstore for each player. It fucked that up, right, then like Oka, I'm gonna use deep research now and have it simulate like an orbit of eight hands.

Speaker 2

So that was a little better.

Speaker 3

It took more compute time, right, Some of the hands were decent, but it you know, it didn't know how to like read aboard correctly and keep track of stacks. But like, but when it was over able to overcome those errors, it also had poor strategy and it gave

various excuses. One thing it said is like, well, the quality of text based content on poker on the internet really sucks, Right, it's aware of that is it says it's worse than go or chess, although when I talk to people who know how AI's played chess, it's also a disaster. Even something like wordle I heard from a user.

You know you think wordle is a word game and it does very well, but like it can't quite figure out the strike sure of the problem, right, it kind of was learning a little bit more by by a rote And to be clear, like this is a very good strategy for like many types of problems. And also if you were to say, okay, here's a bunch of we bought data from PIO solver or gto wizard. These are solvers if you don't know listeners or a database

of high stakes hands. Like if you trained it on that and then had GPTs say okay, poker question, I'm going to call them especially trained database, then it would do well. And maybe if people are criticizing its for on poker, then the labs will start to do that. Right, but it's still kind of like slightly cheating from the

standpoint of super intelligence. Also, it says like when you asked me to do a whole bunch of things at once, right, So I told it, give me characters, give me eight characters who have like they were all kind of like ethnic stereotypes, right, it was.

Speaker 1

They were very funny.

Speaker 3

It was kind of yeah and great. It was kind of at the one hand, so half the players were women. That was very feminist of it, right, But then they're all ethnic stereotypes of different kinds, right, and so it was kind of both woken and unwoke, right the opposite. Yeah, it was kind of good about like matching players playing styles with with their actions, but like, but it couldn't keep track of stack size. It's like, yeah, keep track

of stacks. I have to store a bunch of stuff in memory, and then when you ask me to do a whole bunch of things at once, right, and like you know, and even including like so in the hand I show a sylver bulletin it like it misreads the board, doesn't recognize that a pair of nines has a higher

two pair. If you ask it that out right, then it thinks about it more and gets the question right, right, But like it loses track of things in the context of these scaffolding situations where it has to keep track of a bunch of things at once and it doesn't know it doesn't know what to prioritize right. Like you know, you can make a lot of mistakes in poker if you don't know which hand be twitch hand. That's a more elementary mistake than anything else.

Speaker 1

Absolutely, So there are a few things that stand out to me here.

Speaker 3

One.

Speaker 1

I mean, it's a computer, right, like we humans have working memory capacity problems. It should be able to keep all these things like this is what it's good at. But the problem that you're kind of that you're illustrating is it doesn't know how to prioritize right, and that that is actually that is a major problem. And also

when we keep saying thinking. But one of the first things, so when I was just learning poker, one of the first lessons I had with Phil Galfon, who is one of the people who kind of I worked with and who taught me a lot. I remember very early on he said, you know, I can give you a bunch of charts and a bunch of outputs, and you can memorize them and you'll be a very decent player very quickly. Like I know, I can give you this shit, you can memorize it, you can spit it back at me.

You'll be fine, He's like, but I don't want you to do that, because that's going to make you a fine poker player, and you'll do fine in the short run, but you'll never be a great poker player because you have no idea why you're doing it. You're not actually thinking,

you don't understand. You've just done a rote memorization, which is might make you more money in the short term, but in the long term is actually going to be detrimental to you because you're going to do stuff unthinkingly because you've memorized it.

Speaker 3

He said.

Speaker 1

What I want you to do is instead think through things and figure out, Okay, why right? For every single play, why am I doing this? Why would I play this hand and not this hand? Why would I raise this type of hand and not this type of hand? Why?

Speaker 3

Why? Why? Why? Why?

Speaker 1

And that's something that to this day has stayed with me because it's such an important thing to remember when you're making a decision. Right, And this doesn't even have to be poker, It can be any sort of decision. Why am I choosing this action? Why is it better than all of the other actions? Right? And if you even if you feed you know, piosalvereign gto Wizard outputs

two lllms, they'll be doing some rote of memorization. You know, at this point they're not understanding the why, and so that will lead them still to make big mistakes, which happens when you just learn solver outputs, no matter how sophisticated those outputs might be and how correct they might be in one specific hand, in one specific spot, if you don't understand the why, you're either going to overgeneralize it right or misapply it like you're going to screw

it up because you don't understand the underlying reasoning. And there's the difference between looking like you're thinking and actually thinking. And yeah, you know, as a human you have other pitfalls, and even if you're you know, thinking through things, you can mess up, So you mess up in different ways. But I think that's a really important distinction and probably one of the reasons why AGI is not as close

like poker. Actually illustrates in a very practical way, why the notion that you can just have a lot of data and all of a sudden have this you know, flurry of insight might not be as I guess, as intuitive as one might think.

Speaker 3

When one thing, Chatchip you told me when I asked to like audit itself, is like, well, poker's very difficult

because it's adversarial. You know, understand that there's some game theory there, but it's adversarial, and like there's no one I mean, there is like I guess some superstructure of like a solution for all poker hands if you get very zoomed out, right, but like but like subtle things when you have to calculate like an adversarial equilibrium on the fly basically, and it's that's very difficult.

Speaker 1

Yeah, I mean think about not even poker, but think about trying to do a game theory solution for multiple players.

Speaker 3

Right.

Speaker 1

It's hard enough trying to do a payoff matrix for two players when you're really trying to think through it and you're trying to think of all of the different payoffs and figuring out, you know, how to exactly do you wait them, how exactly do you structure that. Now when it's three players, four players, I mean, it's incredibly difficult to do that accurately. And so even if we just zoom out from poker, like this is just an incredible a very tough problem, even if you understand game theory.

Plus I mean, I think that anyone who has used solvers, and anyone who's talked about this understands anyone who doesn't even play poker but has worked with algorithms, the common saying garbage in, garbage out is absolutely accurate.

Speaker 3

Right.

Speaker 1

The solver works based on what ranges of hands you as a human put in there, right, and what responses you allow or don't allow, right, And yes, it will come up with an equilibrium strategy. But if you were wrong, right, if there's a player who's playing a totally different range, if there's a player who's playing a totally different strategy,

all of a sudden, your solution means nothing. And as a human you can figure that out, and you can kind of make adjustments from baseline, right, if you understand what the baseline strategy as you can adjust as an LM, as an AI who is learning that but doesn't kind of have that nuanced experience at least at this point.

I don't pretend to know what AI is going to be capable of in you know, in five years and ten years, but at least at this point they're not capable of making those sorts of nuanced extrapolations and figuring out, Okay, what were my inputs accurate, right, or were my inputs not quite accurate.

Speaker 3

Look, one thing human beings are good at is that human beings are relatively good estimators, and chatchipt can be sometimes if it's like, Okay, take all this text on the web and kind of give me the average of that,

Like there are some applications where it's pretty good. But like, you know, the other day, I was getting getting a drink with my partner and a guy who looked like he was homeless comes in and like hands like the hosts slash barked like a note saying call nine to one one or something like that, right, and you kind of have to make this judgment call about like is this like a paranoid schizophrenic or is someone actually in danger?

And like I think was pretty clear. I mean they're probably you know, two bartenders and eight customers clear to everybody. Like this guy I think is not like acutely in danger or anything. But I'd never quite experienced situation like that before. And like the fact that we can use like our general intelligence about like human behavior Yeah, Like I think everyone made the right decision that we're not going to call the police either to rat on him

or to say he was in danger. And like you know, and that kind of thing is is is hard for language yells to do, you know.

Speaker 1

No, I mean I think in general, like this just illustrates a really good point, which is that humans, even not the smartest humans, are much smarter and much more capable in very basic ways that we take for granted

than a lot of kind of super intelligences. Right, Like it took forever for a robot to be able to pick up an egg, right, It's something that we don't even think about, but this was like a robotics problem that was absolutely unsolvable, Like how do you get a robot to pick up an egg without without breaking it? And we do it without just without thinking about it.

And we make judgments all the time, just silly judgments that we don't think twice about that are so easy for the human mind, but that are that we don't even realize we're making right things about safety, things about just perceptions of the world. So I think that you know, humans are much smarter than than we think in like in dumb ways, if that makes sense, Like in waste that seemed like they're not hard problems, but those are actually some of the hardest problems to solve.

Speaker 2

I just think of.

Speaker 3

There being four levels, right, Like one more AI performance better than any human, two where AI performs better than like all but expert level humans, Three where AI gives kind of a passable substitute performance but like you know, not quite professional high grade level, and four where it's just inept right to the point of being comically inept.

And like, you know, my heuristic is that like within a few years we might have roughly an even divide between those four and I'm counting, by the way, tasks that involve manipulting the physical environment, which I think AI will mostly be pretty bad at. Right, I don't know how much I should extrapolate from the poker example. It was just so incongruent with these predictions of like imminent AGI, right, where a year ago would have said, okay, yeah, of

course suck at poker. Right, it's not really trained on this and it's a hard problem, and ha ha ha, Right, But like if you want to have these complicated, like structured, nested tasks, like I'm much less worried now in periods of oh, let's say five years of AI like being able to build like atiscal model to like forecast elections or the NFL or whatever else, because that like just requires like a superstructure of lots of little tasks, all of which are fuck upable, And if you get the

superstructure wrong too, then you're kind of just drawing dead. To use a poker term, right if any if you have to chain together twenty steps and any step, there's a ninety percent chance you get it right. Well, point nine to the twentieth power means you're almost.

Speaker 2

Sure to fuck something up.

Speaker 1

Yeah, fuck upable is a really great word to describe this. Nate to kind of to sum this up, you did ask chat GBT why it was bad at poker? What did it tell you?

Speaker 3

Yeah? No, I mean it's said various things. It said, Look, you stressed me out by having me having me have to make up these players and things like that.

Speaker 1

Trust it out, You trusted.

Speaker 3

It said the data that it trained on on the internet is which seems realistic.

Speaker 1

Yeah, it's a data problem. It's a yeah, it's not me.

Speaker 2

It's you. It's it's complicated.

Speaker 3

Inform about like how like it doesn't realize how important stack sizes are because it's just a bunch of tokens. Like all the input you give it is generated into tokens, right, and like the word the is not very important. The fact that Maria has fifty two thousand dollars in chips is very important, right, And it doesn't know how to distinguish those from its transformer architecture. So like, look, it is good at like it is good at verbal reasoning.

Like again, I I, you know we are talking about we don't other segments where we're like amazed by AI progress. Like it's very good at verbal reasoning, or at least faking that. But in ways I if it's faking, it's doing a pretty good job. Right, Like it's very good

at verbal stuff for the most part. Right, And AI is traded on poker, are very good at mimicking Selver solutions, right and like and so like and machine learning can get you a lot of ways when you have good data, but like when you don't have the data in the training set and it's not seeming to extrapolate very well, and then these complex tasks that require more and more compute like, I have not been impressed by deep research, which is where it goes away and says I'm going

to perform several tasks for you that require deeper thought, right, and then it takes fifteen minutes you come back and like, are you fuck this up? Right? And like, anyway, so I think it has affected my priors a little bit.

Speaker 1

Yeah, that's that's really interesting. And I love, by the way, Nate, that you know, even though that this is an AI which is supposed to kind of be better than humans in so many ways, that use such human excuses for why it fucked up. You know, you stressed me out, you gave me that data. It's not my fault.

Speaker 2

This is hard.

Speaker 1

And on that note, let's let's take a break and talk a little bit about Harvard and the other type of intelligence, human intelligence. Nay, We've talked on the show a few times now about kind of the crusade that Donald Trump seems to have against higher education, specifically the Ivy League and even more specifically Harvard University. Last time we spoke was when basically the administration had threatened to pull funding if Harvard didn't comply with a certain set

of demands. Harvard said fuck you and sued and yeah, and tried to go that route, and the funding was frozen. And then the administration last week, so we're taping this on Wednesday, the twenty eighth of May. So last week, the administration decided that it wanted to punish Harvard even more because you know, they didn't like the fact that Harvard wasn't complying, that it was being defiant, that it was taking them to court, and so said, hey, Harvard,

you can no longer have foreign students. Anyone who has visa's effective immediately. Those are going to be revoked. So Harvard's class of twenty twenty five that's graduating next week is going to be your last basically foreign student class, and we're not going to be granting visas anymore to those students. Harvard has sued again, so now we have a second set of lawsuits. The initial court cases have

frozen that. But then after that happened in the last few days, the administration said Marco Rubio actually said that we are going to freeze interviews for all foreign student visas, not just Harvard. So they not only you know, went after Harvard, but doubled down and said, you know what, you can sue us and you can freeze this. But if we don't grant the interviews, last laugh is with us.

Speaker 3

Yeah.

Speaker 2

Were they likely to win the lawsuit or are the likely to?

Speaker 3

Yeah?

Speaker 1

I think so. I think that Harvard is likely to win the lawsuit. Yes, However, from the lee I'm not a lawyer, from the legal analysis that I have seen, they're likely to win the lawsuit. But Nate, let's remember that that requires first of all, a lot of times, right, because how many times can we appeal this? And ultimately, well, if it goes to the Supreme Court, it ends up being these days less of a legal issue and more of a political issue.

Speaker 3

Look, I don't think the Supreme Court is likely to be very sympathetic to Trump on this type of issue, in part because, like you know, clearly they are somewhat transparent about like they are not doing it for like some security concerning and they're doing it because they don't like some of the speech that Harvard is making. But it's almost certainly First Amendment protect.

Speaker 1

Yeah, exactly, and I actually think that the So I don't know if you read Stephen Pinker's op ed in the New York Times, I don't remember what the op ed was called, but he termed this Harvard derangement syndrome, right, kind of playing off of Trump arrangement syndrome and full disclosure. Steve Pinker was my undergrad advisor, so someone I know, well, someone i'm you know, I'm still close with, so I'm very sympathetic. But I thought it was an excellent way

of framing what's going on. And he is someone who has attacked Harvard. He's a tenured professor at Harvard, but he's attacked Harvard many, many times for not standing up for free speech, for not protecting conservative viewpoints, for being a little bit too woke. Right. He has been one of the main people who said, hey, like, we have an issue here. But this his op ed was, this ain't the way, right, this is not the way to

solve this problem. And you're actually attacking the students who are some of my most critical, thinking, least woke people who are on campus.

Speaker 3

But yeah, look, I mean I kind of want to zoom out and say, like, what is the equilibrium here, Like what is a Trump administration trying to achieve? What

is Harvard trying to achieve? Like I support maybe some listeners will get pissed off, right, you know, I think clearly some of these colleges are not complying with like the spirit and maybe the letter of the Supreme Court's affirmative Action decisions, right, and so like, I wouldn't mind if like the Trump administration's like aggressively enforcing the law on those or things like diversity statements, which are enforced

political speech. Although Harvard's a private organization, so it's a little bit more complicated, right, But like I you know, I don't it's such a blunt tool to like target

foreign students. And even if you hate Harvard, you know, as a patriotic American, I want the best and brightest from all around the world to come here and contribute to our economy and pay our taxes and everything else, right, And like it's just like purely destroying the value of these very bright students from helping our economy.

Speaker 1

Yes, it's destroying our intellectual capital. So, by the way, when we're taping this right now, I'm actually in Hongo, and I just found out today that several universities in Hong Kong have offered blanket admissions to all students who can show that they were admitted to Harvard and can no longer go because they are a foreign student. They said, you don't even have to apply, you can just come here, and these are some of the best universities in Hong Kong.

And that's super smart, right, Like, universities all over the world should be doing stuff like that right now, because what the administration is doing is kind of bankrupting at the future of kind of intellectual development in the country by saying, oh, foreign students, you know, we've we've just frozen all these applications. So to me, like that's super smart, right, Like that is the we've we talked about this several months ago, Nate, Right, what can China do to capitalize

on the moment it's doing it. It's being like, come here, like you don't even have to apply, you don't have to do anything to show us the letter that showed you were accepted and welcome. Come.

Speaker 3

Yeah. I went to London School of Economics for my junior year and if I were them, i'd be doing cart wheels. And they always got a lot of foreign students.

Speaker 1

Right absolutely, I mean I think everyone should be doing this right now, and I just do not see I mean a lot of these things. At this point, it just derangement syndrome seems right, because there does not seem

to be any strategic value to this. There's not any any value in terms of trying to get at the types of problems that the Trump administration says it wants to get at, because it's not even addressing that right, Like at the originally anti Semitism was the pretext for this, and this is just like so far divorced from it. Everyone is having their visas revoked or not granted, or their interview is not even granted. So it's just, you know,

the pretense is now crumbling. We always knew it was just a pretense, but at this point, it's just not accomplishing anything other than to further could of create this absolute chasm in higher education and future brain power, future research ability, future creativity of the United States basically mortgaging its future right saying that we were going to give up one of the biggest, if not the biggest edge that the United States has had historically, which is kind

of this brain power creativity place where people can can develop and find support for for for their ideas, and now that is not possible and if this actually stands, right, if foreign students are not allowed into US universities, because right now, like I said, this interview process being paused for everyone, not just Harvard. I mean, that is going to be just detrimental in so many respects to what happens here.

Speaker 3

Yeah, the next So Harvard is not a very sympathetic case for reasons that are partly they're full right, but like you know, so far, the Trump admin has been pretty smart about kind of which schools they they pick on. I mean, I don't know what their in game is here or what Harvard's is really you know, JD. Vans

had it tweet. I don't know where he was like, he's oddly transparent sometimes about his thinking, and he's like, well look just like, look, these colleges aren't serious about complying with the law citing like the students for fair admissions. I think it's called which was the affirmative action decision, which I do think they've kind of not been serious about complying with in some cases case by case basis, and so like, unless we kind of show real muscle,

then what can we do. We have to fight them because they're in the wrong, and so we're overreaching. It is a subtext of that, right, which is like maybe not crazy and they have a real leader because Harvard isn't that sympathetic perceptions of higher education that's gone way down, so it's like harder to marshall public opinion on the side of this, right, and like and like, you know,

I saw some Harvard professors on Twitter. Ever we can be like this is the greatest threat to American I'm like, please don't say that, because it's it's it's maybe more of a threat to American competitoris and not sure it's a threat to democracy per se.

Speaker 1

Right, No, I think there are bigger threats to democracy.

Speaker 3

But like, but what's Harvard's end game? Like I think, ok, if they could go back and tone down some of the things say down for the past ten years, and I think they maybe would, but like now, I mean, if they back down, then they look weak and that also might hurt their recruitment own nothing, they'd have real problems. But like, what if you're the president of Heart What's name Alan Garber? Is that right? Yep? If you're Alan Garber, Maria, what do you do right now? And you know Harvard?

Speaker 1

Yeah, yeah, I mean, you're in a horrible position. I think you're doing what you can, which is trying to stand up to the Trump administration. So Alan Garber is trying to actually walk a very fine line. He's trying to stand up to Trump, and he's you know, spearheading these lawsuits, and he is leading committees into the claims of anti Semitism and kind of and discrimination on campus, the anti wokeness. So he is actually trying to address

all of these things in different ways. And I have no idea, you know, how successful that will end up being, but it is. It's a it's a tough you know, he's been put in a really tough spot because obviously

he hasn't been president for long. The last president departed on lessons tellar circumstances, as did a number of presidents of Ivy League institutions, and you know, you're picking up the presidency at a very difficult point in time where, yeah, you have to acknowledge that Harvard has made mistakes, right, not just Harvard, a lot of universities I think were trying to be I don't even want to use the

word woke, because I don't think that's accurate. Here but we're becoming, you know, places where people were afraid of

voicing their minds. And I wrote, I think almost a decade ago, I wrote a piece for The New Yorker about basically, you know, the problems in academia that come from not having enough conservatives right that it actually has trickle down effects in research and the types of things that are being researched in the findings that are allowed to be voiced, and that it can be incredibly problematic. And so I think you need to acknowledge that, and

those are things that need to be rectified. At the same time, you are currently fighting this battle against you know, Donald Trump, who's trying to destroy higher education in the United States, and so that is a you know, overriding priority, which doesn't dismiss all of the things that have gone wrong and all of the things that are wrong with academia, but it's just a it's a kind of this fine balance where you're like, yeah, I know I'm not sympathetic

because I've fucked up and I've done all of these things. And yet right now I'm one of the people best positioned to fight this, and so let's just try to figure out how to prioritize this and how to not spread ourselves tooth in so that we can actually protect higher education and then start to remedy the problems that we have internally, because I do think those are problems that need to be remedied, right, Like, you can't just be like, well, now we're all united against Trump, so

let's forget we ever did anything wrong. I think that all of these things you remember, Nate, like less than you know. Like six months ago, I was railing against Harvard because I was really pissed at the way they were reacting to a lot of things. Now I think

now I'm supporting it because they're being attacked. It's the classic psychology thing, right where you can actually unite a lot of people who who criticize you when you attack, right, all of a sudden, these various factions, when they feel these life threatening attacks from the outside, all of a sudden, these factions start uniting. And I think that's what's happening

at Harvard right now, as it should be. But we still have to realize that, you know, there are all of these different issues and it's a really I mean, it's a shit show. That's a technical term.

Speaker 2

Nate yeah, Look, I'm not afraid to use a term woke.

Speaker 3

I mean, you know, critical race theory and intersectionality, all these things kind of come from an extremely academic context, right, And Look, the thing I'm probably most concerned about is the fact that I think the quality of academic research is often has lots of problems, but like political bias in predictable directions is among those problems. I'm trying to think of a politically experienced solution that the Trump administration

might like and that people would find tolerable. On the Harvard side, Right, we'll collect data on the political affiliation of our students and faculty. We will make concerted efforts to I'm not going to try to frame this right conservati efforts like higher conservative scholars. Right, maybe you have like you know, if you have an ethnic studies department, you have a conservative stay.

Speaker 1

No, I mean that isn't that you know exactly what people have been trying to reverse by kind of with the affirmative action rulings, right, that you have quotos for certain types of people. So I think that the better way to do it is basically having blind hiring, just

like you have blind admissions. You are willing. You just don't you don't ask any of that, and you don't disqualify people because they have conservative leanings, right, and that it's not even it's not a thing because you don't really care, Like I don't care if my uh, you know, neuroscience professor, Well maybe not neuroscience as insofar as I study psychology, but if my astrophysics professor is, you know, what their political views are, as long as they correctly,

you know, as long as they're one of the best astrophysicists in the world. So so I don't think that that's you know, that what you're proposing is affirmative action for politics.

Speaker 3

I'm just saying, like it's slightly stupid, but like maybe maybe jd.

Speaker 2

Vance would like it.

Speaker 1

But but you're saying, like do something that he would like. Columbia tried to do everything and had all of their funding frozen anyway, right, So like this is the problem that like actually doing trying to appease is not the correct play here. I think that's actually the completely wrong game theoretical way to respond here. That's because it's not working.

Speaker 3

And that's a meta thing across everything that the White House does. You see it playing out Ontari's too, where on the one hand, they reverse course and undercut themselves all the time, right, you know, and they're sloppy enough where they probably maybe if they're doing this right, they could have had more likelihood of success in the courts than they have. Right, on either hand, they don't necessarily

hold up their quid pro quote. This is kind of very transactional and explicit, right, So they're not even like presenting clear enough information to make themselves easy to bargain with, even if it were some theoretical optimal solution or win win.

Speaker 1

Yeah. So I mean, so I guess, you know, we're just in a very sticky spot and we'll have to see how it plays out. I think Harvard is doing its best, and you know, I wish that I could give it, you know, a playbook, Harvard, this is what you should be doing, but but I can't. So let's just let's hold out hope. I think you are.

Speaker 3

I mean, there are real changes that like, yeah, Harvard should, and.

Speaker 1

I think Stephen and I think Stephen Pinker has advocated for a lot of them. Like, as I said, I'm biased, but I think that Harvard and Harvard should listen to Steve Pinker. That's my advice. I think he's one of the smartest people I know who actually knows a lot about this and has very rational views and can advise on this. It is almost eleven pm for me in Hong Kong, so I think we're gonna rap it for this week. The World to Years of Poker has already started.

You and I are not going to be out there for another It's another basically two weeks, I think for us week and a half. So let's wish all of our Risky Business listeners who are already in Vegas for the World Series of Poker good luck. We hope that you all crush it at the tables and we're looking forward to joining you soon. Let us know what you think of the show. Reach out to us at Risky

Business at Pushkin dot fm. And by the way, if you're a Pushkin Plus subscriber, we have some bonus content for you that's coming up right after the credits.

Speaker 3

And if you're not subscribing yet, consider signing up for just six ninety nine a month. What a nice price you get access to all that premium content and ad for listening across Pushkin's entire network of shows.

Speaker 1

Risky Business is hosted by me Maria Kannakova.

Speaker 3

And by me Nate Silver. The show is a co production of Pushkin Industries and iHeartMedia. This episode was produced by Isabelle Carter. Our associate producer is Sonia Gerwin. Sally Helm is our editor, and our executive producer is Jacob Bolstein.

Speaker 2

Mixing by Sarah Bruger.

Speaker 1

Thanks so much for tuning in.

Transcript source: Provided by creator in RSS feed: download file
For the best experience, listen in Metacast app for iOS or Android
Open in Metacast