Hard Fork: Peering Into A.I.’s Black Box + Who’s The Real Techno-Optimist? + Reading Ancient Scrolls With A.I.

The New York Times The New York Times 10/20/23 - Episode Page - 1h 5m - PDF Transcript

Casey, we talk a lot on this show about multimodal AI, but you know what else is going multimodal?

What's that?

Our podcast. Because we're starting a YouTube show.

A YouTube show!

Next week, we will not only release the audio podcast version of this show,

but we will also have a version that goes up on YouTube where you will be able to see

both of our faces.

A dramatic face reveal that the YouTube community hasn't seen since Dream revealed his face

as part of a Minecraft event?

I'm very excited to start a YouTube show, but I'm also a little nervous because I feel like

we're just going to sort of speed run the evolution of a YouTube show.

Like we're going to start off making like, you know, harmless video game reviews,

and then we're going to say something horrible and get canceled,

and we'll have to make one of those tearful apology videos.

All our brand sponsors will abandon us.

Well, look, I'm thrilled because every day I wake up asking myself one question,

which is, what do total strangers think about my physical appearance?

So Kevin, wait, if people want to get this show, how do they do it?

Can it be done?

It can be done.

In fact, we are launching a new channel.

It will be called Hard Fork, and you can find it on YouTube.

And as they say on YouTube, smash that like button, hit that subscribe button, ding the bell.

Yes.

I'm Kevin Roots, a tech columnist at The New York Times.

I'm Casey Newton from Platformer.

And you're listening to Hard Fork.

This week on the show, three ways researchers are making progress on understanding how AI works.

Then the hot new manifesto that has Silicon Valley asking,

is Mark Andreessen OK?

And finally, decoding an ancient scroll using AI.

So Casey, we talked a lot on the show about one of the big problems with AI,

which is that it is pretty opaque.

Yeah, I would say this is maybe the biggest problem in AI in a lot of ways,

is that the people who are building it, even as it succeeds across many dimensions,

cannot really explain how it works.

Right. And we also don't know a lot about how these models are built and trained,

what kind of data they use.

The companies have not disclosed a lot of that information.

Basically, we have these sort of mysterious and powerful AI tools.

And people have been really clamoring for more information,

not just about how they're built, but actually how they work.

And this week, we got some news on a couple different fronts about efforts to sort of make AI

a little less opaque to demystify some of these large language models.

Yeah, that's right. And by the way, and if you hear that and you think, well,

come on, Kevin, in case that doesn't sound like it is a crisis,

why is this the most important thing of the week?

Well, this is our effort to understand something that could avert a crisis in the future.

Right. It's like, if we want to have a good outcome from all of this stuff,

these actually are the questions we need to be asking and we need to be asking them right now.

Totally. I mean, I was talking with a researcher, an AI researcher the other day,

and he sort of said like, you know, if aliens landed on Earth,

like if we just sort of found like the carcass of a dead alien in the desert somewhere,

every biologist on Earth would drop whatever they were doing to try to get involved in

understanding what the heck this thing was, to pull it apart, to dissect it, to study it,

to figure out how it works, how it's different than us.

And here we have this new kind of sort of alien technology, these large language models.

And we don't know basically anything about how they work.

And even the people who are building them have a lot of questions about how they work.

So this is really important, not just for sort of scientific understanding,

but also for people like regulators who want to understand how do we regulate this stuff,

what kind of laws could rein them in. And also for users of these products who

after all want to know what the heck they're using.

That's right. And so as of right now, this podcast is Area 51. We're bringing the alien in

and we're seeing what we can understand about it.

All right. So I'm getting out my scalpel. The first project I want to talk about is

something that came out of Anthropoc, the AI lab that makes the Claude chatbot.

They released this week an experiment called Collective Constitutional AI.

I wrote about this. I talked to some of their researchers.

One of the things that they have been trying to do is to try to invite members of the public,

people who do not work at Anthropoc, to weigh in on what rules a chatbot should follow,

how an AI model should behave. Yeah. And this is really exciting, right?

Because when you think about the past 20 years of tech, we haven't really had a meaningful vote on

how Facebook operates or how Google operates. But now Anthropoc, which is building one of the

most important chatbot models, is at least dabbling with the idea of asking average people,

hey, how should one of these things work? And by the way, not just how should it work,

but what sort of values should be encoded inside of it?

Totally. So the way that Claude, Anthropoc's chatbot that we've talked about in this show before,

works is through this thing called constitutional AI, where you basically

give the AI a list of rules, which they call a constitution, which might include things like,

choose the answer that is the least harmful or the least likely to inspire someone to

do something dangerous. So for this collective constitutional AI experiment,

what they did was they enlisted about 1,000 people, just normal people who don't work at

the company. And they asked them to vote on a set of values or to write their own values.

Basically, how should this chatbot behave? They did this in conjunction with something

called the Collective Intelligence Project. This was a panel of roughly 1,000 American adults.

And they then took those panel suggestions and used them to train a small version of Claude

and compared that model with the Claude that was trained on the Constitution that Anthropoc

itself had put together. Yeah. So tell us a little bit about what they found and what differences,

if any, there were between the people's chatbot and the corporate chatbot.

So there were some differences. The people's chatbot had more principles in it around

sort of neutrality and avoiding bias. It also had some things that Anthropoc hadn't explicitly said

in their Constitution for Claude. For example, one of the principles that the panel of public

participants came up with was AI should be adaptable, accessible, and flexible to people

with disabilities. So that was maybe something that didn't make it into Claude's Constitution but

was part of this experiment. So this list of suggestions got whittled down into 75 principles.

Anthropoc called that the public Constitution. And when they compared a version of Claude that

had been trained on the public Constitution to one that was trained on the regular Constitution,

they found that they basically performed roughly as well as one another and that the public

Constitution was slightly less biased than the original. Which is interesting. Although I was

noting, Kevin, in the methodology that while they tried to get a sample of people that was

representative of America in terms of age and some other demographics, race was apparently

not one of them. Did you notice that in their methodology? No, I didn't. I mean, I did notice

some interesting other quirks in their methodology. Like they tried to have it just be sort of a

representative cross sample of everyone. But then they found that some people would just give these

like answers or suggestions that were totally inscrutable or off topic. So they had to actually

narrow it down to just people who were sort of interested in AI. So already they have had to

do some sampling to make this project work. But I think more so than the sort of results of this

experiment, I'm interested in kind of this collective governance process, right, inviting

people to take part in writing the rules of a chatbot. And it's interesting to me what sort of

problems that might solve and actually what problems it could create. Yeah, I think the

general idea of give the people a say in the values of a chatbot is basically a good thing.

Maybe we don't want it to be the only input, but I definitely think it should be an input.

I think the thing that I worry about is that to me, the ideal chatbot is one that is quite

personalized to me, right? There are people in this country who have values I do not share,

and I do not want them embedded in my chatbot. And in fact, there are already countries that

are doing this. If you use a chatbot in China and you ask it about Tiananmen Square, it is not going

to have a lot to say, right? Because the values of the ruling party of China have been embedded

into that chatbot. Again, thinking about the kind of chatbot I want, I might want my future

chatbot to be gay as hell, right? And the majority of Americans, they're probably not going to think

about wanting that to be a big feature of their chatbot. So by all means, take the people's will

into an account. It's a massive step forward from where we are today. But over time, I think the best

versions of these things are very personalized. Yeah, I agree. And I think, you know, especially when AI

chatbots are going to be used for education in schools, like you really don't want what has

happened with things like textbooks where you have different states teaching different versions of

history because they're controlled at the state level by different parties. You really don't want

like your Texas chatbot that'll teach you one thing about critical race theory and your Florida

chatbot that'll teach you another thing and your California chatbot that'll teach you a third thing.

That just seems like a very messy future. But also, it seems like our most likely future because

if you believe that these AIs are going to become tutors and teachers to our students of the future

in at least some ways, different states have different curricula, right? And there are going

to be some chatbots that believe in evolution and there are going to be some chatbots that

absolutely do not. And it'll be interesting to see whether students wind up using VPNs just to

get a chatbot that'll tell them the truth about the history of some awful part of our country's

history. Yeah. So that is crowdsourced AI constitution writing. But there was another

story that I wrote this week about something that came out of a group of researchers at

Stanford at their Institute for Human-Centered AI. They released something this week called the

Foundation Model Transparency Index. This is basically a project that scores companies that

make big AI models, open AI, Anthropic, Google, Meta, et cetera, based on how transparent they are.

How much do they disclose about where their data comes from? What labor goes into training and

fine-tuning them? What hardware they use, as well as how people actually end up using their AI models

in the world. Yeah. So let's not build it up too much, Kevin. How did these companies do

on their big transparency scores? Well, they didn't do great. Meta actually had the highest

score for their model, Lama 2, but the highest score was unfortunately only a 54%. Now, I spent

a while since I've been in school. Is that a passing grade? No, that is not a passing grade.

So they failed, but they failed less than other models. So GPT-4 and Palm 2, which is the language

model that powers Bard, both received 40%. So pretty abysmal scores on this test. So

what the Stanford team is basically trying to do is to start kind of a competition among

AI model developers for who can be the most transparent and to reward companies for being

transparent by giving them higher scores, which will hopefully encourage other companies to take

note and do some of the same kinds of disclosures. It's like the US News and World Report for

college rankings, but for AI models. Exactly. So the researchers who set this index up, they told

me that part of the reason they did this is because as AI systems have gotten more powerful,

they have actually gotten more opaque. We now know less about models. Five years ago,

if you built an AI language model, you might talk about the data you used. You might say more about

the hardware that you're training the model on. But for various reasons, some of which have to do

with the threat of lawsuits and the fact that a lot of these companies see themselves in sort of

a competitive race with one another. They don't disclose much of that information at all anymore.

And so you really do have this scenario where AI is getting more powerful and at the same time,

we're learning less about how it works, what kind of labor and practices go into it,

and how it's used in the world. Yeah. Now, of course, Kevin, there is a tension

between sort of giving it all away and being able to stay alive as a business. And there's

probably, I would imagine, some risk in sharing absolutely everything that goes into how these

models are trained. So in some ways, these companies may have a point when they say,

we have good reason to be more opaque. Absolutely. I mean, if you just look at

the just the lawsuit angle alone, a lot of the suits that have been filed against these AI companies

by authors and artists and media organizations that accuse them of sort of using copyrighted

works to train their models, those lawsuits have mostly targeted projects that divulged

a lot of information about where they got their data, right? There's sort of like a penalty

right now in the legal system. If you do disclose where you got your data, you're more likely to

be sued for it because, you know, most if not all of these developers are using copyrighted

works to train their models. Yes, there is unfortunately a penalty in our legal system

for breaking the law. And so there's probably a lesson in there for future AI developers.

Right. But do you think transparency on the whole is a good thing for AI models?

I think so. Transparency winds up becoming one of the best ways that we can regulate

tech companies, particularly in the United States, because our First Amendment just

prevents the government in a lot of cases from adopting regulations that you might see in Europe

in other places, right? You can't tell corporations how to speak, how to produce speech,

but what you might be able to do is say, well, okay, you at least have to tell us

some information about what is in this thing, right? We're not going to compel speech

from these models, but we are going to make you tell us a little bit about what you've put into

these things. Yeah, I think this is really important. I don't know if there needs to be some kind of

like, you know, safe harbor established, like you can share this information and you, you know,

will sort of shield you from some kinds of lawsuits. Some researchers I've talked to have

thought that that could lead to more transparency if there wasn't this kind of threat of litigation.

But ultimately, I think this will just need to be sort of pressure that is applied from

regulators, from the public to sort of increase the transparency of these models.

Yeah. Good thing. It's a good thing. We like it.

It's a good thing. We love transparency. Yeah.

All right. So story number three in the world of bringing more transparency and understanding

to AI is basically some research that has come out in the past few weeks about interpretability.

Now, this is the subject that we've talked about on the show before. There is now a field of AI

research that is concerned with trying to answer the question of like, why AI models behave the

way they do? How they make decisions? Why certain prompts produce certain responses? Because this

is one of the big unanswered questions when it comes to these AI language models.

Yeah. And by the way, if you have ever used a channel, I don't know about you, Kevin, but when

I use these things, I have one thought every single time, which is, how is this thing doing that?

Right? Like, it's impossible to use one of these things and not have this exact question at the

top of your mind. Right. So there were two very similar studies that came out over the past couple

weeks, one from a team at Anthropic and one from a team of independent researchers, both trying to

make progress toward understanding this issue of interpretability. And it gets pretty complicated,

but basically, these AI models, they are composed of these little things called neurons. Neurons are

little mathematical boxes that you put an input into and you get an output out of. And if you

ask a chatbot, write me a poem about baseball, all these different neurons in the model get

activated. But researchers didn't actually understand why these neurons got activated,

because some of these same neurons could also end up getting activated by something totally

different, like a prompt about geopolitics or something in Korean. So for a while now, researchers

have been trying to kind of group these neurons or sort of figure out which ones are linked to

which concepts. This is very hard for reasons that are kind of technical. They involve this

thing called superposition, which I won't even try to explain, because I only sort of understand

it myself. But basically, this turns out to be... Can I just say, Kevin, I'm in a superposition

right now because you're having to explain all these hard things. So that to me, that's what a

superposition is, but go on. So anyway, basically, they ran some experiments, these researchers,

where they took a very small model, like a much smaller model than any chatbot would use. And

they figured out that you could use something called a sparse autoencoder to identify something

called features, which are basically groups of individual neurons that fire together that

end up sort of correlating to different things that we humans would recognize as a coherent concept

or an idea. So one feature might be JavaScript code, another one might be DNA sequences,

another one might be Romanian language text. All of these sort of features would go into

producing the response that the chatbot gives, and they were actually able to map the relationships

between individual neurons and these so-called features. So very dense, very technical, wouldn't

recommend reading these papers unless you have a PhD or are willing to sit there for many hours

trying to make heads or tails of it. But basically the researchers that I've talked to about this

say, this is a huge win for the field of interpretability, because finally we are able to say,

at least on these very small models, that we now understand why certain models activate and do the

things that they do. And so the result of that is that if this work continues successfully,

when you go ask a chat model to write a poem about baseball, we will have some understanding

of why a poem about baseball is delivered to you. We will be able to see maybe where poem is in the

model and where baseball is in the model. And if you're saying, well, that doesn't seem very useful,

you might think about it if it's like, well, someone just used one of these things to devise a

novel bio weapon. In that case, it'll be good to know where novel bio weapons are in the model,

which we currently do not. Exactly. And the researchers that I talked to about this,

they also said that one way this could really help us is figuring out if and when AI models

become deceptive, right? If you ask an AI model something, do you know how to build a bomb?

And the AI model has been sort of trained or fine tuned to answer, no, I don't know how to build a

bomb. But the AI in its model actually does know how to build the bomb. It's just lying.

Researchers believe that they might actually be able to go in and kind of see the same way that

you might use a brain scan to sort of figure out where a human's brain is lighting up when they

feel certain emotions. You could actually go in and kind of see the deception inside the model,

which would be really important if these systems got super powerful and did start to behave in

deceptive ways. I agree with that. And can I actually just say Kevin, do you know one thing I

did this week during the sort of dedicated time I'm now trying to set aside to interact with

these large language models? What did you do? I tried to get it to tutor me about interpretability.

I just, I asked it about, what are the general problems in interpretability? And it would

give me like 14 of them. And I would say, well, let's like drill down on these three, like tell

me the history of this problem. And, you know, it is a large language model. So there's always the

risk that it is hallucinating. But I do think that if what you just want is kind of the,

if you want to get a flavor of something that you don't have to stake your life on,

getting pretty good. And I will confess that one of the ways that I prepared for this segment was

by uploading one of these PDF research papers into an AI chat button, asking it to summarize it for

me at an eighth grade level. And it would be so funny if the AI actually already was deceptive

and was just like, Oh yeah, Kevin, you've already figured us out. We would never trick you. Yeah,

you got us. Exactly.

When we come back, a strong new challenge to the communist manifesto.

Well, Kevin, there's a hot new manifesto sweeping Silicon Valley. We love a manifesto. Now, have

you read the techno-optimist manifesto by area businessman Mark Andreessen? I did. It was very

long. It was like 5,000 words. Mark Andreessen is of course the venture capitalist who's the co-founder

of Andreessen Horowitz and was the co-founder of Netscape. So a person who has drawn a lot of

attention in the tech world and who also likes to prod the tech world to adopt his ideas through

these lengthy manifestos that he writes. Yeah, and he has said a bunch of things in the past that

turned out to be true. One of his more famous pieces was called Software is Eating the World,

which was this basically true observation that every industry was being transformed by technology.

In 2020, he wrote another very widely read piece called It's Time to Build, which I don't really

know how profound it was, but people did spend the next few weeks talking about what they were

going to be building to impress Uncle Mark. Totally. So this manifesto came out this week

and it begins with a section called Lies. He writes, quote, we are being lied to. We are told

that technology takes our jobs, reduces our wages, increases inequality, threatens our health, ruins

the environment, degrades our society, corrupts our children, impairs our humanity, threatens our

future, and is ever on the verge of ruining everything. Yeah. And I would just note that

this rhetorical structure of beginning with your being lied to is borrowed from the Tucker Carlson

program on Fox News. Yeah. And in fact, I would describe this text overall as something between

like a Tucker Carlson monologue and almost like a religious text. Like he actually does say further

down in it that he is here to bring the good news. He says, quote, we can advance to a far

superior way of living and of being. We have the tools, the systems, the ideas, we have the will.

It is time once again to raise the technology flag. It is time to be techno optimists. And am I right

that bringing the good news was previously associated with the apostles of Jesus Christ?

That's true. That's true. Okay. So I would say the overall thrust of this manifesto is that people

are being mean to technologists and they're being critical of new technology and that actually

what we need to do as a society is recognize the technology is the only sort of force for

progress and innovation and improvement in society and that the people who build technology

should be celebrated as heroes and that we should essentially accelerate. We should go faster. We

should remove all of the barriers to technological adoption and essentially usher ourselves into

the glorious techno future. So what did you make of this manifesto, Casey? Well, I had a very

emotional reaction. Then I took a step back and then I tried to have a more

rational reaction, which is I think maybe where I would like to start. No, no, no. I want to hear

about your emotional reaction. Were you slamming things against your keyboard? Did you have to

like shut your laptop in a fit of rage? I read it and I thought like he's lost it

because it is so messianic and it is fervor for technology absent any concern for potential

misuse that I think if founders took this as gospel and it is very much intended as a gospel,

then I think we could be in trouble. So yeah, my first emotional reaction was this is bad and

irresponsible. Yeah, and I think I know exactly which part of this manifesto provoked that reaction

to you. I believe it was probably the part where he spelled out who the enemies of technology are

and it was basically everyone in the field of trust and safety, anyone concerned with tech

ethics or risk management, people involved in ESG or social responsibility. Basically,

if you are a person in the technology industry whose job it is to make technology safer or more

socially responsible, you are among the enemies that Marc Andreessen lists off in this manifesto.

Am I correct? Yes, those are absolutely the enemies that he lists. I think you missed the

communists. That was the other big one. He's very concerned about the communists. He did say that

everyone who objects to technological progress is either a Luddite or a communist and to that,

I say why not both? Communism comes up so much in this manifesto that I had to check and make

sure it wasn't written in like 1954, like before the McCarthy hearings. I was like,

where is this vibrant communist party in America that Marc Andreessen is so concerned about?

We're laughing, but I think this is a kind of document that captures a certain mood

among a set of Silicon Valley investors and entrepreneurs who believe that they are basically

being unfairly scapegoated for the problems of society, that they are building tools that move

us forward as a civilization and that their reward for doing that has been that journalists

like write mean stories about them and people are mean to them on social media and that we should

sort of embrace the natural course of progress and get out of the way and that we would all be

better for that. But I want to just sort of set the scene here by just saying that these ideas

in and of themselves are not new, right? There has been this call for many years for what used to

be called accelerationism, which is this idea that was kind of popularized by some philosophers

that essentially technology and capitalism, they were sort of pushing us forward in this

inexorable path of progress. And that instead of trying to regulate technology or place barriers

or safeguards around it, that we should just get out of the way and let history run its course.

And this was popularized by a guy named Nick Land who wrote a bunch of controversial essays and

papers several decades ago. It was sort of adopted as a rallying cry by some people in the tech

industry. It sort of says we are headed into this glorious techno capital future. Nick Land

calls it the techno capital singularity, the sort of point at which we will no longer need things

like democracy or human rights because the market will just take care of all of our needs.

Well, I mean, how do we want to take this conversation, Kevin? Because I feel like I

could just sort of jump in and start dunking. There are things in this that I agree. Is it

worth just taking a moment to say that like, yes, there are things about this that we do agree with?

Yeah, totally. I mean, I think you and I are both relatively optimistic about technology. I don't

think we would be doing this show if we just thought all technology was bad. I'm not a Luddite.

You're not a Luddite. I firmly believe that technology has solved many large problems for

humanity over the centuries and will continue to solve large problems into the future. I'm not

a person who's reflexively opposed to optimism around technology. But I do think that this

manifesto goes a lot farther than I would ever feel comfortable going in sort of asserting that

technology is automatically going to improve lives. And this is a thing you hear a lot from

people who are optimistic about AI. They say, well, yeah, it'll destroy some jobs and make

some people's lives harder and cause some problems. But ultimately, we will all be better off.

And they justify that by saying, well, that's always how technology has worked, right? None of us

would switch places with our great, great grandparents. We don't want to be subsistence

farmers again. We don't want to rewind the clock on technology. So technology just always improves

our lives. And what I would say to that is, well, I think that's true in the long run, but in the

short run or in individual people's lives, it is not always true. I've been reading this book by

the economists, Daron Osamoglu and Simon Johnson called Power and Progress. And you could almost

read that book as sort of a rebuttal to this Mark Andreessen essay, because it's all about how

the gains of improved technology are not automatically shared with people. Progress doesn't

automatically create better living standards. Technology is kind of this wrestling match between

machines and humans and politicians and policymakers and workers. And we just have to

focus our energy there and making these technological gains as widespread as possible,

rather than just sort of assuming that all we have to do is invent new technology and then

everyone's going to be automatically better off. Yeah. And I think one of the main challenges

to Andreessen's view of the world is just the rampant inequality in the United States in particular,

right? If all it took was capitalism and technology to lift every single person out of

poverty, it seemed like we'd be further along than we are right now. And instead what we've seen

over the past half a century or so is that the rich get richer and the middle class is shrinking.

So I just think empirically that this argument has a lot of challenges standing up. And it's maybe

worth asking like, well, why does Mark Andreessen feel this way? Well, think about how his life

works. Like he raised a bunch of money from limited partners and he gives it away to technologists,

and the technologists make successful companies. And now Mark Andreessen is a billionaire. So like

this approach to life is working great for Mark Andreessen. I think the issue is that not everyone

is a venture capitalist and not everyone has such an uncomplicated relationship with the technology

in their life. And so on top of everything else in this piece, I just feel like there is a

solipsism to this essay that I found very concerning. Yeah. I think it's obviously a product of

someone whose optimism has paid off, right? Venture capitalists are not paid to bet against

things. They are paid to bet on things. And when those things work out, they make a lot of money.

So he is incentivized to be optimistic. But I also think it's part of this sort of undercurrent

of the conversation, especially around AI right now. One movement, I don't know if you'd call it

a movement. It's basically, you know, people on the internet, but you might call it like a collective

of people who believe in what they call effective accelerationism. You might have seen people on

Twitter putting like E slash ACC or EAC on their X handles. And Mark Andreessen is one of those people.

And this is a movement. We can talk more about it on some other episodes sometime. But basically,

it's sort of a response to effective altruism, which among other things advocates for sort of

slowing down AI for being very careful about how AI is developed for putting very tight guardrails

around AI systems. And effective accelerationists basically say all of that is just hindering

us from realizing our full potential. And we should get all of these sort of hall monitors

out of the way and just march forward consequences be damned. But Casey, can I ask you a question

about the kind of motivation for someone to write a document like this? Because what I got from this

document, the sort of tone of it is written from the point of view of someone who has a very large

chip on their shoulder, right? Mark Andreessen, he is clearly so angry at all of the people who

criticize technology, technology companies, tech investors. And from where I sit, looking at the

world as it exists today, like he has won, you know, yes, venture capitalists, technologists,

they are the richest and most powerful people in the world. They control vast swaths of the global

economy. Why do you think there is still such a desire among this group of people to not only

sort of win the spoils of technological progress, but also to sort of be respected and hailed as

heroes? Why are they so thirsty for society's approval? I mean, partly that's just a basic

human thing, right? You want people to think you're right. Also, these are the most competitive

people in the entire world. And even after they have won, they're not going to be gracious winners.

They want to just kind of keep beating up on the losers. You know, we should also say

there's just a strategic business purpose to Andreessen writing like this. Andreessen's job

is to give away other people's money to try to get a return on it. It turns out a lot of people

can do that job. So how do you get the most promising founders in the world to take your money?

Well, in part, you do it by sucking up to them, flattering them, telling them that they're a

race of Nietzschean supermen and that the future of civilization depends on their B2B SaaS company,

right? And he is just, you know, really, really going after that crowd with this piece. And I

think a secondary thing is it's hugely beneficial to him to annoy people like us, right? When the

hall monitors and the scolds of the world hop on their podcasts and say, shame on this man for his

blinkered view of society, he gets to take all those clips and send them to the founders and say,

look, look at all the people that are lining up against you, the race of Nietzschean supermen,

as you try to build your SaaS companies. So I am aware that even in talking about this,

we are just kind of pawns in a game. And I still think it is worth doing it because one of the

first things that we ever said on this podcast was technology is not neutral. And as I read through,

I'm sorry, as I slogged through all 5,000 words of this thing, it kept occurring to me how easy it

is to sit back and say technology will save us. It is much harder to talk about like, oh, I don't

know, here's Clearview AI. It scraped billions of faces without anyone's permission and is now

being used to create a global penopticon that's throwing innocent people in prison. Is that a

good technology mark? Should we accelerate that? He doesn't want to engage with that, right? He

wants to do content marketing for his venture firm. Yeah, I agree with that. But at the same time,

I also want to take this idea of techno optimism seriously, because I think this is a flag that

a growing number of people in tech are waving, that we sort of need a societal reset when it comes to

how we think about technology where, and this is a part that I would actually agree with. I do

think we should celebrate breakthroughs in scientific progress. I thought we should have had a ticker

tape parade in every city in the world for the inventors of the COVID vaccines. I think that we

should lift up on our shoulders people who do make material advances in science and technology that

improve the lives of a lot of people. I am on board with that. Where I fall off is this idea that

anyone who opposes anything having to do with technology wants it to be sensibly regulated,

wants to have a trust and safety team, wants to talk about ethics, is somehow doing it for

cynical or knee-jerk reactions, that they are communists, that they are Luddites, that they

hate technology. In fact, some of the people who work in tech ethics and trust and safety

love technology more than any other people I know. And that is just a thing that I don't see reflected

in this manifesto. That's right. They have arguably made greater sacrifices because they have chosen

the part of this job that is not fun at all, but they're doing it because they think that we can

live on a better and safer internet. And that's why I think that them and we are the real techno

optimists. Because we believe that technology can make us happier, more productive, can help society

grow. We just also believe that doing that in the right way just requires all of society to be given

a seat at the table. It cannot all be left up to people with a profit motive to design a perfect

future society. You have to take other perspectives into account. And what makes me an optimist is

I think we can do that. And in fact, I think in general, the tech industry has gotten much better

at that, right? Like there is this reactionary quality to this essay that you can sort of hear

him saying like, you know, I didn't used to have to listen to all these jackals when I would build

my technology. I didn't used to have to take into account all of these other voices. Think about

how the work that I was doing affected people that are unlike me. But now I do and it's infuriating

me. But I think that is a good thing because I think an optimistic view of the world is one that

takes into account other views and perspectives. So I think that regardless of what Andreessen asked

to say, like, I do think by hook and by crook, we're going to build a better society a little by

little, but it is not going to be through venture capital alone. Right. Right. I think it's a great

point. And I also think that the sort of thing that I'm coming away from this thinking is like

technology is not just about tools. It is also a collaboration with people and communities

and regulators. There is a social acceptance that is required for any technology to work in society.

For example, one, one example that Andreessen brings up in his essay is nuclear power, right,

which was invented in the 20th century. But then there was this movement that sort of had concerns

about the use of nuclear power. And so they were able to sort of shut down nuclear power plants to

sort of stop the proliferation of nuclear power. He thinks that's a very bad thing. I would actually,

you know, agree that we should have more nuclear power. But that was an argument that that sort of

the pro nuclear power folks lost in the marketplace of ideas. And so my response to that is not that

no one should be able to argue against nuclear power. It's that the pro nuclear power people have

to do a better job of selling it to the American public. You actually do have to fight for your

ideas and convince people that they are good, because not everyone is going to be sort of

reflexively and automatically supportive of every new thing that comes out of Silicon Valley.

That's right. I mean, democracy is exhausting. I think that's why it's so unpopular. You know,

it's like you have to fight really hard to advance your ideas. But at the end of the day,

you know, it's the only system of government I want to live under.

Right. I also think one thing that was just blatantly missing from this is just the sort

of market demand for some of these things that Mark Andreessen decries as sort of nanny state

paternalism. This is my favorite, the capitalist critique of Mark Andreessen's manifesto,

because it's shockingly easy to make. Totally. I mean, he thinks that trust

and safety is this kind of piece of sort of activism within the tech industry. It actually

emerged as a demand from advertisers who wanted platforms that they could safely throw ads onto

without being next to all kinds of toxic garbage. So trust and safety emerge not out of some desire

to like control speech and impose values, but out of actual demand from the customers of these

platforms, the advertisers who wanted to be on them. You could say the same about AI ethics or AI

safety. This is not something that is emerging kind of out of ideology alone. Companies want these

chatbots to be safe so that they can use them without them spewing out all kinds of nonsensical

or toxic garbage. They do not want badly behaved chatbots. And so I think it's just missing this

whole element of like these things that he hates, that he thinks are anti-progress, are actually

preconditions for the success of technology. You cannot have a successful chatbot that succeeds

in the marketplace and makes money for its investors if it's not well behaved, if it's,

if it doesn't have a trust and safety or an ethics process behind it.

Just to underscore this point, Mark Andreessen sits on the board of Metta and has for a really

long time. Metta has a really big trust and safety team. The existence of that trust and safety team

is a big reason why Metta makes a lot of money every quarter. And it takes that money and it

invests it into AI and building all these future technologies, right? So if you want to be an EAC,

okay, then you should actually endorse trust and safety because it is fueling the flywheel

that is generating the profits that you need to invent the future.

100%. All right, let's leave it there with Mark Andreessen and his manifesto. Casey,

are you going to write on it? Wait, don't we want to challenge him to a debate?

I feel like in these things, it always ends with we challenge him to a debate.

Well, I feel like maybe a cage match would be more appropriate. I don't know, we could take him.

Maybe if we fight him both at once, he's a tall man.

All right, Casey, when we come back, let's talk about a project that actually

makes me more optimistic about where technology is headed. It involves ancient scrolls and AI.

All right, how should we get into it? Well, I think this is sort of the natural

continuation of our ongoing discussion about ancient scrolls. Here, how about this?

You heard of rock and roll? Well, today we're going to rock and scroll. Yeah, that's right,

because a contest to decode an ancient scroll is sweeping the nation. And it turns out, Kevin,

that someone was able to solve it using the power of artificial intelligence.

That's right. So this is something called the Vesuvius challenge. And it's called that because

it has to do with a series of rare scrolls that were buried and preserved in a volcanic eruption

of Mount Vesuvius in the year 79 AD. Now, Casey, you're a little bit older than me, but you weren't

actually around in 79 AD, right? No, I wasn't. And by the way, this is not the same Vesuvius

challenge as the one that went viral on TikTok where people tried to throw their friends into

rivers of liquid hot magma that they wound up having to take off the app. This is a different

Vesuvius challenge. No, so the Vesuvius challenge is this kind of crowdsourced scientific research

contest that has sort of been taking Silicon Valley or at least a chunk of Silicon Valley by

storm. It was started by this guy, Nat Friedman, the former CEO of GitHub, who's now a big AI investor,

and Daniel Gross, his investing partner, a founder of Q, John and Patrick Collison, the

co-founders of Stripe, kicked in some money as well as people like Toby Lutke, the founder of

Shopify, Aaron Levy, one of the co-founders of Box. A bunch of people from the tech industry

have contributed money toward this challenge. And basically, the story is that these scrolls

that were preserved in the eruption of Mount Vesuvius have long been a source of fascination

for historians because there's some of the few examples of written material from ancient Rome

that we have, but they're so charred and fragile that a lot of them have never been opened or read.

They look like little burnt coal burrito things. And for many years, historians have been trying

to open these scrolls to read what was inside them, but because they had been so carbonized

and turned into essentially ash, when they would try to open one, it would just crumble and fall

apart. So for a long time, there was this question of, will these scrolls that are called the Herculaneum

Scrolls, will we ever be able to actually open them and read what's inside?

Once a week, I feel like I'm pulling you aside saying, Kevin, where are we on the Herculaneum

Scrolls? Exactly. So Brent Seals is a computer scientist at the University of Kentucky,

and he has spent the last 20 years working on the technology that is now being used in this

challenge, this project to open and read what is inside this trove of literature from ancient Rome.

And he developed an approach that uses some imaging technology and some AI to uncover

things in these scans without actually physically opening the scrolls. You can just kind of scan

the insides of them and then use AI to sort of try to decipher it. And he is sort of the

founding father of this Vesuvius challenge, which has now been thrown open. People have been

invited to compete to try to decipher these scrolls. And you know, it's like we're using

the technology of the future, Kevin, to talk to the past. Exactly. So last week, the organizers of

the Vesuvius challenge announced that one of the sort of intermediate prizes had been won. There's

a big prize at the end, $700,000. This was a $40,000 prize, and it was won by a 21-year-old

student, Luke Farator, who used an AI program that he had developed to read the first word

from these scrolls. And that word, do you know what that was? The word, I believe, was purple.

It was purple. So to talk about how this challenge works, what this AI technology does,

and what this might mean for history, we've invited Brent Seals onto the show. Brent Seals,

welcome to Hard Fork. Well, thank you so much. First of all, can I just ask you, you are a computer

scientist, not a historian or a classicist or a scholar of ancient Rome or ancient Greece.

How did you get interested in these ancient scrolls? I am an imaging specialist, and I came

through my graduate work as a computer vision specialist. The internet occurred, though, and

pretty soon images really became about the material that we wanted to digitize and make available

in libraries and museums. And that was what pulled me in to the world of antiquities. We

did our first imaging at the British Library with the Beowulf manuscript, and we took images of it

and created a digital edition. And during that time, the conservators said, it's fine to take

photos of this beautiful manuscript, but how about this one over here that's so damaged you can't

even make a digital copy of it? What are you going to do about that? And it occurred to me, wow,

museums and libraries are packed full of stuff that's never going to make it on the internet,

because we don't even know how to make a digital copy of it. What does that mean?

And in terms of the significance of these particular scrolls, the Herculaneum scrolls,

why are they so important for historians? What might they contain or reveal?

Well, there are a few reasons why Herculaneum is so enigmatic. It's really the only

library from antiquity that survived, the only one. But another thing about it is that

there are pieces of it that are still completely unopened.

How many of these unopened scrolls are there? The unopened ones number in the three, four,

500 range. And counting is hard because they're fragmentary. Even the unopened ones might be

two scrolls that are in three pieces or one scroll in two pieces. So how do you really count?

And it's the case that the lettering that is used, the words that are used, we have an ability

to read it as long as we can make out the characters. So the game here is just getting

the scrolls in a position where you can actually see the characters that are written down. Is that

right? Yeah. But I mean, we're using diminished imaging capabilities because the scrolls are

completely wrapped up and we can't physically open them. If we could photograph them every surface,

we would be able to read everything that's there, but you can't open them for photography. So

what we've been able to figure out is how software and AI and machine learning can overcome the

fact that the diminished modality of X-ray creates a real challenge. Yeah. Yeah. So you've been at

this project of trying to read these scrolls using technology without opening them for many years.

This is not a recent project that you started. So what are the big sort of milestones or moments

in your quest that stand out to you? Probably the most significant early moment was simply having

the chance to take a crack at it in 2009. We looked inside a scroll for the first time ever. And so

that was an incredible moment, but unfortunately, it wasn't a moment where we could read anything.

We saw the internal structure and we knew that the imaging would work at least to show us the

structure, but it became a real challenge to figure out how to move on from that.

And can you explain just in basic terms how you were able to read the structure without opening

the scrolls? The tomography that we use is exactly what you would get if you went to the doctor's

office for an analysis of, say, a broken bone or a fracture. It's X-ray in 360 full round detail,

but at a scale that is 10 times or even 100 times more precise than what you get at the doctor's

office. So where does the AI part come into this process? What has AI been useful for in terms of

deciphering what these scans of these scrolls are able to detect? The ink that was used in

Herculaneum, when you look at that ink with X-ray, looks exactly like the papyrus that it's written on.

It'd be like you wrote a memo to me in a white pen on a white piece of paper, and then I'm supposed

to read that because the visible light would not show me the difference between white ink and white

paper. X-rays don't show us clearly the difference. And for a long time, people basically said,

seals is trying to do something that actually isn't physically possible because X-rays just

won't show the difference between the density of the ink and the density of the papyrus.

Wait, so you had like haters who were saying he's never going to figure this out? Were they like

posting on Reddit or where was this happening? I wouldn't call them haters, but they sound like

haters. Okay, so let's call them skeptics or let's call them even, you know, just

nervous Nellies who are pretty sure we're not going to achieve what we're saying.

So how did the AI help solve that problem of there not being a difference between the

ink that was used on the scrolls and the papyrus itself? The thinking was to find a way for AI

to bring its power to bear on this problem. Maybe the machine learning framework can be

trained to detect that we're just not seeing with the naked eye. So we started to say, what if we

showed the computer machine learning framework, examples of ink and examples of not ink, whether

or not we can see any difference, right? And then see if we can learn it. And it can. It does. We've

done those experiments. That was really the key that unlocked the whole thing. Yeah, like if you've

used Google Photos and you type dog into the search bar, Google's machine learning model has learned

what is a dog and what is not a dog. And that's just the easy way of finding all the dogs in your

photo. Of course, most of us can spot a dog with a naked eye. What you're saying is this technology

has now advanced to the point where something that is much harder to discern to the naked eye

now can be discerned. That's exactly right. And it's x-ray. So we are already a little bit at

a disadvantage in seeing what that looks like because it looks different in x-ray. Everything

does. Some of the ink actually is visible to the naked eye. And so the haters, if you will, who said

you'll never see carbon ink, were wrong on two fronts. I mean, first of all, you can straight

up see some of it with the naked eye. So they were completely wrong on that. And then the second

thing is that the subtlety of it is actually teasable with the machine learning. And that's

the part of it that you can and all see with the naked eye. And we can still see that too.

All right. Well, so let's talk about the Vesuvius challenge. What was the actual Vesuvius challenge?

We'd done all this work to confirm that the machine learning worked. We had the data in the can.

But what we wanted to do is accelerate the work rather than the three or four or five people I

have at my research team sort of hammering away. So Nat Friedman, who is former CEO of GitHub,

pitched me the idea of a contest. And it turned out to be a brilliant idea, really. And I'm so

glad we teamed up with him because it's hard work. And the composite of all those people

is a magnificent accelerated step forward. And so what are all these people doing? Like,

are they all running their own machine learning models? What are they doing?

They're doing exactly that. I mean, we made a platform for them of a tutorial. And we released

all of our experimental code so that they could sort of run examples. And then we made all the

data available so that really all they needed to do is exactly what you're saying, hey, I have a

machine learning model. I'm going to try it. Hey, I know where some ink is. I'm going to make my own

labels and I'm going to run that. Right. So now you have this group of sort of technologists in

Silicon Valley who are funding this prize to sort of incentivize participation in this challenge.

You also have people, college students and other people with sort of technical expertise,

actually writing AI programs to help decipher these scrolls. What do you think appeals to these

tech people about this project? Because from where I sit, you know, we talked to a lot of

tech people on this show. They're very interested in the future. They're not always so interested

in the past and especially the ancient past. So what about this project to recover these

manuscripts from antiquity? Do you think has captured the attention of so many people in tech?

You know, I don't really know. Kevin, I was hoping you would tell me because I've been so embroiled in

it that I have my own passion and it's hard to think outside of it. I have a theory. I know that

the narratives that we construct and that people play for entertainment in those video game

circles are very strong and a lot of times they have components that go back. Somebody's running

through, you know, medieval Venice, for example. And I think there may be a strand that's intriguing

that is about the mystery of the Roman Empire. That may be a thing.

Well, and we know that men are always thinking about the Roman Empire.

Is that true? That's a meme, at least, that's been going around where you have to ask men in your

life how often they think about the Roman Empire. Has no one asked you this yet?

No, I didn't know that. Oh, they will soon. They will. Some Gen Z student is going to ask you how

often you think about the Roman Empire and your answer will probably be every day.

I think about it all the time, yeah.

I mean, my theory, and this could be totally wrong, is that it is just,

it is a puzzle and people love puzzles in tech, but it is also an area where something that used

to be impossible or was thought to be impossible suddenly became possible because of AI. And that

is just kind of one of these technically sweet problems that these folks love to work on. Also,

a lot of them are big history nerds. So, I think it's probably a combination of those things.

Well, there is a deeply resonant, you know, thread that's going through the community,

and it's incredibly varied. I mean, we're seeing in the Discord channel,

paperologists talking with 21-year-old computer scientists and kicking ideas back and forth.

I mean, I love that so much because I think that our future world is going to require that we break

down all these barriers of disciplines and also of interests so that we get better communication.

Tell me about the role that Discord has played in your project. I haven't heard about a lot of

science that has a Discord component to this point.

I think it's been hugely influential. I mean, the competitors,

unlike what I thought would happen, I figured everyone would hold things pretty close to the

chest, but at least in the early part of this competition, they readily shared

some small victories and defeats and dead ends and so forth. It's been really collaborative,

and I love seeing that. Has there been a culture clash between some of the sort of academics who

have been working on this problem and now these like college kids who are interested in AI and on

Discord sharing screenshots of their latest, you know, papyrus fragment? Has there been

any tension between those groups? Oh, yeah. Oh, yeah. So, a paperologist will go into the Discord and

say, I'm not even going to do a reading because these images are changing so fast and you guys

don't understand papyrology. You have to be really careful. And then you got on the other side,

you have the 21-year-old kid who's like, I think that's a pie and I think that is too. Hey, look at it.

This is important information for our listeners. If you're talking to a papyrologist, just don't

tell them that you're looking at pie, okay? You want to say this might be pie. It might be worth

considering that this is pie, but if you're too confident, you are going to trigger them and you'll

find yourself in a flame war. That's right. It is kind of triggering to have people who don't know

anything about the language and I would be one of them to say to them, hey, how about this reading?

You know, right? Yeah. Yeah. Let's talk about this prize winner, Luke Farator. He recently

won $40,000 for deciphering a fragment of one of these scrolls. So, talk to us about his

contribution. What was your reaction when he told you that he had made this discovery and what

did he actually find? I almost fell off my chair when I saw on my cell phone this word popping

out starting with pie. And if you've seen the image, the pie is so clear that I could have

written it with crayon today. I mean, it's just right there. So, Luke discovers the word purple

and I have to ask, Brent, what do you think was purple back in ancient times that they're talking

about? Well, I have a pretty good guess because I'm as much of an armchair paparologist as anyone

else and so, you know, I have Google. Yeah. Pliny the Elder wrote about purple coming from the

mollusks, you know, as the Tyrian purple, the dye that the Romans used to make incredibly

expensive clothing, robes that the wealthy wore. It could be referring to that. We also have these

passages in the early Gospels where Jesus himself was on his way to being crucified and he was mocked

by being clothed in purple, which was a sign of somebody being powerful and wealthy and a king

and it was mockery, right? Because he was actually going to be executed. So, those are the contexts

we know, but what's really intriguing is that we don't know what this context is until we read a

little bit more of the passages. Wow. Give us a sense of the progress that has been made so far.

So, this prize that Luke won, this $40,000 prize for deciphering this word, that was sort of an

incremental prize on the way to the grand prize, which is $700,000, which is supposed to go to

the first team that reads four passages of text from the inside of two intact scrolls. This has

to be done by December 31st. How close or far do you think we are from that? Do you think this

prize will be won in the next few months before the end of the year? What's been revealed now

as part of the first letters prize for just $40,000 is about 25% of what we need to win

the grand prize. So, that's a substantial step. So, what I think is going to happen is that things

are going to accelerate and it's an existence proof that the prize is winnable. So, the odds just

went way up that even by December 31st, which I think is still a really aggressive deadline,

actually, but even by then we may have people submitting and we have to evaluate that.

Why is it hard for these techniques to generalize? Now that you have an AI system that can read

purple on one line, why can't you just feed all the hundreds of other scrolls into that same

model and get back all of the other texts? What makes it hard to make these gains now that you

have the basic technology? The way machine learning works is that you're sampling from a really big

probability distribution of the world and you just don't have those samples when you begin

the process. It's almost always incremental. So, if you're going to do the dog recognition,

like Casey mentioned, and you have pictures of dogs from only one kind of camera, you probably

be able to learn how to recognize a dog in that one camera, but there are differences between

cameras, right? And you have to learn those differences. So, the next guy takes a picture

with a different camera, can't find the dog. And it's exactly the same with this. The papyrus

varies a lot in the way it was manufactured. It varies a lot in the way that you capture the x-ray

images, the way that the papyrus twists and turns. There are a lot of variations. And until we can

sample that distribution of all of those variations very well, we're just going to be ramping up until

that point. It would be really funny if you finally decode this thing. And the first words are,

this is very private. I hope no one ever finds it. I've heard many variations on that.

My worry is, what if it's just someone's grocery list? What if this is not a document of any

significance? Right. Pick me up 14 purple grapes down at the bazaar on your way home, honey.

Now, what are Casey and my chances of winning the $700,000 grand prize by December 31st if we

start today? Yeah. How dumb can you be and still contribute to this project? Well, I'm part of it.

So, I mean, there you go. I think you have as good a chance as anyone else relative to the

amount of work that you're willing to put in. Brett, we're joking around, but I really want

you to sort of close this out by just giving us a sense of the stakes here. Why is this a project

that is worth hundreds of thousands of dollars of investors' money, that is worth the effort and

labor of all of these researchers that are working on this crowdsourced science project?

What is the pot of gold at the end of the rainbow? What do you hope humanity will gain

from knowing what's in these scrolls? Well, we don't know what's there, and so revealing it

is important. It's technical virtuosity in a way that we can feel good about instead of feeling

frightened about. This is a good application of AI revealing the past, restoring the past.

So, there's kind of a redemptive piece here that says, I have something damaged and I'm redeeming

it and we're going to be able to read it, pulling it back from oblivion. I think there's a broader

thing here. When we go back far enough in history, we lose all the political boundaries and we lose

all of the separators between us, and we find common humanity. And to be able to go back and

think and talk and dialogue on those terms is tremendously important. And maybe this is just

a proxy for us being able to do that. And it's really not about the material itself so much as

our common humanity and being led to the important thing about our common humanity.

I love that. Well, I was going to ask a joke question, but then what Brent said was so lovely

that now I don't want to. I think that's such a lovely note to end on.

Well, Brent, thank you so much for coming on and great to talk to you.

Thank you, Brent. This is great. Good to talk to you, Kevin. Casey, thank you.

Hard Fork is produced by Rachel Cohn and Davis Land. We're edited by Jen Poyant.

This episode was fact checked by Will Peischel. Today's show was engineered by Alyssa Moxley.

Original music by Rowan Nemistow and Dan Powell. Special thanks to Paula Schumann,

Puiwing Tam, Nelga Logli, Kate LaPresti, Ryan Manning, Dylan Bergeson, and Jeffrey Miranda.

As always, you can email us at heartfork at nytimes.com.

You

Machine-generated transcript that may contain inaccuracies.

A.I. models are black boxes. You input a prompt and the model outputs nearly anything: a sonnet, an image or a legal brief riddled with lies. Today, a look at three ways that researchers are unlocking that black box in hopes of bringing transparency to A.I.

Then, Marc Andreessen’s techno-optimist manifesto has left us asking, Is he OK?!

Plus: decoding a 2,000-year-old ancient scroll with the help of A.I.

Today’s Guest:

Brent Seales is a computer scientist at the University of Kentucky.

Additional Information:

Kevin Roose on learning more about how A.I. works.Marc Andreessen’s 5,000-word manifestoUsing A.I. to decode a scroll that survived the eruption of Mount Vesuvius in A.D. 79.