FYI - For Your Innovation: Big Ideas Monday Mini: Artificial Intelligence

ARK Invest ARK Invest 7/10/23 - Episode Page - 14m - PDF Transcript

Welcome to the Big Ideas Monday mini-series, brought to you by the For Your Innovation

podcast.

Big Ideas is meant to enlighten investors on the long-term impact of innovation.

This annual research report seeks to highlight the technological breakthroughs evolving today

and creating the potential for super-exponential growth tomorrow.

We believe that innovation is taking off now, corroborating our original research and boosting

our confidence that our strategies are on the right side of change.

To learn more, visit arc-invest.com.

Arc Invest is a registered investment advisor focused on investing in disruptive innovation.

This podcast is for informational purposes only and should not be relied upon as a basis

for investment decisions.

It does not constitute either explicitly or implicitly any provision of services or products

by ARC.

All statements may regarding companies or securities are strictly beliefs and points

of view held by ARC or podcast guests and are not endorsements or recommendations by

ARC to buy, sell, or hold any security.

Clients of ARC investment management may maintain positions in the securities discussed in this

podcast.

Welcome to the Artificial Intelligence Overview of Big Ideas.

I'm Will Subberlin.

I co-lead venture investing and also cover AI on the research team at ARC.

And my name is Frank Downing.

I'm the director of research for ARC's next generation internet theme.

So this was clearly an exciting year in AI.

And this presentation will cover three topics primarily.

The first is generative AI.

The second are cost declines and are forward expectations for cost declines through 2030.

And third is the impact in AI on the productivity of knowledge workers.

So starting with generative AI.

Generative AI has come a really long ways in the last few years.

If you go back to just a few years ago and asked a neural net to produce an image of

say a monkey on the moon, the image that was produced would be rather terrible.

It would be grainy and it would probably not look like a monkey on the moon.

But in the last few years, we've seen a lot of architectural advancements in AI.

Transformers being one great example.

And importantly, the cost to train large AI models has continued to decline at a very

rapid rate.

And that's allowed organizations like OpenAI to train much larger models that are much

more performant and capable.

We've seen this in the image space.

Look at Dolly 2, for example.

In this presentation, we give an example of a prompt asking to create a picture of an

astronaut riding a horse.

And you can see here the image that was generated by the neural network.

You can actually use these models yourself.

You can go to Dolly 2 or many platforms that allow you to train or use models like stable

diffusion.

But we've also seen generative models beyond images.

For example, META released a make a video model where you can provide a text prompt

and from that text prompt, the neural net will actually generate a video.

So really incredible capabilities.

We're also seeing innovations on the 3D front and the audio front.

And we expect this pace to continue for the next few years.

While these models are really interesting and cool and fun to show your friends, they're

actually impacting the productivity of knowledge workers already.

In the case of coding, we've seen AI coding assistants like GitHub co-pilot more than

doubling the productivity of software engineers when it comes to coding tasks.

And so these models are abstracting away some of the more redundant tasks that would be

done by software engineers, making them more productive.

We've also seen examples on the design front.

We did a survey and found that if we wanted to have a graphic designer create a relatively

complex image, it would cost about $150 in human labor.

We can ask a neural net to create a similar image.

And after it generates a few images, getting a satisfactory one costs all in about $0.08.

We expect that cost to continue to decline.

And so in the case of image models, we've seen a cost decline from $150 in unassisted

human labor down to $0.08 with generative AI.

Now to talk about cost declines, I'll hand it over to Frank.

Thanks, Will.

A great example of cost declines as applied to image generation.

But if we step back and look at large language models as a whole, we're always looking at

cost declines to try and understand how markets are going to involve around innovative technologies.

And AI is a great example of a technology that is very compute intensive and thus very

expensive to train and run inference on these large models right now.

And one of the things we've been really shocked by is the cost declines that we're seeing

already with AI just in a short term time horizon.

So if you look at GPT-3, which is OpenAI's model that powers chatbots like ChatGPT, that

original base model cost $4.6 million to train in 2020.

And we've already seen a company called MosaicML demonstrate that they can train a model to

similar levels of performance for just $450,000 in 2022, just two years later.

And so that comes to about a 70% annualized cost decline.

And if you compare this to what we put in our Big Ideas presentation last year, where

we thought costs were declining at about 60% per year, that would have been a $740,000

training cost.

So this is one case where the actual numbers that we're seeing are outpacing our original

cost decline forecast from last year.

To double click a little bit further here, what we like to do to understand where these

training efficiencies are coming from is split out hardware cost declines and software

cost declines.

On the hardware side, we see cost declining at 57% per year.

So this is through a variety of innovations, not just Moore's law, but building on top

of adding more transistors on a chip by adding more high speed memory and increasing the

number of what are called tensor cores on a chip.

These are compute cores designed specifically for training AI models that allow, over time,

these chip architects to build more and more efficient chips, specifically tailored for

the task of AI training.

On the software side, we're seeing similar advancements as AI is becoming more prolific

and there's more and more research being poured into it.

There's various algorithmic tricks and efficiencies that AI researchers are coming up with that

have shown to decline costs at 47% independent of the compute.

And when you combine these two forces that are moving in parallel, you get this 70% cost

decline that we think, based on our research, could sustain for the next 10 years through

2030 and beyond.

So these cost declines, which we think will open up a broad market for AI, they beg the

question, what will we do with this cheaper compute?

We won't just be able to run more inferences, but we'll be able to train larger models.

If you look at GPT-3, which is already a large model at 170 billion parameters, there's a

question of how big models will get in the future and what data will be needed to train

those models.

DeepMind, an AI inside of Google, published a paper regarding a model they created called

Chinchilla in the beginning of last year that seeks to come up with a scaling function

for how large models can get based on the amount of training data that we can feed them.

And based on our estimates, the largest model they put forth in that paper is a hypothetical

10 trillion parameter model.

That's 57 times more parameters than the GPT-3 model and 720 times more tokens as this largest

kind of theoretical model that they put forth.

If you overlay our cost declines on that, we think that model, which is orders of magnitude

larger than GPT-3, could be trained for just $600,000 by 2030.

So that's just 13% of the cost to train GPT-3.

Now the trick there is 216 trillion tokens of training data.

So that's larger than the total size of Wikipedia, for example.

There's going to need to be a lot of language data generated to feed models once we get

training capacity this large.

I think that brings about an interesting question in terms of what is most important with getting

an edge in AI training that's particularly related to proprietary data advantages.

And I'll turn it back to Will to talk a little bit more about that.

As Frank mentioned, we believe proprietary data is really important, both in terms of

training a model, but also in terms of fine-tuning and continually improving a model.

When we think about a use case like self-driving, data is really important.

A company like Tesla, in our opinion, is in an advantageous position, in part because

of the data that they've acquired.

They have millions of cars on the road driving billions of miles, and they're using those

cars to continually collect more data that improves their full self-driving capability.

For example, if a driver presses a brake while autopilot is engaged, Tesla marks that as

an error on the part of the AI.

They can then use that example to continually train the model and further improve it.

We believe data advantages in this feedback loop exist in other verticals as well.

For example, in software development, AI coding assistants are continually improving based

on the feedback they're getting from usage.

So a company like Replet that has millions of software engineers on their platform and

is continually improving their AI models, we think will be in an advantage position

moving forward.

When we think about the impact of cost declines on model performance, we look at a variety

of applications of AI, one of which is AI coding assistants.

As we mentioned earlier, AI coding assistants are already doubling the productivity of many

software engineers, and when we take that 70% cost decline and model forward performance

expectations, we think that the models will be performing enough to more than 10x the

productivity of software engineers by 2030.

When we talk to researchers who are building these models, or Amjad, who is the founder

and CEO of Replet, they actually think that a 10x uplift in productivity may be too conservative,

but based on our cost declines, we are confident in our base case of 10x.

To give another example of how we think cost declines will unlock new markets, I think

it's important to touch on chat GPT and similar sophisticated AI chatbots.

So chat GPT has exploded in popularity, reaching over a million users in just its first five

days.

And part of the reason why is because you're able to ask in very natural dialogue a wide

range of questions that the AI can answer based on the corpus of public internet data

that it's trained on.

To give a sense of current performance, and this is changing rapidly as kind of new iterations

of these chatbots are released, chat GPT when launched could already answer SAT questions

at around the national average, which is pretty incredible, and that's only getting better

as more training data is fed into these models and it gets feedback from real use.

And this is already being productized and launched by Microsoft and Google as competing

product coming to market as well.

The question here, the interesting dynamic is doing this inference on these large language

models is relatively expensive.

So we've estimated the cost of one query on chat GPT at about one penny.

It may not sound like much, but when you compare it to the 8.5 billion searches per

day on Google costs that up very quickly.

So if you were looking at doing about a billion inferences per day, right now we think that

would cost about $10 million per day, and that's not even the total scale of Google.

But by 2030 using our cost declines, we think that will come down to about $650.

So a much more reasonable number that will allow these chatbots to be deployed at scale

that we think is going to happen quite quickly.

When we think about the impact of AI on knowledge worker productivity, in our base case we believe

AI will increase the productivity of knowledge workers by more than fourfold by 2030.

In our base case, that would result in companies spending about $14 trillion a year on AI software,

and we believe that could create more than $90 trillion in enterprise value.

In our bull case, at 100% adoption rate, we think companies could spend $41 trillion

on AI software to increase labor productivity by more than $200 trillion.

That's compared to the $32 trillion they currently spend on knowledge worker salaries today.

So we believe that AI could actually break consensus GDP estimates given the potential

for knowledge worker productivity.

That's all.

Thank you for listening to the AI section of Big Ideas.

ARC believes that the information presented is accurate and was obtained from sources

that ARC believes to be reliable.

However, ARC does not guarantee the accuracy or completeness of any information and such

information may be subject to change without notice from ARC.

Historical results are not indications of future results.

Certain of the statements contained in this podcast may be statements of future expectations

and other forward-looking statements that are based on ARC's current views and assumptions

and involve known and unknown risks and uncertainties that could cause actual results, performance

or events that differ materially from those expressed or implied in such statements.

Machine-generated transcript that may contain inaccuracies.