GPT-3 vs Human Brain

preview_player
Показать описание
GPT-3 has 175 billion parameters/synapses. Human brain has 100 trillion synapses. How much will it cost to train a language model the size of the human brain?

REFERENCES:

[1] GPT-3 paper: Language Models are Few-Shot Learners

[2] OpenAI's GPT-3 Language Model: A Technical Overview

[3] Measuring the Algorithmic Efficiency of Neural Networks
Рекомендации по теме
Комментарии
Автор

GPT-3 has 175 billion parameters/synapses. Human brain has 100 trillion synapses. How much will it cost to train a language model the size of the human brain?

lexfridman
Автор

That's interesting because, if the trend continues, it will also cost $5M to train a human brain at college in 2032

engboy
Автор

These short highly focused videos are a nice mental appetizer, and its easy to set aside 5 mins to watch them between consecutive unsuccessful model training runs

georgeprice
Автор

GPT-800: I need your clothes, your boots and your motorcycle 😎😂

xvaruunx
Автор

It would be awesome to see your breakdowns on GPT-3. Explain to us dummies how it works!

twinters
Автор

You forget that the 100 trillion synapses doesn't only do language, it does vision, reasoning, biological function, fine motor control, and much more. The language part (if we can isolate from other parts) probably uses a fraction of those synapses

a_name_a
Автор

Thank you for this post. Powerful topic. Excellent description of the potential for this platform
and hurdles involved.

hubermanlab
Автор

the cost of training will be nothing compared to all the money they make on selling this as a service :o

josephkevinmachado
Автор

I think it is important that computer scientists use the term neuron and synapse very carefully. I am a molecular biologist and to equate neurons in neural networks to biological neurons, or even a synapse, is like calling an abacus a quantum computer. I don't say this to diminish machine learning at all, I use it as a biologist, and I've been showing my whole family AI Dungeon 2 utilizing GPT-3; it really is tremendous. But there is such a large difference between computer neurons (I'll just call them nodes) and biological neurons.

Each neuron itself could be represented as a neural network with probably 100s of trillions of nodes or maybe magnitudes more, and each of those nodes would itself consist of probably thousands or millions of nodes in their own neural network. This is to say that the computation involved in determining whether there is an action potential or not is truly massive. I wish I could put this into more precise words but the complexity of even a single neuron is far, far greater than the complexity of all human systems of all times compiled into even a single object.

I will try to exemplify this using a single example in my field of expertise, microRNA. The synapse consists of multiple protein complexes that work to transmit a chemical signal from outside the cell to inside the cell. In this case, the outside signal is created by another neuron. Every one of those proteins has dozens (and probably a lot more than that) of regulatory steps along the path of its production, localization, and function. These regulatory steps happen over time and themselves consist of other molecules produced/consumed by the neuron, each of which have their own regulation.

Now let's say we have neuron 1 and it is trying to form a synapse with neuron 2. At the position neuron 1 and 2 physically interact, communication has already happened and all the necessary players (small molecules, RNA, and protein) have been recruited to this location. The moment of truth arrives, neuron 1 has an an action potential. Neuron 2 starts to assemble a brand new synapse at that location but this does not end in the production of a new synapse. In neuron 2, perhaps hours or days previous it decoded a complex network of extracellular signals that culminated in the localization of a specific microRNA at the location of this potential synapse. At the same time neuron 2 receives the signal from neuron 1, that microRNA is matured and is made active over a period of minutes. Instead of this new synapse being formed on neuron 2, this specific microRNA causes the production of protein necessary for its completion to stop and the whole process is aborted.

At every step of every process in normally functioning neurons, these seemingly miraculous processes are occurring. They are occurring in our billions of neurons over our entire lives, existing in a body of trillions of cells that are all equally as complex, communicating with each other always and for our entire lives.

I say this not to demean or lessen the work of you, Lex, or any other computer scientist. But I say this to humble us, for us to be a little more careful when we say so casually, "it's just computation."

Jacob_A_OBrien
Автор

2:58 I would love an in-depth GPT-3 video, explaining how it works, the algorithms behind it, the results it has achieved, and its implications for the future.

georgelopez
Автор

This is what I love about 2020 and the Internet. Two decades ago a channel concentrated on the eclectic scientific subjects that Lex covers would have had little activity. But I was thrilled to see that this video, released only hours ago, has a ton of comments and likes on it already, just like a typical YouTube "video star" channel! :D

On the darker side. The millions of dollars required to train a network like GPT-3 does torpedo somewhat the "democratization" of AI initiative. And yes, in X years the power required to train a GPT-3 system might fit in a smart phone. But when that happens there surely will be new hardware as powerful to that coming "genius" smart phone, as the computing cluster that GPT-3 was trained on is to the typical computing resources the average person can afford today. Perhaps it will be some astonishing combination of quantum computing and vast distributed parallel processing (or said more humorously by Marvin in The Hitchhiker's Guide to the Galaxy, a computing platform with the "brain the size of a planet") . Maybe that's just the way the Universe is and always will be??

rag_llm
Автор

Thats the price of our last invention... After that...we might just be at best associate producers on everything.

BlackCat.Designs
Автор

2:30 Looks like Ray Kurzweil’s prediction for the singularity is tracking pretty accurately.

DynamicUnreal
Автор

I thought 175 billion parameters were a lot... They are actually! Wonderful!

VincentKun
Автор

To be honest, I was expecting a figure in the ballpark of the hundreds of trillions of USD, more the entire World's GDP and stuff.

USD 2.6 billion doesn't sound impossible even in 2020. Maybe I'm poisoned by reading about billions too much, and startups like WeWork being worth dozens of billions - but some company/individual investing USD 2.6 bi / USD in 2020 to have a language model, or at least something that is at least hundreds or thousands of times better than GPT-3 sound feasible to me.

canaldoapolinario
Автор

It's not an exaggeration when people say that the most valuable possession you have is your brain...

Bati_
Автор

Not all human synapses are dedicated to image processing

danielro
Автор

At this time there has came out faster and memory efficient training method: There is no need to train every synapse at each iteration, but only subset of them

aiart
Автор

Now this what the whole world needed, Getting a bit of an idea from different articles stating what GPT-3 is but not really we got any update or clue.👋
This is the real thing that you have talked about Lex.👌👌👌👌👌👌
Good one....😺😺😺😺😺😺😺😺😺😺😺😺😺😺😺

ChampionRunner
Автор

GPT-3 Sentence completion: Humanity is...[about to become obsolete] [unprepared for what's coming] [blissfully ignorant of the future they are about to experience]

kayrosis