Overparametrized LLM: COMPLEX Reasoning (Yale Univ)

preview_player
Показать описание
Brand new AI research, published by Yale Univ et al, explores the emergence of intelligence in artificial systems, with a particular emphasis on overparameterized large language models (LLMs) trained on datasets derived from elementary cellular automata (ECA). It posits that exposure to complex yet structured datasets can facilitate the development of intelligence, even in models that are not inherently designed to process explicitly intelligent data. The authors employ ECA rules, specifically from Classes I-IV, to generate training data and evaluate LLM performance on downstream tasks.

The results indicate that models trained on rules operating near the "edge of chaos" (Class IV) demonstrate superior reasoning and chess move prediction capabilities compared to those trained on strictly ordered or purely chaotic data. These findings support the hypothesis that complexity—balanced between order and randomness—fosters the emergence of more sophisticated, generalized behavioral patterns in these models. Furthermore, training on such datasets appears to induce the development of intricate internal representations, as evidenced by attention mechanisms that effectively leverage historical context.

The methodology involves training modified GPT-2 models on ECA-generated datasets, with adaptations to handle binary inputs via linear projection layers. The study employs various complexity measures, including Lempel-Ziv complexity, Lyapunov exponents, and Krylov complexity, to characterize the ECA-generated data. Lempel-Ziv and compression complexities quantify the compressibility of the datasets, while Lyapunov exponents provide insights into the chaotic or stable dynamics of the generated sequences.

The findings suggest that models with overparameterized architectures can naturally explore non-trivial solutions, utilizing their excess capacity to form sophisticated representations of the input space, thereby elucidating their emergent reasoning capabilities. These results underscore that the emergence of intelligence in LLMs is not solely contingent on the nature of the data itself, but rather on the inherent complexity of the data, particularly when situated at the critical juncture between order and chaos.

All rights w/ authors:
INTELLIGENCE AT THE EDGE OF CHAOS

00:00 Intelligence at the Edge of Chaos
02:47 Elementary Cellular Automata Complexity Data
03:43 GPT-4o canvas calculates Cellular Automata Complexities
07:32 Rule 30 vs Rule 90 vs Rule 108
10:12 GPT-4o codes Game of Life Automaton
12:30 Analyze the Findings (complexity and reasoning)
14:09 Overparametrization leads to non-trivial Solutions in LLMs
17:00 Complexity measures
18:00 Concept of Emergence of Intelligence
24:57 Edge of Chaos is crucial for Emergence of Intelligence in LLMs
26:32 How Intelligence Emerges in AI

#airesearch
#emergence
#aiagents
#intelligence
Комментарии
Автор

Thanks for bringing this to my attention. This is very directly related to my work in generating the General Theory of Intelligence and developing a corresponding mathematical and geometric model. This is definitely moving in the right direction. It is this kind of thinking that will result in the next generation of Intelligent Systems.

TheSingularityProject
Автор

For those who are wondering (for me it was not obvious), for the evaluation phase, they freezed the intermediate layers, and they trained only the heads.

vaioslaschos
Автор

Good to see research confirming the intuitive hypothesis that intelligence emerges from the complexity at edge of chaos. Maybe one day AI will capture the essence of intuition itself.

cmw
Автор

Called it.

First we got perplexity. Next we get complexity.

i_forget
Автор

Love the channel and I've learned a lot from your content! Your focus on the reasoning capabilities of models has been particularly enlightening.

However, I need to address a point regarding the 2022 Emergent Abilities of LLMs paper. It does not discuss the complexity of training data. Instead, it only mentions 'quality' of training data, without defining what that means or how it might relate to complexity. This distinction is crucial because 'quality' can imply many things, not necessarily data complexity.

Furthermore, I would advise caution in using this paper as a reference, as its claims have been challenged multiple times. For instance, Rylan Schaeffer et al. in "Are Emergent Abilities of Large Language Models a Mirage?" argue that these so-called emergent abilities might be artifacts of the evaluation metrics used. Similarly, the paper "Are Emergent Abilities in Large Language Models just In-Context Learning?" suggests that what seems like emergent behavior could result from in-context learning, model memory, and linguistic knowledge rather than scale alone.

I hope this adds some nuance to the discussion!

soccerkenshin
Автор

Wow! This is great stuff! As an avid devotee of Wolfram and Tegmark all of this makes complete sense. (Is Wolfram kicking himself now, or jumping up and down excitedly shouting I Told You So! ?) I'm thinking "Artificial" is going to need to be dropped soon, replaced with "Emergent" perhaps. Reading the few comments so far - if you've never heard of this stuff then, yeah, this is going to be all Greek. But damn! So right on! Please do a follow up crash course to get everyone up to speed?

SixTimesNine
Автор

I love your passion, i don't understand everything but love it 😂

OumarDicko-ci
Автор

I have predicted intelligence arising from [LLM internal representations of] complex patterns nearly a year ago. Still getting around to actually building a proof of concept system that would literally sweep every AI researcher off their feet. But I am thinking of it. Complex data is the key and the more the better.

derghiarrinde
Автор

How come Stephen Wolfram was not all over this 3 years ago?

mickelodiansurname
Автор

What a beautiful conjecture. Complexity is all you need ❤

Japneets
Автор

is this the same as the many theories thàt suggest criticality in the brain poises it at the brink of a phase transition between order and randomness to optimize information processing?

blengi
Автор

I'm very interested by the measure of the complexity as it is the key in many domains. Lempel-Ziv, Lyapunov, and Krylov is a good start but not sufficient to fully define the complexity of a given system, a notion of scale is probably necessary as well and I feel some other parameters are missing as well like the interactions with others system

jmirodg
Автор

hey, dude what is the relation this serpinsky triangles(automata shapes) by/with the reasoning llm?
please another video for more data. ty bro

SecurityFirm
Автор

Does this imply that you could train an AI on images of fractals, and intelligent behavior would spontaneously emerge?

meltingscales
Автор

Please add transcript. Need the NotebookLM version. Thank you

boyardosalmon
Автор

i also thinking that true intelligent people are that who able to extrac absctract concept into a structured concept that we can learn from, the introduction of complexity into llm actually suit that concept of how its eaisier to understand something that already have structured rather than making a structured from a abstract thing.

Whic in turn make them smarter at structured concept because its a derevative of that abstracion or in this term are complexity.

onlyms
Автор

Soon we will have Rule 110 training on rule 110 and rule 30, by thr hints Stephen Wolfram gave in his last blog post about Discrete Machine learning systems he develoepd.

wwkk
Автор

I am running a bit slow today so please forgive any stupidity. (I need caffeine badly.) Is this saying that we need to start a new model by giving it synthetic data of increasing complexity, then feed it the normal training data?

s.patrickmarino
Автор

Did you miss the joke? Perhaps complexity is all you need? A sly reference to Attention is all you need.

SixTimesNine
Автор

If one is to assume that their conjecture is true, then the way forward is to design synthetic data generators which can produce data with increasing levels of complexity. The fact that this synthetic data may not be found in the physical world is not relevant to the conjecture.

So this hypothesis can be proved or disproved very quickly (subject to computing resources). One wonders why they didn't take that extra step.

pensiveintrovert
join shbcf.ru