Graph Language Models EXPLAINED in 5 Minutes! [Author explanation 🔴 at ACL 2024]

preview_player
Показать описание
How to make powerful LLMs understand graphs and their structure? 🕸️ With Graph Language Models! They take a pre-trained language model and fit it with the ability to process graphs. Watch if you're curious about how this works (hint: choose the right positional embeddings)!

Follow Moritz Plenz (first author) on:

Thanks to our Patrons who support us in Tier 2, 3, 4: 🙏
Dres. Trost GbR, Siltax, Vignesh Valliappan, Michael, Sunny Dhiana, Andy Ma

Outline:
00:00 LLM for graphs
01:08 Motivation
02:02 Key idea of Graph LLMs
02:25 Relative Positional Encodings
03:00 Method (Graph LLMs)
04:04 Experiments and Evaluation
04:49 Results
06:07 Outro

▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀
🔥 Optionally, pay us a coffee to help with our Coffee Bean production! ☕
Join this channel as a Bean Member to get access to perks:
▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀▀

🔗 Links:

#AICoffeeBreak #MsCoffeeBean #MachineLearning #AI #research

Video editing: Nils Trost
Music 🎵 : Space Navigator – Sarah, the Illstrumentalist
Рекомендации по теме
Комментарии
Автор

Its interesting to see "attention" on graph structures again. I think in the future a more structured knowledge representation may play a role on improving reasoning, as we could leverage logic and rules using engines on them, like compilers aid in code generation.

bensimonjoules
Автор

Excellent! need to go deeper that could be a game changer for reasoning, as it makes more sense to reason on a graph rather than on the next token.

jmirodg
Автор

Thank you for this video Letitia! As always amazing :=)

Side note: Loved the silent Good Bye :)))

bharanij
Автор

I'm interested even more on the generative side, generating large graphs with contained text in them from a prompt, can be useful for modalities which are represented by large graphs. I've yet to see anyone doing this. While you can prompt LLMs to generate small graphs, for larger graphs you see significant performance drops.

vladimirtchuiev
Автор

Interesting. It's almost like two types of tokens: nodes + edges which can each be compressed to a feature vector. But yes, with positional encoding you're left with "random walk with restart" or a traversal depth. Or one could sum node_vector + edge_vector ~= positional distance. but yeah, more graph solutions coming in the future.

jonclement
Автор

It would be interesting how much computation this needs

quebono
Автор

I need to read this more deeply, I don't understand why would just grafting the parameters willy nilly works

yorailevi