Chip Placement with Deep Reinforcement Learning (Paper Explained)

preview_player
Показать описание
The AI Singularity is here! Computers designing new computers! It takes human experts multiple weeks to design new computer chips. What looks like a large game of Tetris is actually a very complex optimization problem. This paper uses Deep Reinforcement Learning to solve this optimization both faster and better than humans.

Abstract:
In this work, we present a learning-based approach to chip placement, one of the most complex and time-consuming stages of the chip design process. Unlike prior methods, our approach has the ability to learn from past experience and improve over time. In particular, as we train over a greater number of chip blocks, our method becomes better at rapidly generating optimized placements for previously unseen chip blocks. To achieve these results, we pose placement as a Reinforcement Learning (RL) problem and train an agent to place the nodes of a chip netlist onto a chip canvas. To enable our RL policy to generalize to unseen blocks, we ground representation learning in the supervised task of predicting placement quality. By designing a neural architecture that can accurately predict reward across a wide variety of netlists and their placements, we are able to generate rich feature embeddings of the input netlists. We then use this architecture as the encoder of our policy and value networks to enable transfer learning. Our objective is to minimize PPA (power, performance, and area), and we show that, in under 6 hours, our method can generate placements that are superhuman or comparable on modern accelerator netlists, whereas existing baselines require human experts in the loop and take several weeks.

Authors: Azalia Mirhoseini, Anna Goldie, Mustafa Yazgan, Joe Jiang, Ebrahim Songhori, Shen Wang, Young-Joon Lee, Eric Johnson, Omkar Pathak, Sungmin Bae, Azade Nazi, Jiwoo Pak, Andy Tong, Kavya Srinivasa, William Hang, Emre Tuncer, Anand Babu, Quoc V. Le, James Laudon, Richard Ho, Roger Carpenter, Jeff Dean

Links:
Рекомендации по теме
Комментарии
Автор

Love how heavily you have been uploading, thank you for all the work, really great to learn from you.

aleks
Автор

In case, anybody is interested, wire routing congestion occurs on a system-on-chip when a lot of wires (or metal lines) are routed in a narrow space.

roman
Автор

thanks for the explanation this is very interesting

doublevgreen
Автор

Can you do a dedicated video on Graph Conv?

So far I havent seen any consensus about the best way to do graph conv.

herp_derpingson
Автор

"How amazing!" *Laughs* "Yeah humanity is doomed..." XD

ribbonwing
Автор

You should make livestream where you try to reproduce paper.

cunningham.s_law
Автор

We are entering the era of Fine tuning.

sahiluppal
Автор

Hey Yannic, could you cover Network Deconvolution? The paper seems like it's going to be a foundational method in a couple years replacing batch norm

sphereron
Автор

Hi, does it have and code implementation or dataset?
I could not find any in their source paper.

mohammadjavadnouri
Автор

Does anyone know if there are simulators for this chip placement anywhere ?

mamalamini
Автор

Hi Yannic great. can you please explain more deep reinforcement learning papers on real world problems.

thirugnanamv
Автор

I wanted more information on converting netlist to graph representation (actual implementation procedure)

bhuvaneshs.k
Автор

Can we get something similar for ... SimCity :)

pooglechen
Автор

Singularity would mean the computer even decide which components to use as well and not just arranging the "given" components. Its way far far from the singularity.

wahabfiles