DETR: End-to-End Object Detection with Transformers | Paper Explained

preview_player
Показать описание

In this video I cover DETR, an end-to-end object detection pipeline with transformers.

The main 2 ideas are:
* Using transformers instead of specialized vision architectures
* Using Hungarian matching and loss to train the system e2e

▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬

⌚️ Timetable:

00:00 Intro: DETR main ideas
00:45 Non-max suppression
03:20 High-level pipeline overview
07:50 Architecture in more detail
12:10 Matching loss
18:35 Hungarian loss
21:00 Results
24:05 Visualization
27:35 Ablations
30:00 Outro: Segmentation results

▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
💰 BECOME A PATREON OF THE AI EPIPHANY ❤️

If these videos, GitHub projects, and blogs help you,
consider helping me out by supporting me on Patreon!

One-time donation:

Much love! ❤️

Huge thank you to these AI Epiphany patreons:
Eli Mahler
Petar Veličković
Zvonimir Sabljic

▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬

💡 The AI Epiphany is a channel dedicated to simplifying the field of AI using creative visualizations and in general, a stronger focus on geometrical and visual intuition, rather than the algebraic and numerical "intuition".

▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
👋 CONNECT WITH ME ON SOCIAL

👨‍👩‍👧‍👦 JOIN OUR DISCORD COMMUNITY:

📢 SUBSCRIBE TO MY MONTHLY AI NEWSLETTER:

💻 FOLLOW ME ON GITHUB FOR COOL PROJECTS:

📚 FOLLOW ME ON MEDIUM:
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬

#detr #objectdetection #transformers
Рекомендации по теме
Комментарии
Автор

bro you're insanely good to explain complex stuff

vincenzo
Автор

17:50 for anyone asking why is that, that's the Generalized Intersection over Union Loss (LGIoU) that is defined as 1 - GIoU.
The GIoU range is [-1, 1]; when GIoU = 1 it means that there is a perfect overlapping (and then LGIoU = 0), meanwhile when GIoU goes to -1 it means that boxes not only do not overlap but their centers are far from each other too (and so LGIoU goes to 2).
So, in the end, best case scenario is LGIoU equal to 0, worst is equal to 2.
So for this reason LambdaIoU is positive (actually it is equal to 2, as found in the Appendix of the paper).

ghostcleanseIRL
Автор

Really nice video! Thank you for doing this! For the IoU part in the loss, I agree that the coefficient must be negative to since better alignment results in higher IoU, so the hyperparameter lamda_iou must be negative.

supang
Автор

I think the IoU loss defined in the appendix actually makes sense, it has 1 - at the beginning so it goes down also

briannlongzhao
Автор

It really helps me a lot!! Especially the match part! thanks!

leoliu
Автор

hey, man. The iouloss part I think it’s a different iouloss function and the paper didn’t show you. In that function when the iou gets bigger the result gets smaller, that make sense.

leoliu
Автор

Thanks for the great explanation! Why doesn’t the author compare results with YOLOv5 or any other single stage models?

ameynaik
Автор

It will be awesome if you also cover the how to use this coding part .

YourAI_agent