BERT: transfer learning for NLP

preview_player
Показать описание
In this video we present BERT, which is a transformer-based language model. BERT is pre-trained in a self-supervised manner on a large corpus. After that, we can use transfer learning and fine-tune the model for new tasks and obtain good performance even with a limited annotated dataset for the specific task that we would like solve (e.g., a text classification task).

Комментарии
Автор

Just binged the entire playlist, helped me understand the intuitions behind the math. I hope you make more videos :)

gan
Автор

Thanks a lot Lennart. What a crisp and clear explanation of BERT.

goelnikhils
Автор

This is the best explanation on Transformer I have found in the web. Can you doing another set of video for T5 ?

rickyebay
Автор

Great video. Are the original word embeddings simple static embeddings? Where do they come from?

JsaintUK