In this video we present BERT, which is a transformer-based language model. BERT is pre-trained in a self-supervised manner on a large corpus. After that, we can use transfer learning and fine-tune the model for new tasks and obtain good performance even with a limited annotated dataset for the specific task that we would like solve (e.g., a text classification task).

The original paper: https://arxiv.org/abs/1810.04805.

Slides used in video: https://chalmersuniversity.box.com/s/iy86xvuieygcxlmikfvivb91hl62zcs9.

Add comment

Your email address will not be published. Required fields are marked *

Categories

All Topics