BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova · 2018
Google AI Language
BERT — bidirectional Transformer pre-training via masked language modeling. Defined the pretrain-then-finetune recipe that dominated NLP until decoder-only LLMs took over.
Metadata
النوع
paper
الموثوقية
مصدر أوّليّ
اللغة
en
تاريخ النشر
11 أكتوبر 2018
المؤسسة
Google AI Language
المؤلفون
Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova