🛠️ Steven Gong

Search

SearchSearch
  • Bidirectional Encoder Representations from Transformers (BERT)
  • RoBERTA

Mar 07, 2024, 1 min read

Bidirectional Encoder Representations from Transformers (BERT)

BERT is a bidirectional Transformer. BERT is not a generative model. It’s an encoder only.

Bert tries to predict the masked token.

Resources

  • https://watml.github.io/slides/CS480680_lecture12.pdf

RoBERTA

This is just training BERT on more images.

Graph View

Backlinks

  • Generative Pre-Trained Transformer (GPT)
  • Natural Language Processing

Created with Quartz, © 2025

  • Blog
  • LinkedIn
  • Twitter
  • GitHub