Introduction


Episode 1: From text to vectors


Figure 1

Embedding of a cat - We have described it along two dimensions: furriness and number of legs

Figure 2

Embeddings of a cat and a dog and a caterpillar

Figure 3


Figure 4

Embeddings of a cat and a dog and a caterpillar - We can describe these animals in many dimensions!

Figure 5

Schematic representations of the different prediction tasks that CBOW and Skip-gram try to solve

Episode 2: BERT and TransformersTransformersBERTBERT ArchitectureBERT as a Language ModelBERT for Text ClassificationUnderstanding BERT ArchitectureBERT for Token Classification———- END HERE ??? ———-


Figure 1

Transformer Architecture
Transformer Architecture

Figure 2

BERT Architecture
BERT Architecture

Figure 3

BERT Language Modeling
BERT Language Modeling

Figure 4

BERT as an Emotion Classifier
BERT as an Emotion Classifier

Figure 5

BERT as an Emotion Classifier
BERT as an Emotion Classifier

Figure 6

The Encoder-Decoder Attention Mechanism
The Encoder-Decoder Attention Mechanism

Figure 7

The Encoder Self-Attention Mechanism
The Encoder Self-Attention Mechanism

Figure 8

BERT as an NER Classifier
BERT as an NER Classifier