Introduction
Episode 1: From text to vectors
Figure 1
![](../fig/emb3.png)
Embedding of a cat - We have described it along
two dimensions: furriness and number of legs
Figure 2
![](../fig/emb5.png)
Embeddings of a cat and a dog and a
caterpillar
Figure 3
![](../fig/emb12.png)
Figure 4
![](../fig/emb6.png)
Embeddings of a cat and a dog and a caterpillar
- We can describe these animals in many dimensions!
Figure 5
![](../fig/emb13.png)
Schematic representations of the different
prediction tasks that CBOW and Skip-gram try to solve
Episode 2: BERT and TransformersTransformersBERTBERT ArchitectureBERT as a Language ModelBERT for Text ClassificationUnderstanding BERT ArchitectureBERT for Token Classification———- END HERE ??? ———-
Figure 1
![Transformer Architecture](../fig/trans1.png)
Transformer Architecture
Figure 2
![BERT Architecture](../fig/bert3.png)
BERT Architecture
Figure 3
![BERT Language Modeling](../fig/bert1b.png)
BERT Language Modeling
Figure 4
![BERT as an Emotion Classifier](../fig/bert4.png)
BERT as an Emotion Classifier
Figure 5
![BERT as an Emotion Classifier](../fig/bert4b.png)
BERT as an Emotion Classifier
Figure 6
![The Encoder-Decoder Attention Mechanism](../fig/trans3.png)
The Encoder-Decoder Attention Mechanism
Figure 7
![The Encoder Self-Attention Mechanism](../fig/trans5.png)
The Encoder Self-Attention Mechanism
Figure 8
![BERT as an NER Classifier](../fig/bert5.png)
BERT as an NER Classifier