Ahmadzei's picture
added 3 more tables for large emb model
5fa1a76
Pretraining methods involve a
self-supervised objective, which can be reading the text and trying to predict the next word (see causal language
modeling) or masking some words and trying to predict them (see masked language
modeling).