Ahmadzei's picture
added 3 more tables for large emb model
5fa1a76
This is very analogous to tokenization - you generally get the
best performance for inference or fine-tuning when you precisely match the tokenization used during training.