About model_max_length

#1
by hongwen11 - opened

in tokenizer_config.json it says: "model_max_length": 1000000000000000019884624838656,

Can you kindly tell me the length distribution of train set? That will help me to adjust better chunk length when testing your model.

Our query and passage are trained with a length of 512, but the maximum length of the query with examples is set to 2048.

Thanks for reply. Do you train all your passages with a length of 512, or their lengths are different while averages around 512?

Just like other LLM-based embedding models, we set all passages with a length of 512.

Sign up or log in to comment