How does this compare to ms-marco cross encoders?

#3
by hiranya911 - opened

Thanks for sharing this model with the community.

I couldn't find a direct comparison of re-ranking performance with cross-encoder/ms-marco-MiniLM-L-12-v2 (and other models in the same family). Has anybody looked into it?

Beijing Academy of Artificial Intelligence org
edited Sep 15, 2023

Thanks for your interest in our work!
The aim of bge-reranker is to train a multilingual model (mainly for English and Chinese currently), and you can see the training dataset at https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/reranker. cross-encoder/ms-marco-MiniLM-L-12-v2 is only for english.
The current evaluation is based on Chinese data and cross-language data, lacking evaluation for English. We plan to extend the evaluation dataset and add more models in the future.

Is there any update about the comparation between cross-encoder/ms-marco-MiniLM-L-12-v2 and BAAI/bge-reranker-base? Thank you.

Beijing Academy of Artificial Intelligence org

Is there any update about the comparation between cross-encoder/ms-marco-MiniLM-L-12-v2 and BAAI/bge-reranker-base? Thank you.

Apologies, but we have not conducted this comparison.
If you have any experimental results, we would greatly appreciate your feedback.

Sign up or log in to comment