FlagEmbedding icon indicating copy to clipboard operation
FlagEmbedding copied to clipboard

bge-reranker-v2-m3 training details

Open sreenija07 opened this issue 11 months ago • 1 comments

Could you please provide details about the process used to train the reranker model available at https://huggingface.co/BAAI/bge-reranker-v2-m3? Specifically, I'm interested in the pipeline employed for training, the dataset utilized, as well as information on the learning rate, base model, and training group size.

sreenija07 avatar Mar 19 '24 05:03 sreenija07

The training pipeline is the same as the https://github.com/FlagOpen/FlagEmbedding/tree/master/examples/reranker The dataset utilized is the same as BAAI/bge-reranker-v2-gemma and BAAI/bge-reranker-v2-minicpm-layerwise (see https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/llm_reranker) The learning rate is 4e-5, the base model is BAAI/bge-m3 (https://huggingface.co/BAAI/bge-m3) And other settings are simliar to BAAI/bge-reranker-v2-gemma and BAAI/bge-reranker-v2-minicpm-layerwise (see https://github.com/FlagOpen/FlagEmbedding/tree/master/FlagEmbedding/llm_reranker)

545999961 avatar Mar 19 '24 07:03 545999961

Could you please provide details about the process used to train the reranker model available at https://huggingface.co/BAAI/bge-reranker-v2-minicpm-layerwise? I'm mostly interested in how the model is trained

passionate11 avatar Mar 27 '24 12:03 passionate11