FlagEmbedding icon indicating copy to clipboard operation
FlagEmbedding copied to clipboard

Retrieval and Retrieval-augmented LLMs

Results 622 FlagEmbedding issues
Sort by recently updated
recently updated
newest added

Do you plan to train and release multilingual embedding models in the near future?

query:请帮我找出大于0.5的数据 知识库:0.1,0.2,0.3,0.8,0.6 召回结果:0.6,0.8 向量模型能做到这种程度吗?

长文本向量化后需要存储,节省之后每次计算的时间,请问官方有没有像gensim的save()那样保存词向量的方法呢

Hi, what is the name of the loss function used for fine-tuning? And could you please point me to where it is defined? Edit: looks like CrossEntropyLoss with In-Batch Negative...

如果我有一个ES,那么还需使用这个函数要计算这一项吗? compute_lexical_matching_score

https://github.com/vllm-project/vllm/issues/2676

只有最后4k插入needle可以通过,前面插入全部失败。 测试输入长度16k,32k,128k. [Needle In A Haystack](https://github.com/gkamradt/LLMTest_NeedleInAHaystack) 下面是29.5k长度测试输出 `******************** Input Length: 29605 Position: 0k Prediction: There are many great things to do in San Francisco! Here are some suggestions: 1. Visit...

BGE中的retroMAE阶段,为什么encode阶段不用position embedding呢?而decoder阶段都有用到,是为什么呢?非常期盼解答!

我在本地试用bge-reranker-base模型推理三条数据平均花了2.8秒,我的cpu是i5-7500,请问有提升推理速度的方法吗,如果要升级硬件配置,大概什么样的配置能达到推理十条数据在两秒之内呢