text-embeddings-inference icon indicating copy to clipboard operation
text-embeddings-inference copied to clipboard

A blazing fast inference solution for text embeddings models

Results 180 text-embeddings-inference issues
Sort by recently updated
recently updated
newest added

### Feature request splade is not supported for BAAI/bge-m3 ### Motivation splade is not supported for BAAI/bge-m3 ### Your contribution splade is not supported for BAAI/bge-m3

### System Info Question: **The reasoning results of TEI and sentence-transformers are inconsistent.** TEI version: 1.2 sentence-transformers: 2.3.1 embedding-model: thenlper/gte-small-zh ### Information - [X] Docker - [ ] The CLI...

### Model description The model can be found at: https://huggingface.co/nvidia/NV-Embed-v2. `NV-Embed-v2` is a generalist embedding model that ranks No. 1 on the Massive Text Embedding Benchmark ([MTEB benchmark](https://huggingface.co/spaces/mteb/leaderboard))(as of Aug...

### Feature request Is it possible to support HuggingFace mirror website? Such as env `HF_ENDPOINT` . Like `huggingface_hub` library, it has a environment variable `HF_ENDPOINT` which can use huggingface mirror...

### System Info - text-embeddings-inference version: 1.5 - OS: Windows/Debian 11 - Deployment: Docker - Model: [BAAI/bge-m3](https://huggingface.co/BAAI/bge-m3/tree/main) ### Information - [X] Docker - [ ] The CLI directly ### Tasks...

### Model description jina-embeddings-v3 is a multilingual multi-task text embedding model designed for a variety of NLP applications. Based on the [Jina-XLM-RoBERTa architecture](https://huggingface.co/jinaai/xlm-roberta-flash-implementation), this model supports Rotary Position Embeddings to...

# What does this PR do? This PR adds the ability to set the pooling method to one of cls, mean or last token when using the python backend. The...

### Feature request The python backend currently only supports CLS pooling ### Motivation I am currently deploying on intel gaudi hardware, and their fork of TEI currently only has support...

### Feature request Hello, Thank you for releasing this inference server! I have two requests, either of which would solve my specific problem: - Ability to specify which GPU to...