text-embeddings-inference icon indicating copy to clipboard operation
text-embeddings-inference copied to clipboard

A blazing fast inference solution for text embeddings models

Results 180 text-embeddings-inference issues
Sort by recently updated
recently updated
newest added

### Feature request Hello, My question is that if/how we support multiple models selection by a single api endpoint. For example, I have gte, bge etc. and I can deploy...

### Feature request The repo should add a CONTRIBUTING.md ### Motivation To help new contributors ### Your contribution I can help with adding one if there is a widely used...

### Feature request Does it make sense for TEI to add a cache layer for embeddings? Not sure if TEI supports this already. If not, I'd be curious if it...

### Feature request In the concurrent scenario, I tried to reduce a batching_task, and the batchsize of each embed is larger, so that the inference performance is better.In the single-concurrency...

### Feature request support Volta gpu ### Motivation support Volta gpu ### Your contribution ....

### Model description Hello! Thanks for this great work :) Previously, I implemented [mpnet-rs](https://github.com/NewBornRustacean/mpnet-rs) and found a related issue(feature request) #33 If there is no on-going work for the mpnet...

### System Info Colab Pro T4 ### Information - [ ] Docker - [X] The CLI directly ### Tasks - [X] An officially supported command - [ ] My own...

# What does this PR do? Enable CPU device for python backend Fixes # (issue) ## Before submitting - [ ] This PR fixes a typo or improves the docs...

### Feature request When attempting to use this reranker model [mxbai-rerank-large-v1](https://huggingface.co/mixedbread-ai/mxbai-rerank-large-v1) from huggingface with TEI, I got the following error message: ``` Error: Could not create backend Caused by: Could...

### Feature request Add cli option to auto-format input text with config_sentence_transformers.json prompt settings (if provided) before toknizing. ### Motivation A lot of models now expect a prompt prefix so...