openlimit icon indicating copy to clipboard operation
openlimit copied to clipboard

Maximize your usage of OpenAI models without hitting rate limits

Results 10 openlimit issues
Sort by recently updated
recently updated
newest added

For example 40k TPM is converted to 666.66 TPS I've a summarization usecase where the incoming token is 800+, in which case the logic seems to wait indefinitely. Has anyone...

@shobrook would you mind relicensing the software as MIT, Apache, etc.?

- Use latest redis version since aioredis is depreciated - Add simple implementation of wait_for_capacity_sync to use async and wait

Hello, The package does not work on Python >=3.11 ``` >>> from openlimit import EmbeddingRateLimiter Traceback (most recent call last): File "", line 1, in File "lib/python3.11/site-packages/openlimit/__init__.py", line 1, in...

aioredis has been [deprecated](https://github.com/aio-libs-abandoned/aioredis-py#-aioredis-is-now-in-redis-py-420rc1-) as it is now included in the official redis Python package. As aioredis [fails to import](https://github.com/aio-libs-abandoned/aioredis-py/issues/1409) on Python 3.11+, it prevents me from using openlimit, even...

Issue Description Problem: When using the new OpenAI library in my asynchronous application, I've encountered an issue related to rate limiting. Description: It appears that when I run my asynchronous...

Following (v0.3.0) ```python #!/usr/bin/python3 import sys import traceback import openai from openlimit import ChatRateLimiterWithRedis openai.api_key = "" request_timeout = 3 rate_limiter = ChatRateLimiterWithRedis( request_limit=3500, token_limit=90000, redis_url="redis://localhost:6379/0" ) @rate_limiter.is_limited() def call_openai(**chat_params):...

``` rate_limiter = ChatRateLimiterWithRedis( request_limit=1, token_limit=40000, redis_url="redis://localhost:6379" ) chat_params = { "model": "gpt-4", "messages": [{"role": "user", "content": "Hello!"}], } with rate_limiter.limit(**chat_params): response = openai.ChatCompletion.create(**chat_params) ``` The code above is hanging...

Hi, thanks for the implementation! We were planning to add this feature in zeno-build library as well ([#132](https://github.com/zeno-ml/zeno-build/issues/132)). I was curious how well do the OpenAI servers work in practice...

Hi there, can you support me to run this with https://pypi.org/project/pyautogen/ ?