மனோஜ்குமார் பழனிச்சாமி
மனோஜ்குமார் பழனிச்சாமி
Yes, after clearing the docker image
Run this to check LLM response time. ```python import warnings warnings.filterwarnings("ignore") import tomllib as toml from litellm import completion from datetime import datetime file_path=r'config.toml' config = toml.load(open(file_path,'rb')) messages = [{...
Add the following code to `opendevin\llm\llm.py` and run ```py if __name__ == '__main__': llm = LLM() messages = [{"content": "42?", "role": "user"}] response = llm.completion(messages=messages) print('\n' * 4 + '--'...
@enyst In some low-end devices with 8GB RAM, even to generate "Hello", it took around ~3 mins for a ~6GB model.
> Add the following code to `opendevin\llm\llm.py` and run > > ```python > if __name__ == '__main__': > llm = LLM() > messages = [{"content": "42?", "role": "user"}] > response...
> @enyst In some low-end devices with 8GB RAM, even to generate "Hello", it took around ~3 mins for a ~6GB model. When testing the LLM manually. So, I thought...
Is the mic working properly?
@DamianFekete, would you test #408?
Should you use `tqdm.autonotebook` here?
Would you run with the following? ```py import os os.environ['BROWSER_USE_LOGGING_LEVEL']='debug' ```