Kai Ruan

Results 21 comments of Kai Ruan

Is there any plan to support Mistral 8-bit quantization recently?

> @x66ccff can you try updating to ollama `0.1.38`? The issue persist in 0.1.38

The issue seems to be caused by two Python scripts importing the ollama library differently. One uses `llama_index.llms.ollama`, while the other imports `ollama.chat` directly. This difference in import methods might...

Each method can be called repeatedly from multiple scripts, but two different methods cannot be called at the same time, which will cause the model to be reloaded for different...

Change the ctx length in ollama.Options from `8192` to `3900` (same with llama_index default) solved this problem

i got the same error, i install by using .deb , 1.8.0 and 1.8.2 both run into this. Can not figure out why.

> Just to clarify, are you following our custom sandbox guide? https://docs.all-hands.dev/modules/usage/how-to/custom-sandbox-guide hmm, i didnt use the `config.toml` in the entire process. i m not sure whether it is correct....

Sigh, i ve tried many commands and all failed using pytorch together with openhands. i also tried first copy a docker image from a running openhands container, it fails too....

@mamoodi hi, i got this error ``` (openhands) kent@kent-Super-Server:~/_Project/openhands$ docker run --gpus all -it --pull=always \ -e SANDBOX_BASE_CONTAINER_IMAGE=pytorch/pytorch:2.4.1-cuda12.4-cudnn9-devel \ -e SANDBOX_USER_ID=$(id -u) \ -e WORKSPACE_MOUNT_PATH=$WORKSPACE_BASE \ -v $WORKSPACE_BASE:/opt/workspace_base \ -v...