werruww

Results 37 issues of werruww

I want to buy a card rtx 3050 Does the cuda work with it? Can it speed up the running of a linguistic model 7b and train models

Failed to post request http://Localhost:11434 (b) C:\Users\m\Desktop\1>ollama serve Error: listen tcp 127.0.0.1:11434: bind: Only one usage of each socket address (protocol/network address/port) is normally permitted. ollama not work with add...

feature request

(base) C:\Users\m>pip install llama-cpp-python Collecting llama-cpp-python Using cached llama_cpp_python-0.2.85.tar.gz (49.3 MB) Installing build dependencies ... done Getting requirements to build wheel ... done Preparing metadata (pyproject.toml) ... done Requirement already...

bug

not work

How can the model be modified to the free model with ollama or groq?

How do I run a gguf multi-part model on ollaama? https://huggingface.co/Qwen/Qwen2.5-7B-Instruct-GGUF/blob/main/qwen2.5-7b-instruct-fp16-00004-of-00004.gguf qwen2.5-7b-instruct-fp16-00001-of-00004.gguf qwen2.5-7b-instruct-fp16-00002-of-00004.gguf qwen2.5-7b-instruct-fp16-00003-of-00004.gguf qwen2.5-7b-instruct-fp16-00004-of-00004.gguf how to run safetensors model on ollama ?

feature request

it is need VCRUNTIME140.DLL

bug

### System Info ```Shell colab t4 https://huggingface.co/docs/accelerate/concept_guides/ https://huggingface.co/docs/accelerate/concept_guides/big_model_inference If I have a single 16 GB Vega and a processor, how do I run a larger model of Vega on the...

docker pull docker.all-hands.dev/all-hands-ai/runtime:0.13-nikolaik m@DESKTOP-VKFHU30:~$ sudo docker run -it --privileged --rm --pull=always \ > --network host \ > -e LLM_API_KEY="ollama" \ LLM_> -e LLM_BASE_URL="http://127.0.0.1:11434" \ > -e LLM_OLLAMA_BASE_URL="http://127.0.0.1:11434" \ > -e...

!pip install udocker !udocker --allow-root install !udocker --allow-root pull docker.all-hands.dev/all-hands-ai/openhands:main !udocker --allow-root run docker.all-hands.dev/all-hands-ai/openhands:main ollama on colab !pip install colab-xterm #https://pypi.org/project/colab-xterm/ %load_ext colabxterm %xterm !curl -fsSL https://ollama.com/install.sh | sh !pip...