ollama icon indicating copy to clipboard operation
ollama copied to clipboard

Support Mistral's new visual model: Pixtral-12b-240910

Open awaescher opened this issue 1 year ago • 43 comments

Mistral AI just dropped Pixtral, their 12b model with vision support.

  • https://github.com/mistralai/mistral-common/releases/tag/v1.4.0
  • https://www.reddit.com/r/LocalLLaMA/comments/1fe3x1z/mistral_dropping_a_new_magnet_link/

awaescher avatar Sep 11 '24 08:09 awaescher

Gosh I love them.

Qualzz avatar Sep 11 '24 09:09 Qualzz

Support. I want to see this model to be usable on Ollama.

Kerimcanak avatar Sep 11 '24 15:09 Kerimcanak

Hope too ~

tuanlda78202 avatar Sep 12 '24 08:09 tuanlda78202

+1

briansan avatar Sep 12 '24 17:09 briansan

Kudos to the Ollama team :heart:

image

(source)

awaescher avatar Sep 13 '24 14:09 awaescher

Related to Pixtral, but more generally regarding Multimodal support in Ollama:

From my experiments today, Ollama is still supporting multi-modal chat with LLaVa (retried today with v0.3.10).

There were indeed some changes in the Llama.cpp server a while back.

I was genuinely interested to understand how Ollama can still handle it while Llama.cpp reportedly cannot anymore. Was Ollama relying on llama-cli, not llama-server?

Turns out it's not relying on either. Ollama is integrating directly with the llama.cpp code base. As for LLaVa support, they lifted the LLaVa support directly from the Llama.cpp server codebase and have been maintaining that in addition to everything else since then...

Ollama team are truly unsung heroes in this technological revolution.

abenmrad avatar Sep 16 '24 16:09 abenmrad

Is it available at Ollama now?

BennisonDevadoss avatar Sep 19 '24 05:09 BennisonDevadoss

Is it available at Ollama now?

ollama pull pixtral
pulling manifest Error: pull model manifest: file does not exist

Not yet.

webboty avatar Sep 19 '24 12:09 webboty

Is it available at Ollama now?

ollama pull pixtral pulling manifest Error: pull model manifest: file does not exist

Not yet.

Same for me

dropout-developer avatar Sep 19 '24 18:09 dropout-developer

Tried to convert and add this with ollama https://huggingface.co/DewEfresh/pixtral-12b-8bit/tree/main but it seems the architecture is not supported by Ollama (yet).

K:\AI\DewEfresh\pixtral-12b-8bit>ollama create Pixtral-12B-2409 transferring model data 100% converting model Error: unsupported architecture

K:\AI\DewEfresh\pixtral-12b-8bit>ollama create --quantize q8_K_M Pixtral-12B-2409 transferring model data 100% converting model Error: unsupported architecture

CRCODE22 avatar Sep 19 '24 18:09 CRCODE22

I tried to run the Pixtral with python code using an RTX 4060 with 16GB, but it was not possible :(. Perhaps it would work with a 4090 with 24GB.

torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 60.00 MiB. GPU 0 has a total capacity of 15.70 GiB of which 34.69 MiB is free. Including non-PyTorch memory, this process has 15.64 GiB memory in use. Of the allocated memory 15.50 GiB is allocated by PyTorch, and 10.20 MiB is reserved by PyTorch but unallocated.

celobusana avatar Sep 19 '24 20:09 celobusana

Cant wait to try the gguf version of Pixtral man Ollama gonna deliver a christmas present early for us

thecr7guy2 avatar Sep 20 '24 13:09 thecr7guy2

I tried to run the Pixtral with python code using an RTX 4060 with 16GB, but it was not possible :(. Perhaps it would work with a 4090 with 24GB.

torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 60.00 MiB. GPU 0 has a total capacity of 15.70 GiB of which 34.69 MiB is free. Including non-PyTorch memory, this process has 15.64 GiB memory in use. Of the allocated memory 15.50 GiB is allocated by PyTorch, and 10.20 MiB is reserved by PyTorch but unallocated.

You need to quantize to run an 12b model on 16GB hardware.

ddpasa avatar Sep 22 '24 13:09 ddpasa

any news ?

Qualzz avatar Sep 25 '24 22:09 Qualzz

+1 for this feature

ep5000 avatar Sep 27 '24 12:09 ep5000

any news ?

A-REDA-EL avatar Sep 27 '24 15:09 A-REDA-EL

please stop spamming here.

multiple people are subscribed to this issue and patiently wait until it is done. all of them receive an email for every "+1" - it is annoying and for sure doesn't speed up anything. thanks for your understanding.

(sorry for another mail, subscribers...)

frostworx avatar Sep 27 '24 15:09 frostworx

Anyone else checking the model library a couple times a day waiting for Pixtral, Llama3.2 9B and Molmo-7B to drop? 😄

robert-mcdermott avatar Oct 01 '24 21:10 robert-mcdermott

im not sure if the team is actively working on multimodal support or if they're focusing on something else at the moment. What is certain is that multimodal capabilities will become increasingly essential in the near future, and many users may switch to alternatives that offer this functionality.

On Tue, Oct 1, 2024, 23:11 Robert McDermott @.***> wrote:

Anyone else checking the model library a couple times a day waiting for Pixtral, Llama3.2 9B and Molmo-7B to drop? 😄

— Reply to this email directly, view it on GitHub https://github.com/ollama/ollama/issues/6748#issuecomment-2387083244, or unsubscribe https://github.com/notifications/unsubscribe-auth/AIMKMGFPFQGUUDYTKWQV67TZZMFWZAVCNFSM6AAAAABOANEDJGVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGOBXGA4DGMRUGQ . You are receiving this because you commented.Message ID: @.***>

Qualzz avatar Oct 01 '24 23:10 Qualzz

in the near future, and many users may switch to alternatives that offer this functionality.

Not with local models, I guess.

sr-tream avatar Oct 02 '24 15:10 sr-tream

How much long any hint pls?

Rakhsan avatar Oct 03 '24 13:10 Rakhsan

Anyone else checking the model library a couple times a day waiting for Pixtral, Llama3.2 9B and Molmo-7B to drop? 😄

pretty much

wijjj avatar Oct 10 '24 11:10 wijjj

+1 to this feature :)

nuryslyrt avatar Oct 11 '24 14:10 nuryslyrt

Anyone else checking the model library a couple times a day waiting for Pixtral, Llama3.2 9B and Molmo-7B to drop? 😄

pretty much

there is no Llama3.2 9B THERE ARE 1, 3, 11, 90

Rakhsan avatar Oct 11 '24 16:10 Rakhsan

Anyone else checking the model library a couple times a day waiting for Pixtral, Llama3.2 9B and Molmo-7B to drop? 😄

pretty much

there is no Llama3.2 9B THERE ARE 1, 3, 11, 90

yeah. Thanks. You know what we mean. Anyways: waiting patiently.

wijjj avatar Oct 15 '24 13:10 wijjj

please stop spamming here.

multiple people are subscribed to this issue and patiently wait until it is done. all of them receive an email for every "+1" - it is annoying and for sure doesn't speed up anything. thanks for your understanding.

(sorry for another mail, subscribers...)

Please do your Daily checks for "the drop" without spamming. Holy flipping cow.

semidark avatar Oct 15 '24 19:10 semidark

Just to point out an alternative for now - LM Studio just released with Pixtral support: https://lmstudio.ai/blog/lmstudio-v0.3.5

pbasov avatar Oct 22 '24 16:10 pbasov

@pbasov Do I understand it right that this is Apple-Silicone only?

oderwat avatar Oct 22 '24 22:10 oderwat

@oderwat I believe so, yes, since it's enabled by the MLX engine and llama.cpp still doesn't support it. https://github.com/ggerganov/llama.cpp/issues/9440

But I'm sure ollama is going to get Pixtral support very soon, seeing that llama3.2 vision support is being rolled in 0.4

pbasov avatar Oct 23 '24 15:10 pbasov

guys I ma save you some time. I learned that pixtral will COME by end of december. top secret. now stop wasting your time checking every few days

Rakhsan avatar Oct 27 '24 07:10 Rakhsan