Warp icon indicating copy to clipboard operation
Warp copied to clipboard

Make Warp work with Local Language Models (like Ollama models)

Open hmdz105 opened this issue 1 year ago • 69 comments

Discord username (optional)

No response

Describe the solution you'd like?

Due to safety concerns, many users are getting paranoid about Warp's forced login and online AI assistance. Since terminals are used to access critical data on local machines and servers, adding the ability to use local language models like Llama 2 using Ollama instead of online AI engines would be a great plus to Warp terminal emulator.

Is your feature request related to a problem? Please describe.

Mostly safety concerns when using Warp for accessing password protected systems and documents.

Additional context

No response

How important is this feature to you?

4

Warp Internal (ignore) - linear-label:39cc6478-1249-4ee7-950b-c428edfeecd1

None

hmdz105 avatar Feb 26 '24 07:02 hmdz105

Some local LLM tools even expose an OpenAI API-compatible server, which I believe could allow for rudimentary support with just a setting to change the OpenAI endpoint?

Absolucy avatar Mar 09 '24 03:03 Absolucy

Thanks for this feature request @hmdz105!

To anyone else interested in this feature, please add a 👍 to the original post at the top to signal that you want this feature, and subscribe if you'd like to be notified.

dannyneira avatar Apr 03 '24 19:04 dannyneira

This is the feature that would bring me to use Warp

kn0wmad avatar Jan 19 '25 16:01 kn0wmad

As a warp user and system's administrator, this feature alone would make getting buy in from our CTO 100x easier. I love using Warp in my home lab / personal environments but no way in its current iteration could I even get it in my professional environment to demo a proof of value. Having the option to show case the AI integrations with local models would allow us to begin the discussion on an enterprise plan with zero data retention.

Ryj3l avatar Jan 20 '25 19:01 Ryj3l

Yes, please. Ollama is easiest. LMStudio great too. Groq (not the X model) would be rather interesting. Let people use the LLM they want to use. Warp team, this INCREASES the value of this tool greatly! Your value is NOT in being a niche gateway to popular LLMs.

kentlangley avatar Jan 24 '25 02:01 kentlangley

I think this is one of the important features missing in warp. 2 reasons - Most of the big companies dont let access gpt or any other LLM which captures data. Also, if they allow, there is only so many request that can be made for free.

adityagaddhyanphonepe avatar Jan 29 '25 08:01 adityagaddhyanphonepe

Would be game changer.

beamop avatar Jan 30 '25 12:01 beamop

vscode with LmStudio is amazing, WARP could integrate very easily!

mbalasko avatar Feb 05 '25 21:02 mbalasko

Consider using the keyboard shortcut action to generate prompt based on last entered command with printed output and ask it each time in order to avoid using too much CPU/data processing usage.

lulunac27a avatar Feb 20 '25 04:02 lulunac27a

This is a perfect feature for those developer who have security concerns. It's inevitable, soon enough all devices will use LLM natively just like its part of a system by default (like a RAM, CPU).

Highly suggest they would add this to the roadmap, it's easy to deploy an LLM server locally nowadays.

If they won't allow other services to connect to warp, maybe they can add this to the build natively: https://cortex.so/

So everyone is happy and not leaving warp desktop app just to run an llm server, its within the warp app itself.

rodrigoisabelo avatar Feb 22 '25 00:02 rodrigoisabelo

This shoud have been like 2nd on your list of features to implement right after plugging OpenAI API... There is no way anybody would run such a tool as an agent over a production server over ssh You can put as much safechecks as you want, you can't prove that OpenAi or other company won't save anything that goes through the endpoint, I'm also pretty sure that everything is routed through your own enpoints which are prone to attacks as well. If I'm not trusting my paid VPN nor my company one, there is no way I'd could trust an outsourced/offshore LLM messing with critical data. Regardless of what it can actually run on its own or not.

Semt3x avatar Feb 25 '25 00:02 Semt3x

I concur, I'm already running Ollama for handling other things, and ideally I'm already using qwen2.5-coder so just adding this to my terminal would mean I wouldn't need to pop out of my terminal to check something and I don't intend to use OpenAI if I can help it.

atomicangel avatar Feb 26 '25 18:02 atomicangel

Agreed that this option, especially at the free level, would allow for greater adoption in corporations if anything to validate the concept especially in organizations where they run their own LLM.

JuliusPIV avatar Feb 26 '25 22:02 JuliusPIV

This feature will make me enable AI in the app because it allows to use local AI instead of US-hosted AI(which is a huge privacy issue for me, i trust China more than US and running local is the best for me). Yes, i seen no training policy but the data can be used not only for training but for example for US government organizations to extract code parts and find vulnerabilities that can be secretly exploited by them.

PtDragon avatar Feb 26 '25 22:02 PtDragon

Warp without this feature is a dealbreaker for me. I work with too much sensitive information to not have an option to utilize a local LLM.

aidanjsamuels avatar Feb 26 '25 22:02 aidanjsamuels

Sorry, but I will never be able to trust any cloud-based AI with any kind of sensitive information (personal or professional). I've checked the network logs, and I see calls to your GraphQL API with pretty much the output of all my previous commands in it...

Having the ability to plug the terminal into my locally ran Ollama would help me alleviate my concerns. But so far, it's unfortunately a no-go for me.

jeefuji avatar Feb 27 '25 00:02 jeefuji

+1

e-p-s avatar Feb 27 '25 06:02 e-p-s

This seems extremely important to users, would be nice to hear from the warp team whether or not they intend to implement? If not, we can save ourselves the time watching this thread.

kn0wmad avatar Feb 27 '25 14:02 kn0wmad

ABSOLUTELY! I only run local LLMs, for security and fiscal reasons.

MrMxyzptlk avatar Feb 27 '25 16:02 MrMxyzptlk

Just learned about warp and have been giving it a test drive. If I could link it back to my Ollama instance, it would be a daily driver for sure!

Zelixo avatar Feb 27 '25 20:02 Zelixo

Running warp on a local LLM would be amazing.

codinkai avatar Feb 28 '25 09:02 codinkai

Agree with all of the above. Amazing product otherwise, but being able to integrate with a private, local LLM would make it perfect for competent privacy minded users.

Leaving it as an 'Enterprise Only' feature is just a cash grab. Not cool.

vZeroG avatar Feb 28 '25 16:02 vZeroG

I am a novice programmer at best and I'm trying to learn by doing. I've used the free version of Warp and ran out of tokens in less than a week. I really like Warp and it has helped me learn but since I already am paying for Github Copilot and a couple more services, I am not going to add another subscription in which the AI can only be used for one purpose. So I'm uninstalling it and just using VS Code terminal or other method where I can use my existing subscriptions or use my local llm.

lukesdaddy avatar Mar 02 '25 15:03 lukesdaddy

It can be used to avoid using API data limits and without processing data to server.

lulunac27a avatar Mar 02 '25 22:03 lulunac27a

Warp is great, but I will not use it until we can use local LLMs

TheLazyLemur avatar Mar 03 '25 07:03 TheLazyLemur

agreed, when working with many proprietary dataset, it's not possible to take advantage of warp at all regardless of the privacy/security policies associated with warp.

faux123 avatar Mar 06 '25 16:03 faux123

Adding my voice to the din: this looks like a really cool product but this feature is a hard requirement. Years ago, I was initially turned away by the always-online account-required policy, but it looks like that has been updated. That said, it's a nonstarter for me without being able to target any OpenAI-compatible backend.

jbohanon avatar Mar 07 '25 01:03 jbohanon

Love it. Have so much shit to pay for though.

ProJamKiller avatar Mar 12 '25 15:03 ProJamKiller

it would be a really complete tool with that feature.

pflajszer avatar Mar 13 '25 07:03 pflajszer

local models please 🙏

gagichce avatar Mar 20 '25 19:03 gagichce