ChatLLM-Web
ChatLLM-Web copied to clipboard
đŖī¸ Chat with LLM like Vicuna totally in your browser with WebGPU, safely, privately, and with no server. Powered by web llm.
ChatLLM Web
English / įŽäŊ䏿
đŖī¸ Chat with LLM like Vicuna totally in your browser with WebGPU, safely, privately, and with no server. Powered By web-llm.

Features
-
đ¤ Everything runs inside the browser with no server support and is accelerated with WebGPU.
-
âī¸ Model runs in a web worker, ensuring that it doesn't block the user interface and providing a seamless experience.
-
đ Easy to deploy for free with one-click on Vercel in under 1 minute, then you get your own ChatLLM Web.
-
đž Model caching is supported, so you only need to download the model once.
-
đŦ Multi-conversation chat, with all data stored locally in the browser for privacy.
-
đ Markdown and streaming response support: math, code highlighting, etc.
-
đ¨ responsive and well-designed UI, including dark mode.
Instructions
-
đ To use this app, you need a browser that supports WebGPU, such as Chrome 113 or Chrome Canary. Chrome versions ⤠112 are not supported.
-
đģ You will need a GPU with about 6.4GB of memory. If your GPU has less memory, the app will still run, but the response time will be slower.
-
đĨ The first time you use the app, you will need to download the model. For the Vicuna-7b model that we are currently using, the download size is about 4GB. After the initial download, the model will be loaded from the browser cache for faster usage.
-
âšī¸ For more details, please visit mlc.ai/web-llm
Roadmap
-
[â ] LLM: using web worker to create an LLM instance and generate answers.
-
[â ] Conversations: Multi-conversation support is available.
-
[] Desktop: Build a desktop version with Tauri, which will use system cache (not just browser cache) and run locally and offline.
-
[] UI: Dark and Light Theme.
Deploy to Vercel
Development
git clone https://github.com/Ryan-yang125/ChatLLM-Web.git
cd ChatLLM-Web
npm i
npm run dev
Screenshots


LICENSE
MIT