web-llm icon indicating copy to clipboard operation
web-llm copied to clipboard

High-performance In-browser LLM Inference Engine

Results 251 web-llm issues
Sort by recently updated
recently updated
newest added

I ran the init 3 times, while fetching the cache it ran into a network error every time at 18%. Is that something others have run into? ``` [System Initalize]...

Hello everyone, I decided to create this topic because I didn't understand how to install it. I'm completely new and didn't understand how to run AI on my macOS m1....

Where stores Web LLM initial weights on a Windows PC? It is > C:\Users\UserName\AppData\Local\Google\Chrome Beta\User Data\Default\Service Worker\CacheStorage correct? This folder occupies 4.5 GB for me. Windows 10 Pro, version 22H2

RWKV Raven 7B Gradio DEMO: https://huggingface.co/spaces/BlinkDL/Raven-RWKV-7B CPU INT4: https://github.com/saharNooby/rwkv.cpp 100% CUDA version: https://github.com/harrisonvanderbyl/rwkv-cpp-cuda ONNX convertor: https://github.com/harrisonvanderbyl/rwkv-onnx Github project: https://github.com/BlinkDL/ChatRWKV Please let me know if you have any questions :)

I noticed that the readme mentioned that we need "6.4G GPU memory" to run the demo. However, my Mac Pro only has a 4GB memory, just wondering whether there is...

When WebLLM Chat is loaded in two different tabs (same url) the `System Initialize` restarts, reloading the shards into memory: ``` [System Initalize] Fetching param cache[163/163]: 4020MB fetched. 100% completed,...

Hi, Linux can be installed and run on Android smartphones. For example, I have Ubuntu Version 20.04.3 LTS (Focal Fossa) and Debian version_ID=10 (buster) - with certain restrictions under the...

Error when running `npm run dev` of: `ReferenceError: require is not defined in ES module scope, you can use import instead This file is being treated as an ES module...

### Environment - Nuxt project using [vitesse-nuxt3](https://github.com/antfu/vitesse-nuxt3) template - Example repo in [https://github.com/k2m5t2/tvm-bot-web](https://github.com/k2m5t2/tvm-bot-web) - rough code (may result in other errors) ### Problem **Using `web-llm` in a Nuxt project results...

Just wanted to inform that I installed [Safari Technology Preview](https://developer.apple.com/safari/technology-preview/) to check its support for WebGPU and I noticed that the output of any model in Web-LLM appears scrambled like...