Issues
- 10
- 3
Engine Reuse Fails with Different JSON Schemas - "Module has already been disposed" Error
#560 opened by SMarioMan - 5
Error: Module has already been disposed
#486 opened by talperetz - 1
- 1
New error: DXGI_ERROR_DEVICE_HUNG (0x887A0006)
#489 opened by time2bot - 0
Convert gpt 2 models
#562 opened by ArpanDhot - 3
Usage Stats in Intermediate Steps
#559 opened by jdp8 - 5
Use subgroup operations when possible
#553 opened by beaufortfrancois - 5
Feature request: engine.preload()
#529 opened by flatsiedatsie - 2
Deploy small LLM in a chrome extension
#510 opened by ArpanDhot - 1
- 4
support concurrent inference from multiple models
#512 opened by mikestaub - 5
Support concurrent requests to a single model instance
#522 opened by LEXNY - 0
DuckDB-NSQL-7B Model
#554 opened by PureEngineering - 1
Model request (Aya-23)
#483 opened by time2bot - 1
vercel/ai provider integration
#551 opened by louis030195 - 1
- 1
- 4
Sending raw text to the model
#507 opened by loristns - 3
Are old models being removed?
#487 opened by flatsiedatsie - 16
Gemma 2 2B crashes on mobile phone
#524 opened by flatsiedatsie - 3
- 4
LLama 3.1 Error: Device was lost during reload. This can happen due to insufficient memory or other GPU constraints. Detailed error: [object GPUDeviceLostInfo]. Please try to reload WebLLM with a less resource-intensive model.
#517 opened by djaffer - 1
- 0
- 3
How to let the user cancel loading the model and stop it from fetching params
#499 opened by JohnReginaldShutler - 0
anyone tried to run web-llm in tauri?
#515 opened by louis030195 - 4
Deply llama 3 40 billion parameters model
#506 opened by ArpanDhot - 1
Example for using web worker with next js
#496 opened by djaffer - 2
Which LLM models can run on 6GB RTX 4050?
#500 opened by radiantone - 0
Can I initialize existing model with random weights?
#505 opened by lostmsu - 1
wasm optimization?
#494 opened by 137591 - 0
TOO SLOW in downloading models from huggingface when running 'mlc_llm package'
#504 opened by Foursheepsir - 1
[Bug] Converted model outputs gibberish text
#502 opened by gulan28 - 0
Error: Failed to execute 'mapAsync' on 'GPUBuffer'
#497 opened by hanlily666 - 0
- 2
- 3
How to actually use WebLLM
#493 opened by mirdulvultr - 1
Inconsistent and unreliable outputs on mobile as opposed to on pc/laptop for -1k models
#485 opened by JohnReginaldShutler - 3
- 0
model request: Llama-3-8B-Web
#491 opened by talperetz - 0
How to fine tuning the model in the browser?
#488 opened by Bert0324 - 0
- 0
Model request: moondream (tiny vision model)
#482 opened by darkvertex - 3
Please ensure you have called `MLCEngine.reload(model)` to load the model before initiating chat operations
#477 opened by radiantone - 7
TypeError: Failed to execute 'add' on 'Cache': Request failed [RedPajama-INCITE-Chat-3B-v1-q4f32_1-MLC-1k]
#474 opened by JohnReginaldShutler - 10
Running the MLCengine completion in the service worker results in `Receiving end does not exist`
#469 opened by talperetz - 0
Allow specifying optional `onDisconnect` callback when creating ServiceWorkerMLCEngine
#480 opened by t83714 - 0
- 1
"Only one of context_window_size and sliding_window_size can be positive"
#478 opened by flatsiedatsie