Local Gemma Chat
Ask questions and get local AI replies in your browser using WebGPU. The model runs on-device after it loads.
Checking browser support…
Model settings
Use a web-compatible
*-web.task or *-Web.litertlm file with CORS enabled.
Higher values allow longer answers and chat history; lower is safer on mobile.
Checking model cache…
If the model is already cached by this app, loading can start from local storage instead of downloading again.
Idle