Local Gemma Chat

Ask questions and get local AI replies in your browser using WebGPU. The model runs on-device after it loads.

Checking browser support…
Model settings
Use a web-compatible *-web.task or *-Web.litertlm file with CORS enabled.
Higher values allow longer answers and chat history; lower is safer on mobile.
Checking model cache…
Idle
If the model is already cached by this app, loading can start from local storage instead of downloading again.