🇬🇧WebFileTools/Browser AI
Blog🇫🇷 French
🧠

Local AI in your browser

Run a real LLM entirely in your browser via WebGPU. No API key, no server, no data leaves your device.

Choose a model

🔒

The model is downloaded from HuggingFace and cached locally in your browser (IndexedDB). After the first download, everything works offline. No conversation is ever sent to a server.

What is Browser AI?

Browser AI runs a language model (Llama, Phi, Mistral) directly in your browser using WebGPU. No server, no data leakage, fully offline once downloaded.

How to use this tool?

Select a model and download it (first time only, ~0.7–4 GB). Once loaded, chat with the AI in real-time — everything stays on your device.

Benefits

  • 100% local, no data sent
  • WebGPU acceleration
  • Llama, Phi & Mistral models
  • No account required

Frequently Asked Questions

How does it work without a server?
The model runs entirely in your browser via WebGPU. Your data never leaves your device.
Which browser is required?
Chrome or Edge 113+ with WebGPU enabled. Firefox is not yet supported.
Are my conversations private?
100% private — nothing is sent to any server. The model runs locally in your browser.

Similar Tools