A Boring Standard That Changes Everything
WebLLM is an open protocol that makes AI universal. Users choose their provider. Developers skip the infrastructure. The web gets AI everywhere.
AI should be a utility.
Instead it's a mess.
Users pay 5x
Same model, different apps, multiple bills
Developers rebuild
Identical infrastructure, over and over
Wikipedia can't afford
What individual users pay for
3 companies control
How we get access to AI
Stop paying multiple times for the same technology
Users pay for GPT-4 five times across different apps. One subscription should work everywhere — like having one internet bill, not paying each website separately.
After: Pay once → Use everywhere
Total: $75/mo
Same GPT-4 model
Everyone deserves AI, not just the venture-funded
AI shouldn't be a luxury for funded startups. When users bring their own AI, every website can be intelligent.
All AI-enabled now
Make use of the computers people already have
Millions of powerful GPUs and M-series Macs sit idle while their owners pay for cloud AI. That's backwards.
• Privacy bonus: Your data never leaves your device
• Speed bonus: Often faster than API calls
Your hardware finally wakes up and works for you
RTX 4090
82 TFLOPS AI
M3 Max
40 TFLOPS AI
Stop rebuilding the same plumbing
Every company spends weeks building identical infrastructure: auth, billing, rate limiting, error handling. Collective waste in the millions of developer hours.
Ship features, not infrastructure
Before
4 weeks
After
navigator.llm4 lines
Write once, evolve forever
Stop hardcoding model names that become obsolete in 6 months. Describe what you need, let the system evolve.
When GPT-5 or Claude 4 launches, your code automatically uses it. No refactoring. No deployment.
Standards outlive products
Hardcoded approach:
Constant refactoring
Standard approach:
navigator.llmAuto-upgrades to best model
AI at the speed of thought
Every AI call shouldn't require a round-trip to your server, which often just relay requests to other AI services. Tools should execute where the data lives — in the browser.
• Instant tools: Manipulate DOM, update state immediately
• Real-time: Forms that complete as you think
The backend was always unnecessary overhead
Traditional:
~700ms round trip
With WebLLM:
~10ms instant
Data sovereignty for users
Privacy shouldn't depend on corporate promises. When AI runs locally, sensitive data never leaves your control.
Privacy by architecture, not by policy
Medical
Local only
Financial
Local only
Work docs
Local only
Personal
Local only
Users control their tools, not vice versa
You shouldn't need permission to use your preferred AI. Your tools should adapt to you, not lock you in.
• Switch anytime: No app breakage when you change
• Mix and match: Different AI for different tasks
Your preferences, your choice, everywhere
Your AI preferences:
Claude
Primary choice
GPT-4
Backup
Gemini
Available
Works on every website
Small teams can compete with giants
Infrastructure advantage disappears when users bring their own AI. Now it's about ideas, not capital.
• Open source: Compete with commercial products
• Students: Build without credit cards
David and Goliath use the same AI
BigCorp
1000 devs
Solo Dev
1 person
Same AI capabilities
Proprietary products become universal utilities
The web thrives on open standards. AI should be no different — universal, interoperable, and user-controlled.
Standards boring. Standardization revolutionary.
<video>fetch()WebGLnavigator.llm527
Developers building
10,241
Users connected
142
Commits this week
See what the community thinks about our LLM Gateway
Plus Chrome's built-in Gemini Nano, and many more through OpenAI-compatible APIs
Discover how developers are building intelligent applications with browser-native AI.
AI-powered card interpretation for brainstorming, idea generation, card games, and activities like tarot reading. Let users explore and discuss cards with conversational AI.
Chrome extension alpha
Core protocol design
10+ providers integrated
Popular SDK integration
Model routing system
Configuration console
Request playground
Demo projects
Public Browser extension release
Better permissions UX
Gateway node system
Image models
Embeding models
Sharable computing tokens
Desktop Installer
Cloud compute program
Partners announcement
Image, audio & midi models
Better Tool calling
Model management & download tools
WebGPU acceleration for local models
P2P resource pooling
Web standards alignements
Mobile SDK (iOS/Android)
standards always win
Find out all the essential details about WebLLM and how it can serve your needs.
WebLLM is an open protocol that brings AI directly to your browser. Instead of each app having its own AI integration, WebLLM provides a standard SDK that lets users bring their own AI provider to any website.
No! That's exactly the problem WebLLM solves. Instead of paying $20/month to ChatGPT, $10 to Notion AI, $30 to Grammarly — all using the same underlying model — you pay once and use it everywhere through WebLLM.
Privacy is built into the architecture. You control which AI provider handles your data. For maximum privacy, you can run models locally on your own hardware — your data never leaves your device.
Yes! Chrome 138+ includes Gemini Nano — free, on-device AI for desktop browsers. WebLLM automatically uses it when available, with graceful fallback to other providers. Users with Chrome AI get free local AI; everyone else uses their configured provider.
It's just a few lines of code. Install the WebLLM SDK, create a session, and start making requests. No API keys to manage, no billing to set up — WebLLM handles it all through the user's configured provider.
WebLLM supports 30+ providers including OpenAI, Anthropic, Google, Mistral, Groq, Together.ai, OpenRouter, Chrome's built-in Gemini Nano, and local options like Ollama and LM Studio. Users choose their preferred provider and can switch anytime.
WebLLM itself is completely free and open source. Users pay only for their chosen AI provider (or nothing if using Chrome AI or local models). Developers can add AI features to their apps without any infrastructure costs.
AI as a utility needs builders, users, advocates, and believers.
How will you help build the grid?
→ See Quick Start (4 lines of code)
→ View Examples
→ Join Discord
→ Add to Chrome
→ Configure Your AI
→ Browse Normally
→ Tweet Template
→ Email Template
→ Success Stories
This is a community effort, privately funded.
We believe it has the potential to disrupt the market and shift value to small players and individual users.
Utilities aren't built by companies.
They're built by communities.
Are you in?