AI & Local

Why local AI is the future of personal productivity — and how Bun Agents fits in

Bun Agents Team April 2025 8 min read

The year 2025 marked an inflection point. AI productivity tools are no longer a curiosity — they are embedded in virtually every software category. Microsoft Copilot sits inside Word, Excel, and Teams. Notion AI rewrites your pages and summarizes meeting notes. Google Duet answers questions inside Gmail and Docs. The numbers are staggering: over 77% of the Fortune 500 are now paying for access to Microsoft's Copilot suite, and the global AI-in-productivity-software market is projected to exceed $35 billion by 2027.

On the surface, this looks like a golden era for knowledge workers. But underneath the polished interfaces and impressive demos lies a structural problem that nobody talks about in the product launch videos: every single one of these tools sends your data to the cloud.

The AI productivity boom — and the privacy trade-off hiding in plain sight

When you ask Notion AI to summarize your meeting notes, those notes travel over the network to a remote server — likely running GPT-4o or Claude — where they are processed and returned to you. When Microsoft Copilot drafts an email based on your calendar context and recent conversations, your calendar, your conversations, and your writing style are all inputs to a model query happening on Microsoft's infrastructure.

This is not a bug. It is, by design, how cloud AI works. Large language models require significant compute resources — a 70-billion-parameter model running inference on a consumer device would be impractical. Sending data to the cloud and getting results back is currently the only way to deliver that level of capability at scale.

But what does "sending your data to the cloud" actually mean in practice? It means:

Notion's privacy policy, for instance, explicitly states that user content may be processed by third-party AI providers. Microsoft's data processing addendum for Copilot runs to dozens of pages. Most users never read any of it — and most users do not realize that every "smart suggestion" represents a data transfer event.

The fundamental tension: AI features are genuinely useful. But the useful ones today require your data to leave your device. This is a trade-off most productivity apps are quietly asking you to make without telling you clearly.

GDPR, training data, and the consent problem

The regulatory situation is complex and evolving. The EU's GDPR requires a lawful basis for processing personal data. When you enable an AI feature in a productivity app, does clicking "Enable AI" constitute valid informed consent for your data to be sent to a third-party model provider? The answer is genuinely unclear — and several European data protection authorities are actively investigating.

Italy temporarily blocked ChatGPT in early 2023 over training data concerns. The Irish Data Protection Commission has been probing multiple AI providers on data retention and training practices. The UK's ICO published guidance in March 2024 specifically warning organizations about the risks of using cloud AI tools with employee data.

For individual users, the concern is less about fines and more about trust. When you journal about a difficult period in your life inside Notion, you are trusting that content won't resurface somewhere unexpected. When you log your habits — including the ones you're not proud of — in a productivity app with AI features, that data deserves to stay private. Today, it largely does not.

The on-device AI wave: what's actually changing in 2025

Here is where the story gets genuinely exciting. The last 18 months have seen a dramatic shift in what is possible on consumer hardware. Models are getting dramatically smaller without proportionally sacrificing capability, and browser technology has evolved to support running inference natively — with no server involved.

Gemini Nano and Chrome's built-in AI

Starting with Chrome 120, Google shipped Gemini Nano directly into the browser. This is not a gimmick. Gemini Nano is a 1.8-billion-parameter model that runs entirely on-device, accessed via the Chrome built-in AI API (window.ai). It can summarize text, answer questions about page content, and assist with writing — all without a network request. Google has opened this API to developers through the Chrome origin trial program, and it is now available in stable Chrome on supported hardware.

The hardware requirement matters: Gemini Nano requires a device with at least 22 GB of storage and a GPU with 4 GB of VRAM. That limits deployment to mid-range and above consumer devices — but the set of qualifying devices expands every year.

WebLLM: running LLMs entirely in the browser with WebGPU

The MLC AI project's WebLLM library is one of the most impressive pieces of browser technology released in recent years. WebLLM uses WebGPU — the modern successor to WebGL for GPU compute — to run quantized LLMs directly in the browser, with performance approaching native inference speeds on supported hardware.

WebLLM has already demonstrated running Llama 3.2 1B and 3B in-browser at usable speeds on consumer GPUs. The models are downloaded once and cached locally. After the initial download (which ranges from 600 MB for a 1B model to 2 GB+ for a 3B), every inference request happens entirely on-device. No API key required. No network round-trip. No data leaves the browser.

The MLC AI project supports a growing catalog of models, including:

Phi-3-mini: Microsoft's most interesting local AI bet

Microsoft's Phi-3-mini deserves special mention. It is a 3.8-billion-parameter model trained with a novel "textbooks are all you need" methodology — using extremely high-quality synthetic training data to achieve performance that rivals models 10x its size on many benchmarks. Phi-3-mini can be quantized to run via ONNX Runtime Web in the browser with reasonable performance on modern hardware.

The practical implication: you can summarize a 2,000-word document, extract action items, or generate a short draft email entirely in-browser, on your own hardware, with a model that has never seen your data and never will.

What "local AI for productivity" actually looks like

The marketing vision of AI productivity often involves magical assistants that anticipate your needs. The realistic near-term picture of local AI for productivity is more grounded — but also more private, more reliable, and more trustworthy:

None of these require a 100-billion-parameter model. They are well within the capability range of the models running in browsers today, and they can be done with zero data transmission.

Local AI capability checklist for 2025

  • Text summarization (1,000–5,000 words): available today via WebLLM / Gemini Nano
  • Sentiment and pattern analysis on structured data: available today
  • Short-form text generation and rewriting: available today
  • Semantic similarity search (embeddings): available today via ONNX Runtime Web
  • Full document Q&A with large context windows: emerging (limited by VRAM)
  • Real-time voice transcription on-device: available on supported devices (Whisper via WebAssembly)

Our roadmap: local AI features coming to Bun Agents

Bun Agents is already built on an offline-first architecture. Your data lives in an OPFS-backed SQLite database inside your browser — it never touches our servers (unless you explicitly opt into cloud sync). This foundation makes us uniquely positioned to layer local AI capabilities on top without any architectural compromise.

Here is what we are building toward:

Private smart suggestions

We are exploring using on-device embedding models to understand the semantic content of your notes and tasks — and surface related items you might have forgotten, suggest tags automatically, and detect when two tasks describe the same underlying goal. All of this will run locally, using the OPFS database as its source, with no data ever transmitted.

Offline habit insights

Your habit data contains patterns you cannot see at a glance. We are building a local analysis layer — using small statistical models and, where hardware allows, on-device LLMs — that can surface weekly summaries of your habit streaks, identify your highest-risk days, and suggest timing adjustments based on historical completion rates. Your habit journal is among the most personal data you create. It should stay with you.

Local summarization of notes and journals

For users on capable hardware (a mid-range desktop or a recent MacBook with Apple Silicon), we plan to integrate WebLLM to enable one-click summarization of your notes. Select a week, a project, or a time range — and get a concise summary generated locally in seconds. For users on less capable hardware, we will provide a graceful fallback that uses statistical summarization (extractive rather than generative) with no model download required.

The fundamental argument: AI should augment your private data, not harvest it

There is a version of AI-powered productivity that is genuinely empowering: one where intelligent features help you understand your own patterns, surface your own forgotten ideas, and make sense of your own data — without ever requiring you to hand that data to a third party.

That version is not science fiction. The models are small enough, the browser APIs are mature enough, and the devices in people's pockets are powerful enough. The only thing standing between today's cloud-AI status quo and a genuinely private AI-powered productivity future is product decisions — and a willingness to do the harder engineering work of making local inference actually fast and useful.

We think the trade-off is worth it. Your notes about your goals, your struggles, your health, your relationships — these deserve the same protection as your messages and your financial data. AI features do not require you to give that up. The wave of on-device AI is here. We plan to ride it.

Interested in following our progress on local AI features? We post updates on the blog and announce new capabilities in-app. Try Bun Agents free — no sign-up required to start.

Try Bun Agents — your data stays yours

Offline-first productivity tools that work without an internet connection. Notes, habits, tasks, and goals — all stored locally on your device.

Open Bun Agents