Voice-first note capture that integrates with existing tools, designed for capturing fleeting thoughts on the go
How might we help people capture fleeting thoughts and ideas in the moment—without interrupting flow, pulling out their phone, or switching between apps?
How might we design a notes app that adds value without asking users to abandon their existing note-taking systems and workflows?
The world doesn't need another notes app—yet here we are with hundreds of them. Most compete by adding more features, more organization systems, more complexity. But this creates a paradox: the more capable the app, the more friction in capturing a quick thought.
This was a 0→1 mobile-first app designed around a single insight: the best note-taking app is the one you're already using. Rather than compete with Notion, Apple Notes, or Google Keep, this app accepts that users have existing systems—and focuses on solving the specific problem of on-the-go capture.
I interviewed people about their note-taking habits, focusing on when and why they failed to capture thoughts:
Rather than build another full-featured notes app, I focused on two jobs-to-be-done:
Press home button → speak → done. Zero navigation.
Send notes to where they belong in existing systems
This meant accepting that this app wouldn't be the long-term home for notes—it's a capture tool that hands off to existing workflows.
I designed the primary flow around Google Assistant integration, enabling truly hands-free capture:
While voice is primary, the app needed a visual interface for:
I designed this as a lightweight utility—not a destination app. Users should spend 30 seconds here, not 30 minutes.
Press home button, speak, done. No app launching, no navigation, no typing. The entire interaction takes 2-3 seconds, making it possible to capture thoughts without breaking stride—literally.
AI detects intent from natural speech: "Buy milk" becomes a task, "Meeting thoughts" becomes a note, "Call Sarah tomorrow" becomes a reminder. No need to specify format or destination—the system infers from content.
Connect to existing tools (Google Keep, Notion, Todoist, Apple Notes) and configure routing rules. Work notes to Notion, personal to Keep, tasks to Todoist. Notes land where you'll actually see and use them.
Recent captures appear in a simple feed within the app. Quickly scan for transcription errors, re-route misclassified items, or add context before they sync to destination apps.
Within the app, captured items are organized into two tabs: Notes and Tasks. This provides quick visual separation when reviewing recent captures, making it easy to scan what you've recorded and ensure items are categorized correctly.
This was a concept project with functional prototype testing. The following metrics are from user testing with the prototype.
average capture time from trigger to completion
transcription accuracy in varied acoustic environments
of test participants said they'd use this daily
"I've tried voice notes before, but they just sit in a list and I never look at them. This actually gets my thoughts to where I'll use them."— Beta tester, product manager
The most successful product strategy wasn't building the best notes app—it was building the best bridge between capture and existing systems. Users don't want to change their workflows; they want their workflows to work better.
Designing for voice meant rethinking feedback, error states, and confirmation patterns. Visual designers default to screens—voice design requires thinking in time, sound, and minimal interruption. I learned to use haptics and brief audio cues instead of pulling users into the app.
The simpler the interface, the more complex the underlying system needed to be. One-button capture required sophisticated AI, routing logic, and integration management. Simple user experience meant complex product architecture.
If I could revisit this project, I'd spend more time on the "what happens when it's wrong?" scenarios. The design assumes AI categorization works well, but transcription errors and mis-routed notes could be frustrating. I'd build more forgiving correction flows and better confidence indicators for AI decisions. I'd also explore collaborative capture—what if you could quickly share a voice note with someone else without switching apps?