Tuesday, November 18, 2025

Where's the future

Voice command has been “the future” for almost 30 years, and somehow it’s still treated like a secret feature hidden behind keyboard shortcuts no one can use without… a keyboard.

Why it feels like they’re resisting voice input

Because in a way, they are.

Voice threatens the old interaction model. If you can control everything with speech, then:

menus become unnecessary

UI complexity becomes exposed

companies lose control over guiding you through their “engagement funnels”

hardware makers don’t get to upsell keyboards, accessories, and touch devices

software makers have to build clean, predictable command structures

So instead of making voice the primary interface, they shoved it into tiny corners:

“press Win+H to dictate”

“press Command twice for Siri”

“press this obscure function key”

“open this random menu to enable the feature no one told you existed”

All of it assumes a keyboard — which is absurd in 2025, when half our devices aren’t meant to have one.

The deeper problem

Voice was never treated as a core, first-class input method (except on phones). On desktop OSes, it’s an add-on, almost like accessibility tech they grudgingly allow rather than embrace.

The infrastructure for full voice navigation is there — the OS can hear you, transcribe you, and run commands — but the interface is stuck in a 1990s mindset:

> “Voice exists only after the keyboard gives permission.”

What you actually want

And what the industry has failed to give:

A way to say “Start listening” without touching anything

A unified, universal voice command layer, not 12 separate systems

A voice-first mode that bypasses the UI entirely

A dictation system you can launch with your voice, not a hidden hotkey

A device that trusts your intention instead of requiring ritual combinations of inputs

We should have had that by 2005.
We still don’t.

No comments:

Post a Comment