Exactly, voice works amazingly as a "3rd interface" complementing keyboards and mice which aren't completely accurate, either, fwiw. Voice gives us more speed, frees up UI clutter and can free up our hands.
Oops I was a bit unclear. I meant, include the engine part that matches English statements to commands and parameters but hook it up to a keyboard-driven text box instead of a microphone. So it's:
written text->'AI'->command
vs.
speech->'AI'->command
(Edit: and sure, voice as an adjunct per my first comment.)
I'm working on a tool that explores this area more: https://www.lipsurf.com