I'm able bodied, but years back suffered tremendous pain, that led to me having to lie down on my back and use the computer with a keyboard. And TBH even that was a pain. I got quite far under OSX tiger, and browse the web. Use a terminal etc. But at the time it led me into thinking about alternative interfaces. And the one thing that did catch my eye was 'dasher' for text entry. Which could be controlled with a simple pointer. And it did make me wonder how far you could go with a pointer based interface. I wasn't that fast while trying it out. But could construct sentences with a bit of effort.

https://www.inference.org.uk/dasher/

I've been thinking about it some more recently, as AI could assist with word completion, and augmenting it with some AI helpers could really improve things. And if the idea was extended.

Not wanting to sound like an able-ist snob, but interfacing with traditional computers with a keyboard is hard. And the interface is clumsy. Even touch devices and keyboards. My Mum could never use a computer, but she has worked out how to use a tablet and find videos on Youtube. I keep meaning to introduce her to voice input. As this would really benefit her.

Despite the huge tech leaps with smartphones and tablets and what not, I do feel there has been a huge regression in basic communication between people only exacerbated by the pandemic. There's knowledge available easily at people's fingertips and that's great, along with new channels of communication. But text based comms have retarded many people.

Since when we talk, our tongues tap patterns on the roof of the mouth and the back of the teeth, I wonder if AI processing could infer what words you are shaping from these sensors. Maybe it’s possible to input text by mouthing words silently, but without opening your mouth. Kind of like how it’s possible to eavesdrop from just the sound of keyboard clicks:

https://github.com/ggerganov/kbd-audio