I'm sure there are lots of fun things to do with an Arduino that could recognize gestures; I'm very surprised nobody's "excited" by this?
For example, I made a photobooth simple device based on a Pi and a regular DSLR; the person takes a photo by pressing a button, then the Pi checks the camera and sends the latest image in a web gallery somewhere. The problem is the button: it needs a remote. If wired, it risks destroying the whole apparatus if someone pulls on the cord; if wireless, it risks being lost.
A gesture-based trigger would be super cool; but having it run on the Pi through the DSLR risks damaging the camera, so it should run on a different device, cheap and not too power hungry, such as an Arduino.
What about some kind of MIDI-controller based on an Arduino that could recognize gestures?
I also made a webapp to learn sight-reading (babeloop.com) but it requires users clicking or tapping on the screen, which is not natural.
It would be cool to be able to listen to users reading notes out loud and detect if they're right or wrong, on the fly, locally in a browser or on a phone. A light general speech recognition model such as VOSK is 40Mb and is able to recognize most phonemes; but for music sight reading, there are only 7 syllabes, so one should be able to make a much smaller model? I don't know how hard it would be though to train my own model...?
Embedded platforms generally have very little memory and very few hardware threads.
Tiny SoCs will likely need accelerators to run more interesting models. For example, ARM is working on a ML coprocessor to pair with their Cortex-M chips.
> I'm sure there are lots of fun things to do with an Arduino that could recognize gestures
Arduinos can recognize gestures, faces, etc. right now with well-established tech, no servers needed. I've had a couple of robots running around my place that do this sort of thing for a number of years now.
For example, I made a photobooth simple device based on a Pi and a regular DSLR; the person takes a photo by pressing a button, then the Pi checks the camera and sends the latest image in a web gallery somewhere. The problem is the button: it needs a remote. If wired, it risks destroying the whole apparatus if someone pulls on the cord; if wireless, it risks being lost.
A gesture-based trigger would be super cool; but having it run on the Pi through the DSLR risks damaging the camera, so it should run on a different device, cheap and not too power hungry, such as an Arduino.
What about some kind of MIDI-controller based on an Arduino that could recognize gestures?
I also made a webapp to learn sight-reading (babeloop.com) but it requires users clicking or tapping on the screen, which is not natural.
It would be cool to be able to listen to users reading notes out loud and detect if they're right or wrong, on the fly, locally in a browser or on a phone. A light general speech recognition model such as VOSK is 40Mb and is able to recognize most phonemes; but for music sight reading, there are only 7 syllabes, so one should be able to make a much smaller model? I don't know how hard it would be though to train my own model...?