Engineering student Federico Terzi has built an impressive computer interface device reminiscent of a Wiimote.
When talking in person, you can express meaning using facial expressions, and your hands. Usually this acts to add emphasis to a statement or perhaps to point out a certain object, but what if you could actually type letters based on how your hands move?
Terzi’s aptly named “Gesture Keyboard” does just this, using an Arduino Pro Micro, an MPU-6050 accelerometer, and an HC-06 Bluetooth module for sending signals to his laptop. A Python library using Scikit-learn’s SVM (Support Vector Machine) algorithm then translates the motion readings into characters that appear on the screen.
Each Monday is ArduinoMonday here at Adafruit! Be sure to check out our posts, tutorials and new Arduino related products. Adafruit manufactures the Arduino right here in the United States in cooperation with arduino.cc. We have a huge selection of Arduino accessories and all the code and tutorials to get you up and running in no time!
Have an amazing project to share? Join the SHOW-AND-TELL every Wednesday night at 7:30pm ET on Google+ Hangouts.
Join us every Wednesday night at 8pm ET for Ask an Engineer!
Learn resistor values with Mho’s Resistance or get the best electronics calculator for engineers “Circuit Playground” – Adafruit’s Apps!
Maker Business — Alibaba to invest $15b in tech, set up research labs around the world
Wearables — Hand beading mimicry
Electronics — Trigger happy oscilloscope?
Biohacking — Biohacking: Visioneer – AI Glasses to Assist the Visually Impaired
No comments yet.
Sorry, the comment form is closed at this time.