Gesture Wand on the Edge with Particle’s Xenon and TF #Tensorflowlite #machinelearning #AI #artificialintelligence #particle
Particle recently posted several guides for machine learning on MCUs. Most recently, Brandon Satrom created a gesture tracking device using Particle’s Xenon board and Adafriut’s LSM9DS1 3 axis accelerometer. The device is able to differentiate between three gestures “O”, “W”, and “left slope” and lights up an LED on the Xenon board when a gesture is detected. A pre-trained TensorFlow Lite gesture model was used to power the device. This model is a convolutional neural network trained on gesture data from several people.
The model, which was created by the TensorFlow team, is a 20 KB convolutional neural network (or CNN) trained on gesture data from 10 people performing four gestures fifteen times each (ring, wing, slope, and an unknown gesture).
Besides being a fun and easy example of ML for IoT, Satrom also highlights the utility of performing machine learning computations on edge devices:
If you’re working with a high-precision sensor and you need to make a decision based on raw data, you often cannot shoulder the cost of sending all that data to the cloud. And even if you could, the round trip latency runs counter to the frequent need for real-time decision-making.
By performing prediction on a microcontroller, you can process all that raw data on the edge, quickly, without sending a single byte to the cloud.
If you would like to learn more check out the GitHub repo for this work.