Motion Gestures provides powerful embedded AI-based gesture recognition software for different sensors. Unlike conventional solutions, our platform does not require any training data collection or programming and utilizes advanced machine learning algorithms. As a result, gesture software development time and costs are reduced by 10x while gesture recognition accuracy is increased to nearly 100%. We currently support touch, motion (i.e. IMU), and vision sensors. In the camera category, we support RGB (i.e. color), NIR (i.e Near Infrared), and Depth (i.e. 3D) cameras.
Motion Gestures’ software can be used to add sophisticated gestures-based user interface to any product using touch, motion, or vision sensors. We support applications in all major verticals involving any type of gesture, whether static or dynamic. Software’s capabilities can be scaled according to deployment hardware. For camera-based systems, sophisticated hand tracking and gesture recognition is available using a 21 joints-based skeleton that provides positional coordinates of all joints in real-time. Free SDK is available for evaluation upon request. For demos of our technology, please visit our YouTube channel.
Gesture recognition software is available for all Synopsys ARC® processors.
Learn more about how Motion Gestures and Synopsys work together.
Quickly identify and access the right IP solutions for your project needs.
Find embedded memory and logic IP for your SoC design.
Find silicon-proven NVM IP for your SoC design needs.