Gesture and Positional Tracking

Take control of your reality

Gesture and positional interactions are key to making your VR experience more lifelike and immersive. To accurately capture a user's movement, parallel data capture and processing is critical.

Lattice's wide range of programmable product offerings provides the ability to capture and pre-process data from multiple video streams/sensor arrays while easily balancing hardware and software computation.

Jump to

Block Diagrams

Sensor-based Gesture Tracking with iCE40

  • Concurrent multi sensor acquisition and data processing to enable precise and accurate tracking.
  • Low power, small form factor, low cost.
  • Flexibility to to integrate computation or perform simple pre-processing such as time-stamping.

Click here for more information on Sensor Management

Visual-based Gesture Tracking

  • Multiple product families capable of directly interfacing with multiple image sensors through HiSPi, sub-LVDS or MIPI CSI-2 interface protocols.
  • Readily available video pre-processing IP and development kits to simplify your design.
  • Wide range of programmable logic densities and integrated embedded features to accommodate your algorithm.

See below for more information

Design Resources for Gesture and Positional Tracking

Support

Technical Support

Need Help? We're Here to Assist You

Quality & Reliability

Reference Material to Help Answer Your Questions

Like most websites, we use cookies and similar technologies to enhance your user experience. We also allow third parties to place cookies on our website. By continuing to use this website you consent to the use of cookies as described in our Cookie Policy.