Bookmark and Share

"Selecting and Exploiting Sensors for Sensor Fusion in Consumer Robots," a Presentation from Daniel Casner

Register or sign in to access the Embedded Vision Academy's free technical training content.

The training materials provided by the Embedded Vision Academy are offered free of charge to everyone. All we ask in return is that you register, and tell us a little about yourself so that we can understand a bit about our audience. As detailed in our Privacy Policy, we will not share your registration information, nor contact you, except with your consent.

Registration is free and takes less than one minute. Click here to register, and get full access to the Embedded Vision Academy's unique technical training content.

If you've already registered, click here to sign in.

See a sample of this page's content below:

Daniel Casner, formerly a systems engineer at Anki, presents the "Selecting and Exploiting Sensors for Sensor Fusion in Consumer Robots" tutorial at the May 2019 Embedded Vision Summit.

How do you design robots that are aware of their unstructured environments at a consumer price point? Excellent sensing is required but using low cost sensors is also necessary. By fusing data from multiple sensors and making sure every sensor does multiple jobs (like the hidden sensors in your camera ISP), we can achieve rolling shutter correction, drivable space understanding and other applications. Careful sensor selection with a focus on ultimate customer utility, data bus architecture with a plan for fusion, and full understanding of the inner workings of each sensor can make all these fusion tasks easier and reduces computational complexity.

This talk covers the pragmatic approach used for hardware design and some of the design choices to enable sensor fusion made across multiple generations of products. Casner also gives an overview of how information is fused from multiple sensors in order to make robots that feel alive and aware of their surroundings.