fbpx

“Fundamentals of Monocular SLAM,” a Presentation from Cadence

Shrinivas Gadkari, Design Engineering Director at Cadence, presents the “Fundamentals of Monocular SLAM” tutorial at the May 2019 Embedded Vision Summit.

Simultaneous Localization and Mapping (SLAM) refers to a class of algorithms that enables a device with one or more cameras and/or other sensors to create an accurate map of its surroundings, to determine the device’s location relative to its surroundings and to track its path as it moves through this environment. This is a key capability for many new use cases and applications, especially in the domains of augmented reality, virtual reality and mobile robots.

Monocular SLAM is a type of SLAM that relies exclusively on a monocular image sequence captured by a moving camera. In this talk, Gadkari introduces the fundamentals of monocular SLAM algorithms, from input images to 3D map. He takes a close look at key components of monocular SLAM algorithms, including Oriented Fast and Oriented Brief (ORB), Fundamental Matrix-based Pose Estimation, stitching together poses using translation estimation and loop closure. He also discusses implementation considerations for these components, including arithmetic precision required to achieve acceptable mapping and tracking accuracy.

Here you’ll find a wealth of practical technical insights and expert advice to help you bring AI and visual intelligence into your products without flying blind.

Contact

Address

1646 N. California Blvd.,
Suite 360
Walnut Creek, CA 94596 USA

Phone
Phone: +1 (925) 954-1411
Scroll to Top