A robot with a view—how drones and machines can navigate on their own [video]

Robots have captured our imaginations for more than 70 years. When you think about a robot, you might picture Rosie from the Jetsons or a robotic arm in a manufacturing facility. However, the next generation of robots will be very different. These robots and drones will be seamlessly integrated into our everyday lives. Think small flying cameras that will change the way we take selfies, or home-assistant robots that will simplify the tedium of everyday, time-consuming tasks.

Why now? Historically, robots have been limited to industrial settings. But thanks to the same technology powering our smartphones, robots are poised to evolve into intelligent, intuitive machines capable of perceiving their environments, understanding our needs, and helping us in unprecedented ways.

For years Qualcomm Research, the R&D division of Qualcomm Technologies, Inc., has been at the forefront of sensor fusion, computer vision and machine learning technologies—innovations that will enable smarter robots and drones to “see” in 3D, sense their surroundings, avoid collision, and autonomously navigate their environments.

How we’re enabling robots to move autonomously 

In order for a robot to autonomously navigate its environment, it has to accurately estimate its relative position and orientation while moving through an unknown environment. Known as position and orientation tracking with 6 degrees of freedom (6-DOF pose tracking), this capability is essential not only for robotics, but also for many other applications, such as virtual reality (VR), augmented reality (AR) gaming, and indoor navigation.

To solve this challenge, Qualcomm Research uses a technique called Visual-Inertial Odometry (VIO). This fuses information from a camera and inertial sensors, specifically gyroscopes and accelerometers, to estimate device pose without relying on GPS and GNSS (Global Navigation Satellite System).

VIO takes advantage of the complementary strengths of the camera and inertial sensors. For example, a single camera can estimate relative position, but it cannot provide absolute scale—the actual distances between objects, or the size of objects in meters or feet. Inertial sensors provide absolute scale and take measurement samples at a higher rate, thereby improving robustness for fast device motion. However, sensors, particularly low-cost MEMS varieties, are prone to substantial drifts in position estimates when compared with cameras. So VIO blends together the best of both worlds to accurately estimate device pose.

Technology designed for the mobile environment

At Qualcomm Research, we’ve designed VIO from the ground up for power-efficient operation on mobile and embedded devices, and we’ve achieved a very high level of accuracy across a wide variety of device motions and scene environments. All this was made possible through our breakthrough algorithmic innovations and optimizations using the Qualcomm Snapdragon processor’s vector processing and parallel computation abilities. The result? Faster execution time and lower memory consumption.

Our optimizations also made VIO work across a wide range of smartphones, despite several impairments, including rolling shutter, inaccurate sensor time stamps, and limited field of view (FOV) lenses.

Qualcomm Research’s joint work with the University of Pennsylvania’s GRASP Lab is a testament to what’s possible using only a common smartphone. We recently demonstrated in a video the world’s first smartphone autonomously flying a Quad-copter with all processing, including our VIO.

Additionally, the video below illustrates the level of robustness our VIO solution was able to maintain across a wide variety of device motions and scene environments, including walking, running, biking, and VR-like head motions both indoors and outdoors. 

The demo video is using a global shutter camera with a wide FOV lens and accurate time stamps. By using VIO to combine landmark measurements from the camera with inertial sensor measurements in an extended Kalman filter (EKF) framework, we were able to accurately estimate not only the device pose, but also inertial calibration parameters (biases, scale factors, and so on).

What we see is the trajectory of the device projected on a horizontal plane. The device tracks visual landmarks, such as corners, and displays the estimated depth (distance) and associated uncertainty of this estimation (shown as orange-colored numbers). At the end of the video, the device returns to its starting location with a less than 1 percent residual error in computed trajectory—end-to-end drift—for the total trajectory length. This highlights the accuracy and robustness of our VIO solution across different user motions.

By taking advantage of the heterogeneous compute capabilities of Snapdragon, we are further optimizing VIO to enable breakthrough experiences at lower power consumption. VIO is just one way Qualcomm Research is bringing the future forward faster. In upcoming blog posts, we’ll walk you through other technologies we’re developing to build smarter and safer robots, drones, cars, and many other machines and devices. 


转自:https://www.qualcomm.com/news/onq/2015/12/16/robot-view-how-drones-and-machines-can-navigate-their-own

1、资源项目源码均已通过严格测试验证,保证能够正常运行; 2、项目问题、技术讨论,可以给博主私信或留言,博主看到后会第一时间与您进行沟通; 3、本项目比较适合计算机领域相关的毕业设计课题、课程作业等使用,尤其对于人工智能、计算机科学与技术等相关专业,更为适合; 4、下载使用后,可先查看README.md或论文文件(如有),本项目仅用作交流学习参考,请切勿用于商业用途。 5、资源来自互联网采集,如有侵权,私聊博主删除。 6、可私信博主看论文后选择购买源代码。 1、资源项目源码均已通过严格测试验证,保证能够正常运行; 2、项目问题、技术讨论,可以给博主私信或留言,博主看到后会第一时间与您进行沟通; 3、本项目比较适合计算机领域相关的毕业设计课题、课程作业等使用,尤其对于人工智能、计算机科学与技术等相关专业,更为适合; 4、下载使用后,可先查看README.md或论文文件(如有),本项目仅用作交流学习参考,请切勿用于商业用途。 5、资源来自互联网采集,如有侵权,私聊博主删除。 6、可私信博主看论文后选择购买源代码。 1、资源项目源码均已通过严格测试验证,保证能够正常运行; 2、项目问题、技术讨论,可以给博主私信或留言,博主看到后会第一时间与您进行沟通; 3、本项目比较适合计算机领域相关的毕业设计课题、课程作业等使用,尤其对于人工智能、计算机科学与技术等相关专业,更为适合; 4、下载使用后,可先查看README.md或论文文件(如有),本项目仅用作交流学习参考,请切勿用于商业用途。 5、资源来自互联网采集,如有侵权,私聊博主删除。 6、可私信博主看论文后选择购买源代码。
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值