当前位置: X-MOL 学术Optik › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
An improved visual SLAM based on affine transformation for ORB feature extraction
Optik ( IF 3.1 ) Pub Date : 2020-11-23 , DOI: 10.1016/j.ijleo.2020.165421
Lecai Cai , Yuling Ye , Xiang Gao , Zhong Li , Chaoyang Zhang

Aiming at the problems of the existing robot vision SLAM(Simultaneous Localization and Mapping), such as the small number of feature point extraction and the easy loss of keyframes, which leads to the problem of trajectory deviation, many existing visual SLAM methods based on keyframes only propose a holistic system solution in the scheme, no detailed research is carried out on the feature extraction of the front-end visual odometry. In this paper, an affine transformation based ORB feature extraction method(Affine-ORB) is used and applied to existing robot vision SLAM methods, and an improved visual SLAM method is proposed. In the proposed SLAM, we first use the BRISK method for feature point description; Secondly, the mathematical method of affine transformation is introduced into the ORB feature extraction; Finally, the sample is normalized and the image is restored. By requiring a handheld camera to take the vision SLAM experiment, it is judged that the keyframe loss rate of the proposed algorithm is significantly reduced. Through evaluation experiments with TUM, KITTI and EUROC datasets, the keyframe extraction effect and positioning accuracy of the SLAM algorithm set out in the present paper are compared with PTAM, LSD-SLAM and ORB-SLAM, respectively. The frame loss rate of feature extraction SLAM based on affine transformation has decreased from 0.5 % to 0.2 %, and the root mean square error (RMSE) of the running trajectory has been drastically reduced. That means the key frame extraction speed is faster, the keyframe loss rate is lower at the same moving speed, and the positioning accuracy is higher



中文翻译:

一种基于仿射变换的改进的Visual SLAM用于ORB特征提取

针对现有的机器人视觉SLAM(同时定位定位)问题,如特征点提取数量少,关键帧容易丢失等问题,从而导致轨迹偏差问题。目前已有很多基于关键帧的视觉SLAM方法。该方案仅提出了整体系统解决方案,对前端视觉里程表的特征提取没有进行详细研究。本文提出了一种基于仿射变换的ORB特征提取方法(Affine-ORB),并将其应用于现有的机器人视觉SLAM方法,并提出了一种改进的视觉SLAM方法。在提出的SLAM中,我们首先使用BRISK方法进行特征点描述。其次,将仿射变换的数学方法引入ORB特征提取中。最后,对样本进行归一化,然后还原图像。通过要求手持相机进行视觉SLAM实验,可以判断所提出算法的关键帧丢失率显着降低。通过对TUM,KITTI和EUROC数据集的评估实验,将本文提出的SLAM算法的关键帧提取效果和定位精度分别与PTAM,LSD-SLAM和ORB-SLAM进行了比较。基于仿射变换的特征提取SLAM的帧丢失率从0.5%降低到0.2%,并且运行轨迹的均方根误差(RMSE)大大降低。这意味着在相同的移动速度下关键帧提取速度更快,关键帧丢失率更低,定位精度更高 通过要求手持相机进行视觉SLAM实验,可以判断所提出算法的关键帧丢失率显着降低。通过对TUM,KITTI和EUROC数据集的评估实验,将本文提出的SLAM算法的关键帧提取效果和定位精度分别与PTAM,LSD-SLAM和ORB-SLAM进行了比较。基于仿射变换的特征提取SLAM的帧丢失率从0.5%降低到0.2%,并且运行轨迹的均方根误差(RMSE)大大降低。这意味着在相同的移动速度下关键帧提取速度更快,关键帧丢失率更低,定位精度更高 通过要求手持相机进行视觉SLAM实验,可以判断所提出算法的关键帧丢失率显着降低。通过对TUM,KITTI和EUROC数据集的评估实验,将本文提出的SLAM算法的关键帧提取效果和定位精度分别与PTAM,LSD-SLAM和ORB-SLAM进行了比较。基于仿射变换的特征提取SLAM的帧丢失率从0.5%降低到0.2%,并且运行轨迹的均方根误差(RMSE)大大降低。这意味着在相同的移动速度下关键帧提取速度更快,关键帧丢失率更低,定位精度更高 判断该算法的关键帧丢失率明显降低。通过对TUM,KITTI和EUROC数据集的评估实验,将本文提出的SLAM算法的关键帧提取效果和定位精度分别与PTAM,LSD-SLAM和ORB-SLAM进行了比较。基于仿射变换的特征提取SLAM的帧丢失率从0.5%降低到0.2%,并且运行轨迹的均方根误差(RMSE)大大降低。这意味着关键帧提取速度更快,在相同移动速度下关键帧丢失率更低,并且定位精度更高 判断该算法的关键帧丢失率明显降低。通过对TUM,KITTI和EUROC数据集的评估实验,将本文提出的SLAM算法的关键帧提取效果和定位精度分别与PTAM,LSD-SLAM和ORB-SLAM进行了比较。基于仿射变换的特征提取SLAM的帧丢失率从0.5%降低到0.2%,并且运行轨迹的均方根误差(RMSE)大大降低。这意味着在相同的移动速度下关键帧提取速度更快,关键帧丢失率更低,定位精度更高 将本文提出的SLAM算法的关键帧提取效果和定位精度分别与PTAM,LSD-SLAM和ORB-SLAM进行了比较。基于仿射变换的特征提取SLAM的帧丢失率从0.5%降低到0.2%,并且运行轨迹的均方根误差(RMSE)大大降低。这意味着关键帧提取速度更快,在相同移动速度下关键帧丢失率更低,并且定位精度更高 将本文提出的SLAM算法的关键帧提取效果和定位精度分别与PTAM,LSD-SLAM和ORB-SLAM进行了比较。基于仿射变换的特征提取SLAM的帧丢失率从0.5%降低到0.2%,并且运行轨迹的均方根误差(RMSE)大大降低。这意味着在相同的移动速度下关键帧提取速度更快,关键帧丢失率更低,定位精度更高

更新日期:2020-12-04
down
wechat
bug