ORB-SLAM2源码解析(一):ORB算法

ORB-SLAM2源码解析(一):ORB算法


目录

一、前言

二、ORB算法原理

三、ORB-SLAM2中ORB算法代码注释分析

四、与opencv中ORB对比测试

五、总结

六、参考


一、前言

        目前在学习ORB-SLAM2中,看了一段时间却仅仅学到皮毛,于是决定深入代码,逐步蚕食各个部分,今天从ORB-SLAM2中使用的ORB算法开始。这里记录我的学习过程,同时增加了一些个人理解注释。


二、ORB算法原理

         ORB算法原理主要来自于文章《ORB an efficient alternative to SIFT or SURF》。在参考文献(1)中对该文献进行了大概的翻译,这里不在再重复,总结来说ORB是基于FAST角点和BRIEF描述子,准确来说也加入了Harris角点的角点评分方法,不过在ORB-SLAM2代码中并没用到,ORB-SLAM2使用的它的贡献在与:

  1. 使用灰度质心法计算了角点的方向,为描述子增加了旋转不变性;
  2. 使用Harris角点的角点评分方法为FAST角点增加角点评分,(在opencv可以选择使用Harris角点的角点评分方法还是使用Harris角点的角点评分方法)。

  如果你对BRIEF特征点描述子、FAST和Harris角点没有基础的认识,参考文章2、3、4。

  总的原理以文章1为主,接下来进入代码分析,这里给出代码的完全注释,具体参考了文章5\6\7\8,有兴趣的可以看一看。


三、ORB-SLAM2中ORB算法代码注释分析

#include <opencv2/core/core.hpp>
#include <opencv2/highgui/highgui.hpp>
#include <opencv2/features2d/features2d.hpp>
#include <opencv2/imgproc/imgproc.hpp>
#include <vector>
#include <iterator>
 
#include "ORBextractor.h"
#include <iostream>
 
 
using namespace cv;
using namespace std;
 
namespace ORB_SLAM2
{
 
const int PATCH_SIZE = 31;//用来做灰度质心的圆圆的直径
const int HALF_PATCH_SIZE = 15;//圆的半径15+15+1(1为中心点的那个像素)
const int EDGE_THRESHOLD = 19; //边界阈值
 
 
//灰度质心法(IC)计算特征的旋转
static float IC_Angle(const Mat& image, Point2f pt,  const vector<int> & u_max)
{
    int m_01 = 0, m_10 = 0;
 
    const uchar* center = &image.at<uchar> (cvRound(pt.y), cvRound(pt.x));  //cvRound 返回跟参数最接近的整数值;
 
//我们要在一个圆域中算出m10(u)和m01(v),计算步骤是先算出v=0的m10,然后在平行于x轴算出m10和m01,一次计算相当于图像中的同个颜色的两个line。参考文章7
// Treat the center line differently, v=0   横坐标:-15-----+15
    for (int u = -HALF_PATCH_SIZE; u <= HALF_PATCH_SIZE; ++u)
        m_10 += u * center[u];
 
    // Go line by line in the circuI853lar patch
    int step = (int)image.step1();    //opencv中概念,计算每行的元素个数
    for (int v = 1; v <= HALF_PATCH_SIZE; ++v)
    {
        // Proceed over the two lines
        int v_sum = 0;
        int d = u_max[v];
        for (int u = -d; u <= d; ++u)
        {
            int val_plus = center[u + v*step], val_minus = center[u - v*step];
            v_sum += (val_plus - val_minus);
            m_10 += u * (val_plus + val_minus);
        }
        m_01 += v * v_sum;
    }
//返回计算的角度
    return fastAtan2((float)m_01, (float)m_10);
}
 
 
//弧度制与角度的转换
const float factorPI = (float)(CV_PI/180.f);
 
//计算描述子
static void computeOrbDescriptor(const KeyPoint& kpt,
                                 const Mat& img, const Point* pattern,
                                 uchar* desc)
{
    float angle = (float)kpt.angle*factorPI;
    float a = (float)cos(angle), b = (float)sin(angle);
 
    const uchar* center = &img.at<uchar>(cvRound(kpt.pt.y), cvRound(kpt.pt.x));
    const int step = (int)img.step;
 
    #define GET_VALUE(idx) \
        center[cvRound(pattern[idx].x*b + pattern[idx].y*a)*step + \
               cvRound(pattern[idx].x*a - pattern[idx].y*b)]//这里使用旋转不变性
 
 
    for (int i = 0; i < 32; ++i, pattern += 16)
    {
       。。。
    }
 
    #undef GET_VALUE//取消宏定义
}
 
 //这里的点是经过机器学习训练过的点,用来做二进制描述子提取,具体参考文章2BRIEF描述子
static int bit_pattern_31_[256*4] =
{
   。。。。
};
 
ORBextractor::ORBextractor(int _nfeatures, float _scaleFactor, int _nlevels,
         int _iniThFAST, int _minThFAST):
    nfeatures(_nfeatures), scaleFactor(_scaleFactor), nlevels(_nlevels),
    iniThFAST(_iniThFAST), minThFAST(_minThFAST)
{
/***********确定每一层的特征点数,采用等比数列**************/
//定义每一层的尺度和逆尺度
    mvScaleFactor.resize(nlevels);
    mvLevelSigma2.resize(nlevels);
    mvScaleFactor[0]=1.0f;
    mvLevelSigma2[0]=1.0f;
    for(int i=1; i<nlevels; i++)
    {
        mvScaleFactor[i]=mvScaleFactor[i-1]*scaleFactor;
        mvLevelSigma2[i]=mvScaleFactor[i]*mvScaleFactor[i];
    }
 
    mvInvScaleFactor.resize(nlevels);
    mvInvLevelSigma2.resize(nlevels);
    for(int i=0; i<nlevels; i++)
    {
        mvInvScaleFactor[i]=1.0f/mvScaleFactor[i];
        mvInvLevelSigma2[i]=1.0f/mvLevelSigma2[i];
    }
 
    mvImagePyramid.resize(nlevels);
 
    mnFeaturesPerLevel.resize(nlevels);
    float factor = 1.0f / scaleFactor;
 
//第一层特征点数,以后每一层成等比数列,利用的是等比数列求和公式
	float nDesiredFeaturesPerScale = nfeatures*(1 - factor)/(1 - (float)pow((double)factor, (double)nlevels));
//所有层数的特征点数量加起来是nfeatures
    int sumFeatures = 0;
    for( int level = 0; level < nlevels-1; level++ )
    {
        mnFeaturesPerLevel[level] = cvRound(nDesiredFeaturesPerScale);  //取整
        sumFeatures += mnFeaturesPerLevel[level];
        nDesiredFeaturesPerScale *= factor;
    }
    mnFeaturesPerLevel[nlevels-1] = std::max(nfeatures - sumFeatures, 0);//保证特征点数
 
 
//复制训练的模板
    const int npoints = 512;//256对点*2
    const Point* pattern0 = (const Point*)bit_pattern_31_;//这里的点是经过机器学习训练过的点,用来做二进制描述子提取
    std::copy(pattern0, pattern0 + npoints, std::back_inserter(pattern));
 
    //This is for orientation
    // pre-compute the end of a row in a circular patch
   //计算方向时,每个v对应的最大的u坐标
    umax.resize(HALF_PATCH_SIZE + 1);
 
// 将v坐标划分为两部分进行计算,主要为了确保计算特征主方向的时候,x,y方向对称
    int v, v0, vmax = cvFloor(HALF_PATCH_SIZE * sqrt(2.f) / 2 + 1);//cvFloor含义是取不大于参数的最大整数值  
    int vmin = cvCeil(HALF_PATCH_SIZE * sqrt(2.f) / 2);           //cvCeil含义是取不小于参数的最小整数值
 
//利用勾股定理计算坐标
	const double hp2 = HALF_PATCH_SIZE*HALF_PATCH_SIZE;   //patch圆半径的平方
    for (v = 0; v <= vmax; ++v)
        umax[v] = cvRound(sqrt(hp2 - v * v));  //每一个v坐标,最大的U坐标
 
    // Make sure we are symmetric  确保是圆
    for (v = HALF_PATCH_SIZE, v0 = 0; v >= vmin; --v)
    {
        while (umax[v0] == umax[v0 + 1])
            ++v0;
        umax[v] = v0;
        ++v0;
    }
}
 
 
//计算每个关键点的角度
static void computeOrientation(const Mat& image, vector<KeyPoint>& keypoints, const vector<int>& umax)
{
    for (vector<KeyPoint>::iterator keypoint = keypoints.begin(),
         keypointEnd = keypoints.end(); keypoint != keypointEnd; ++keypoint)
    {
        keypoint->angle = IC_Angle(image, keypoint->pt, umax);
    }
}
 
void ExtractorNode::DivideNode(ExtractorNode &n1, ExtractorNode &n2, ExtractorNode &n3, ExtractorNode &n4)
{
    const int halfX = ceil(static_cast<float>(UR.x-UL.x)/2);
    const int halfY = ceil(static_cast<float>(BR.y-UL.y)/2);
 
    //Define boundaries of childs
    n1.UL = UL;
    n1.UR = cv::Point2i(UL.x+halfX,UL.y);
    n1.BL = cv::Point2i(UL.x,UL.y+halfY);
    n1.BR = cv::Point2i(UL.x+halfX,UL.y+halfY);
    n1.vKeys.reserve(vKeys.size());
 
    n2.UL = n1.UR;
    n2.UR = UR;
    n2.BL = n1.BR;
    n2.BR = cv::Point2i(UR.x,UL.y+halfY);
    n2.vKeys.reserve(vKeys.size());
 
    n3.UL = n1.BL;
    n3.UR = n1.BR;
    n3.BL = BL;
    n3.BR = cv::Point2i(n1.BR.x,BL.y);
    n3.vKeys.reserve(vKeys.size());
 
    n4.UL = n3.UR;
    n4.UR = n2.BR;
    n4.BL = n3.BR;
    n4.BR = BR;
    n4.vKeys.reserve(vKeys.size());
 
    //Associate points to childs
    for(size_t i=0;i<vKeys.size();i++)
    {
        const cv::KeyPoint &kp = vKeys[i];
        if(kp.pt.x<n1.UR.x)
        {
            if(kp.pt.y<n1.BR.y)
                n1.vKeys.push_back(kp);
            else
                n3.vKeys.push_back(kp);
        }
        else if(kp.pt.y<n1.BR.y)
            n2.vKeys.push_back(kp);
        else
            n4.vKeys.push_back(kp);
    }
 
    if(n1.vKeys.size()==1)
        n1.bNoMore = true;
    if(n2.vKeys.size()==1)
        n2.bNoMore = true;
    if(n3.vKeys.size()==1)
        n3.bNoMore = true;
    if(n4.vKeys.size()==1)
        n4.bNoMore = true;
 
}
 
 
//计算fast特征点并进行筛选
//先用(maxX-minX)/(maxY-minY)来确定四叉数有几个初始节点,这里有 bug,如果输入的是一张宽高比小于0.5 的图像,
//nIni 计算得到 0,下一步计算 hX 会报错,例如round(640/480)=1,所以只有一个初始节点,(UL,UR,BL,BR)就会分布到被裁掉边界后的图像的四个角。
//把所有的关键点分配给属于它的节点,当节点所分配到的关键点的个数为1时就不再进行分裂,当节点没有分配到关键点时就删除此节点。
//再根据兴趣点分布,利用四叉树方法对图像进行划分区域,当bFinish的值为true时就不再进行区域划分,首先对目前的区域进行划分,
 
//把每次划分得到的有关键点的子区域设为新的节点,将nToExpand参数加一,并插入到节点列表的前边,删除掉其父节点。
//只要新节点中的关键点的个数超过一个,就继续划分,继续插入列表前面,继续删除父节点,直到划分的子区域中的关键点的个数是一个,
//然后迭代器加以移动到下一个节点,继续划分区域。
 
//当划分的区域即节点的个数大于关键点的个数或者分裂过程没有增加节点的个数时就将bFinish的值设为true,不再进行划分。
//如果以上条件没有满足,但是满足((int)lNodes.size()+nToExpand*3)>N,则说明将所有节点再分裂一次可以达到要求。
//vSizeAndPointerToNode 是前面分裂出来的子节点(n1, n2, n3, n4)中可以分裂的节点。
//按照它们特征点的排序,先从特征点多的开始分裂,分裂的结果继续存储在 lNodes 中。每分裂一个节点都会进行一次判断,
//如果 lNodes 中的节点数量大于所需要的特征点数量,退出整个 while(!bFinish) 循环,如果进行了一次分裂,
//并没有增加节点数量,不玩了,退出整个 while(!bFinish) 循环。取出每一个节点(每个区域)对应的最大响应点,即我们确定的特征点。
 
 
vector<cv::KeyPoint> ORBextractor::DistributeOctTree(const vector<cv::KeyPoint>& vToDistributeKeys, const int &minX,
                                       const int &maxX, const int &minY, const int &maxY, const int &N, const int &level)
{
    // Compute how many initial nodes 计算有多少初始节点  
    const int nIni = round(static_cast<float>(maxX-minX)/(maxY-minY));  //round四舍五入取整,水平划分格子的数量
 
    const float hX = static_cast<float>(maxX-minX)/nIni;       //水平划分格子的宽度
 
    list<ExtractorNode> lNodes;
 
    vector<ExtractorNode*> vpIniNodes;
    vpIniNodes.resize(nIni);
 
    for(int i=0; i<nIni; i++)
    {
        ExtractorNode ni;
        ni.UL = cv::Point2i(hX*static_cast<float>(i),0);
        ni.UR = cv::Point2i(hX*static_cast<float>(i+1),0);
        ni.BL = cv::Point2i(ni.UL.x,maxY-minY);
        ni.BR = cv::Point2i(ni.UR.x,maxY-minY);
        ni.vKeys.reserve(vToDistributeKeys.size());
 
        lNodes.push_back(ni);
        vpIniNodes[i] = &lNodes.back();
    }
 
    //Associate points to childs
    for(size_t i=0;i<vToDistributeKeys.size();i++)
    {
        const cv::KeyPoint &kp = vToDistributeKeys[i];
        vpIniNodes[kp.pt.x/hX]->vKeys.push_back(kp);
    }
 
    list<ExtractorNode>::iterator lit = lNodes.begin();
 
    while(lit!=lNodes.end())
    {
        if(lit->vKeys.size()==1)
        {
            lit->bNoMore=true;
            lit++;
        }
        else if(lit->vKeys.empty())
            lit = lNodes.erase(lit);
        else
            lit++;
    }
 
    bool bFinish = false;
 
    int iteration = 0;
 
    vector<pair<int,ExtractorNode*> > vSizeAndPointerToNode;
    vSizeAndPointerToNode.reserve(lNodes.size()*4);
 
    // 根据兴趣点分布,利用N叉树方法对图像进行划分区域
    while(!bFinish)
    {
        iteration++;
 
        int prevSize = lNodes.size();
 
        lit = lNodes.begin();
 
        int nToExpand = 0;
 
        vSizeAndPointerToNode.clear();
 
        // 将目前的子区域经行划分
        while(lit!=lNodes.end())
        {
            if(lit->bNoMore)
            {
                // If node only contains one point do not subdivide and continue
                lit++;
                continue;
            }
            else
            {
                // If more than one point, subdivide
                ExtractorNode n1,n2,n3,n4;
                lit->DivideNode(n1,n2,n3,n4); // 再细分成四个子区域
 
                // Add childs if they contain points
                if(n1.vKeys.size()>0)
                {
                    lNodes.push_front(n1);                    
                    if(n1.vKeys.size()>1)
                    {
                        nToExpand++;
                        vSizeAndPointerToNode.push_back(make_pair(n1.vKeys.size(),&lNodes.front()));
                        lNodes.front().lit = lNodes.begin();
                    }
                }
                if(n2.vKeys.size()>0)
                {
                    lNodes.push_front(n2);
                    if(n2.vKeys.size()>1)
                    {
                        nToExpand++;
                        vSizeAndPointerToNode.push_back(make_pair(n2.vKeys.size(),&lNodes.front()));
                        lNodes.front().lit = lNodes.begin();
                    }
                }
                if(n3.vKeys.size()>0)
                {
                    lNodes.push_front(n3);
                    if(n3.vKeys.size()>1)
                    {
                        nToExpand++;
                        vSizeAndPointerToNode.push_back(make_pair(n3.vKeys.size(),&lNodes.front()));
                        lNodes.front().lit = lNodes.begin();
                    }
                }
                if(n4.vKeys.size()>0)
                {
                    lNodes.push_front(n4);
                    if(n4.vKeys.size()>1)
                    {
                        nToExpand++;
                        vSizeAndPointerToNode.push_back(make_pair(n4.vKeys.size(),&lNodes.front()));
                        lNodes.front().lit = lNodes.begin();
                    }
                }
 
                lit=lNodes.erase(lit);
                continue;
            }
        }       
 
        // Finish if there are more nodes than required features
        // or all nodes contain just one point
        if((int)lNodes.size()>=N || (int)lNodes.size()==prevSize)
        {
            bFinish = true;
        }
        // 当再划分之后所有的Node数大于要求数目时
        else if(((int)lNodes.size()+nToExpand*3)>N)
        {
 
            while(!bFinish)
            {
 
                prevSize = lNodes.size();
 
                vector<pair<int,ExtractorNode*> > vPrevSizeAndPointerToNode = vSizeAndPointerToNode;
                vSizeAndPointerToNode.clear();
 
                // 对需要划分的部分进行排序, 即对兴趣点数较多的区域进行划分
                sort(vPrevSizeAndPointerToNode.begin(),vPrevSizeAndPointerToNode.end());
                for(int j=vPrevSizeAndPointerToNode.size()-1;j>=0;j--)
                {
                    ExtractorNode n1,n2,n3,n4;
                    vPrevSizeAndPointerToNode[j].second->DivideNode(n1,n2,n3,n4);
 
                    // Add childs if they contain points
                    if(n1.vKeys.size()>0)
                    {
                        lNodes.push_front(n1);
                        if(n1.vKeys.size()>1)
                        {
                            vSizeAndPointerToNode.push_back(make_pair(n1.vKeys.size(),&lNodes.front()));
                            lNodes.front().lit = lNodes.begin();
                        }
                    }
                    if(n2.vKeys.size()>0)
                    {
                        lNodes.push_front(n2);
                        if(n2.vKeys.size()>1)
                        {
                            vSizeAndPointerToNode.push_back(make_pair(n2.vKeys.size(),&lNodes.front()));
                            lNodes.front().lit = lNodes.begin();
                        }
                    }
                    if(n3.vKeys.size()>0)
                    {
                        lNodes.push_front(n3);
                        if(n3.vKeys.size()>1)
                        {
                            vSizeAndPointerToNode.push_back(make_pair(n3.vKeys.size(),&lNodes.front()));
                            lNodes.front().lit = lNodes.begin();
                        }
                    }
                    if(n4.vKeys.size()>0)
                    {
                        lNodes.push_front(n4);
                        if(n4.vKeys.size()>1)
                        {
                            vSizeAndPointerToNode.push_back(make_pair(n4.vKeys.size(),&lNodes.front()));
                            lNodes.front().lit = lNodes.begin();
                        }
                    }
 
                    lNodes.erase(vPrevSizeAndPointerToNode[j].second->lit);
 
                    if((int)lNodes.size()>=N)
                        break;
                }
 
                if((int)lNodes.size()>=N || (int)lNodes.size()==prevSize)
                    bFinish = true;
 
            }
        }
    }
 
    // Retain the best point in each node
    // 保留每个区域响应值最大的一个兴趣点
    vector<cv::KeyPoint> vResultKeys;
    vResultKeys.reserve(nfeatures);
    for(list<ExtractorNode>::iterator lit=lNodes.begin(); lit!=lNodes.end(); lit++)
    {
        vector<cv::KeyPoint> &vNodeKeys = lit->vKeys;
        cv::KeyPoint* pKP = &vNodeKeys[0];
        float maxResponse = pKP->response;
 
        for(size_t k=1;k<vNodeKeys.size();k++)
        {
            if(vNodeKeys[k].response>maxResponse)
            {
                pKP = &vNodeKeys[k];
                maxResponse = vNodeKeys[k].response;
            }
        }
 
        vResultKeys.push_back(*pKP);
    }
 
    return vResultKeys;
}
 
 
 
//对影像金字塔中的每一层图像进行特征点的计算。具体计算过程是将影像网格分割成小区域,每一个小区域独立使用FAST角点检测
//检测完成之后使用DistributeOcTree函数对检测到所有的角点进行筛选,使得角点分布均匀
void ORBextractor::ComputeKeyPointsOctTree(vector<vector<KeyPoint> >& allKeypoints)
{
    allKeypoints.resize(nlevels);
 
    const float W = 30;//窗口大小
 
    // 对每一层图像做处理
    for (int level = 0; level < nlevels; ++level)   //计算边界
    {
        const int minBorderX = EDGE_THRESHOLD-3; //裁边19-3=16,
        const int minBorderY = minBorderX;
        const int maxBorderX = mvImagePyramid[level].cols-EDGE_THRESHOLD+3;
        const int maxBorderY = mvImagePyramid[level].rows-EDGE_THRESHOLD+3;
 
        vector<cv::KeyPoint> vToDistributeKeys;
        vToDistributeKeys.reserve(nfeatures*10);
 
//宽度和高度
        const float width = (maxBorderX-minBorderX);
        const float height = (maxBorderY-minBorderY);
 
//宽高方向的网格数量
        const int nCols = width/W;
        const int nRows = height/W;
 
//网格的宽与高
		const int wCell = ceil(width/nCols);
        const int hCell = ceil(height/nRows);
 
 
//遍历每一个窗口
		for(int i=0; i<nRows; i++)
        {
            const float iniY =minBorderY+i*hCell;
            float maxY = iniY+hCell+6;//最后计算的 Cell 的宽或高不会小于7。因为 FAST 计算的邻域是直径为7的 BressenHam 圆。
                                       //数字7与代码中出现的数字6对应。
 
//超出区域,进行下一个循环
            if(iniY>=maxBorderY-3)
                continue;
 
//最大Y超出边界就使用计算最宽的边界
			if(maxY>maxBorderY)
                maxY = maxBorderY;
 
//计算每列的位置
            for(int j=0; j<nCols; j++)
            {
                const float iniX =minBorderX+j*wCell;
                float maxX = iniX+wCell+6;
                if(iniX>=maxBorderX-6)
                    continue;
                if(maxX>maxBorderX)
                    maxX = maxBorderX;
 
                // FAST提取兴趣点, 自适应阈值
                vector<cv::KeyPoint> vKeysCell;
                FAST(mvImagePyramid[level].rowRange(iniY,maxY).colRange(iniX,maxX),
                     vKeysCell,iniThFAST,true);
 
//如果没有找到FAST特征点,就降低阈值重新计算
                if(vKeysCell.empty())
                {
                    FAST(mvImagePyramid[level].rowRange(iniY,maxY).colRange(iniX,maxX),
                         vKeysCell,minThFAST,true);
                }
 
//找到特征点,就将其放到vToDistributeKeys
				if(!vKeysCell.empty())
                {
                    for(vector<cv::KeyPoint>::iterator vit=vKeysCell.begin(); vit!=vKeysCell.end();vit++)
                    {
                        (*vit).pt.x+=j*wCell;
                        (*vit).pt.y+=i*hCell;
                        vToDistributeKeys.push_back(*vit);
                    }
                }
 
            }
        }
 
        vector<KeyPoint> & keypoints = allKeypoints[level];
        keypoints.reserve(nfeatures);
 
        // 根据mnFeaturesPerLevel,即该层的兴趣点数,对特征点进行剔除,采用Harris角点的score进行排序
        keypoints = DistributeOctTree(vToDistributeKeys, minBorderX, maxBorderX,
                                      minBorderY, maxBorderY,mnFeaturesPerLevel[level], level);
 
        const int scaledPatchSize = PATCH_SIZE*mvScaleFactor[level];
 
        // Add border to coordinates and scale information
        const int nkps = keypoints.size();
        for(int i=0; i<nkps ; i++)
        {
            keypoints[i].pt.x+=minBorderX;
            keypoints[i].pt.y+=minBorderY;
            keypoints[i].octave=level;
            keypoints[i].size = scaledPatchSize;
        }
    }
 
    // compute orientations
    for (int level = 0; level < nlevels; ++level)
        computeOrientation(mvImagePyramid[level], allKeypoints[level], umax);
}
 
 
//计算关键点
void ORBextractor::ComputeKeyPointsOld(std::vector<std::vector<KeyPoint> > &allKeypoints)
{
    allKeypoints.resize(nlevels);
 
    float imageRatio = (float)mvImagePyramid[0].cols/mvImagePyramid[0].rows;//图像纵横比
 
    for (int level = 0; level < nlevels; ++level)
    {
        const int nDesiredFeatures = mnFeaturesPerLevel[level];
 
        const int levelCols = sqrt((float)nDesiredFeatures/(5*imageRatio)); //论文中提到的每个网格5个点吗?
        const int levelRows = imageRatio*levelCols;
 
//得到每一层图像进行特征检测区域的上下两个坐标
        const int minBorderX = EDGE_THRESHOLD;
        const int minBorderY = minBorderX;
        const int maxBorderX = mvImagePyramid[level].cols-EDGE_THRESHOLD;
        const int maxBorderY = mvImagePyramid[level].rows-EDGE_THRESHOLD;
 
//将待检测区域划分为格子的行列个数
        const int W = maxBorderX - minBorderX;
        const int H = maxBorderY - minBorderY;
        const int cellW = ceil((float)W/levelCols);
        const int cellH = ceil((float)H/levelRows);
 
        const int nCells = levelRows*levelCols;
        const int nfeaturesCell = ceil((float)nDesiredFeatures/nCells);//每一个cell中特征点的个数
 
 
//Vector<T> v(n,i),向量V中含有n个值为i 的元素       			means cellKeypoint has levelRows层,每一层中又有levelCols层,均初始化为0
        vector<vector<vector<KeyPoint> > > cellKeyPoints(levelRows, vector<vector<KeyPoint> >(levelCols));
 
        vector<vector<int> > nToRetain(levelRows,vector<int>(levelCols,0));
        vector<vector<int> > nTotal(levelRows,vector<int>(levelCols,0));
        vector<vector<bool> > bNoMore(levelRows,vector<bool>(levelCols,false));
        vector<int> iniXCol(levelCols);
        vector<int> iniYRow(levelRows);
        int nNoMore = 0;
        int nToDistribute = 0;
 
 
        float hY = cellH + 6;
 
        for(int i=0; i<levelRows; i++)
        {
            const float iniY = minBorderY + i*cellH - 3;//第i个cell的第一个Y
            iniYRow[i] = iniY;// vector<int> iniYRow(levelRows)
 
            if(i == levelRows-1)//如果循环到最后一个
            {
                hY = maxBorderY+3-iniY;//hY=3+Ymax-iniY=3+Ymax-(Ymin+(levelRows-1)*cellH -3)=6+Ymax-Ymin-H+cellH=cellH+6  
                if(hY<=0)         //hY牵扯到后面cellimage的大小 范围从iniY到 iniY+hY,不可能为负值
                    continue;     //continue 只管for、while,不看if,不管多少if都直接无视;如果小于直接跳出本次循环,根据上一个注释的式子,正常是不会小于的 
            }
 
            float hX = cellW + 6;
 
            for(int j=0; j<levelCols; j++)
            {
                float iniX;
 
                if(i==0)
                {
                    iniX = minBorderX + j*cellW - 3; //和上面计算的y其实是关于某条线对称的
                    iniXCol[j] = iniX;
                }
                else
                {
                    iniX = iniXCol[j];// 和第一行的x值对齐 
                }
 
 
                if(j == levelCols-1)
                {
                    hX = maxBorderX+3-iniX;
                    if(hX<=0)
                        continue;
                }
 
 
                Mat cellImage = mvImagePyramid[level].rowRange(iniY,iniY+hY).colRange(iniX,iniX+hX);
 
                cellKeyPoints[i][j].reserve(nfeaturesCell*5);//论文中至少5个点
 
                FAST(cellImage,cellKeyPoints[i][j],iniThFAST,true);//FAST检测关键子
 
                if(cellKeyPoints[i][j].size()<=3)
                {
                    cellKeyPoints[i][j].clear();
 
                    FAST(cellImage,cellKeyPoints[i][j],minThFAST,true);//降低阈值重新检测关键子
                }
 
 
                const int nKeys = cellKeyPoints[i][j].size();//网格中到底有多少关键点
                nTotal[i][j] = nKeys;
 
                if(nKeys>nfeaturesCell)    //网格中的关键点比需要的要多
                {
                    nToRetain[i][j] = nfeaturesCell;    //保存预先计算好的关键点
                    bNoMore[i][j] = false;
                }
                else
                {
                    nToRetain[i][j] = nKeys;
                    nToDistribute += nfeaturesCell-nKeys;
                    bNoMore[i][j] = true;
                    nNoMore++;
                }
 
            }
        }
 
 
        // Retain by score  如果 总共的离散点数大于0并且 未达到阈值的cell数目比总共的格网数小;直到不需要离散 不需要加点为止  
 
        while(nToDistribute>0 && nNoMore<nCells)
        {
            int nNewFeaturesCell = nfeaturesCell + ceil((float)nToDistribute/(nCells-nNoMore)); //不够的cell需要加入后 新的点的数目(旧的加上均分的新的)
            nToDistribute = 0;
 
            for(int i=0; i<levelRows; i++)
            {
                for(int j=0; j<levelCols; j++)
                {
                    if(!bNoMore[i][j]) //有足够点数的cell
                    {
                        if(nTotal[i][j]>nNewFeaturesCell)	//总数目甚至比新的要求的点数还要多(当所有cell都执行这个条件语句,while循环就可以终止了) 
                        {
                            nToRetain[i][j] = nNewFeaturesCell;//只保存新要求的点的数目即可  
                            bNoMore[i][j] = false;
                        }
                        else
                        {
                            nToRetain[i][j] = nTotal[i][j];
                            nToDistribute += nNewFeaturesCell-nTotal[i][j];//还要离散的点的数目
                            bNoMore[i][j] = true;   //还需要在加点
                            nNoMore++;
                        }
                    }
                }
            }
        }
 
        vector<KeyPoint> & keypoints = allKeypoints[level];
        keypoints.reserve(nDesiredFeatures*2);
 
        const int scaledPatchSize = PATCH_SIZE*mvScaleFactor[level];
 
        // Retain by score and transform coordinates	换算特征点真实位置(添加边界值),添加特征点的尺度信息
        for(int i=0; i<levelRows; i++)
        {
            for(int j=0; j<levelCols; j++)
            {
                vector<KeyPoint> &keysCell = cellKeyPoints[i][j];
                KeyPointsFilter::retainBest(keysCell,nToRetain[i][j]);//保存最佳点
                if((int)keysCell.size()>nToRetain[i][j])
                    keysCell.resize(nToRetain[i][j]);
 
 
                for(size_t k=0, kend=keysCell.size(); k<kend; k++)
                {
                    keysCell[k].pt.x+=iniXCol[j];
                    keysCell[k].pt.y+=iniYRow[i];
                    keysCell[k].octave=level;
                    keysCell[k].size = scaledPatchSize;
                    keypoints.push_back(keysCell[k]);
                }
            }
        }
 
//特征点还多的话,在进行一次滤波
        if((int)keypoints.size()>nDesiredFeatures)		
        {
            KeyPointsFilter::retainBest(keypoints,nDesiredFeatures);
            keypoints.resize(nDesiredFeatures);
        }
    }
 
    // and compute orientations
    for (int level = 0; level < nlevels; ++level)
        computeOrientation(mvImagePyramid[level], allKeypoints[level], umax);
}
 
 
//计算描述子
static void computeDescriptors(const Mat& image, vector<KeyPoint>& keypoints, Mat& descriptors,
                               const vector<Point>& pattern)
{
    descriptors = Mat::zeros((int)keypoints.size(), 32, CV_8UC1);
 
    for (size_t i = 0; i < keypoints.size(); i++)
        computeOrbDescriptor(keypoints[i], image, &pattern[0], descriptors.ptr((int)i));
}
 
 
//输入的变量
// _image:获取的灰度图像
// _mask:掩码
// _keypoints:关键点
// _descriptors:描述子
//括号运算符输入图像,并且传入引用参数_keypoints,_descriptors用于计算得到的特征点及其描述子
// 这种设计使得只需要构造一次ORBextractor就可以为为所有图像生成特征点
void ORBextractor::operator()( InputArray _image, InputArray _mask, vector<KeyPoint>& _keypoints,
                      OutputArray _descriptors)
{ 
    if(_image.empty())
        return;
 
    Mat image = _image.getMat();
    assert(image.type() == CV_8UC1 );   //若错误则终止程序
 
    // Pre-compute the scale pyramid
    // 构建图像金字塔
    ComputePyramid(image);
 
    // 计算每层图像的兴趣点
    vector < vector<KeyPoint> > allKeypoints; // vector<vector<KeyPoint>>
    ComputeKeyPointsOctTree(allKeypoints);
    //ComputeKeyPointsOld(allKeypoints);
 
    Mat descriptors;
 
    int nkeypoints = 0;
    for (int level = 0; level < nlevels; ++level)
        nkeypoints += (int)allKeypoints[level].size();
    if( nkeypoints == 0 )
        _descriptors.release();
    else
    {
        _descriptors.create(nkeypoints, 32, CV_8U);
        descriptors = _descriptors.getMat();
    }
 
    _keypoints.clear();
    _keypoints.reserve(nkeypoints);
 
    int offset = 0;
    for (int level = 0; level < nlevels; ++level)
    {
        vector<KeyPoint>& keypoints = allKeypoints[level];
        int nkeypointsLevel = (int)keypoints.size();
 
        if(nkeypointsLevel==0)
            continue;
 
        // preprocess the resized image 对图像进行高斯模糊
        Mat workingMat = mvImagePyramid[level].clone();
        GaussianBlur(workingMat, workingMat, Size(7, 7), 2, 2, BORDER_REFLECT_101);
 
        // Compute the descriptors 计算描述子,采用高斯分布取点,就是上面的patten
        Mat desc = descriptors.rowRange(offset, offset + nkeypointsLevel);
        computeDescriptors(workingMat, keypoints, desc, pattern);
 
        offset += nkeypointsLevel;
 
        // Scale keypoint coordinates对关键点的位置坐做尺度恢复,恢复到原图的位置
        if (level != 0)
        {
            float scale = mvScaleFactor[level]; //getScale(level, firstLevel, scaleFactor);
            for (vector<KeyPoint>::iterator keypoint = keypoints.begin(),
                 keypointEnd = keypoints.end(); keypoint != keypointEnd; ++keypoint)
                keypoint->pt *= scale;
        }
 
//在_keypoints.end()前面插入区间keypoints.begin(), keypoints.end()的所有元素
		// And add the keypoints to the output
        _keypoints.insert(_keypoints.end(), keypoints.begin(), keypoints.end());
    }
}
 
/**
 * 构建图像金字塔
 * @param image 输入图像
 */
void ORBextractor::ComputePyramid(cv::Mat image)
{
    for (int level = 0; level < nlevels; ++level)
    {
        float scale = mvInvScaleFactor[level];
        Size sz(cvRound((float)image.cols*scale), cvRound((float)image.rows*scale));
        Size wholeSize(sz.width + EDGE_THRESHOLD*2, sz.height + EDGE_THRESHOLD*2);
        Mat temp(wholeSize, image.type()), masktemp;
        mvImagePyramid[level] = temp(Rect(EDGE_THRESHOLD, EDGE_THRESHOLD, sz.width, sz.height));
 
        // Compute the resized image
        if( level != 0 )
        {
            resize(mvImagePyramid[level-1], mvImagePyramid[level], sz, 0, 0, cv::INTER_LINEAR);
 
            copyMakeBorder(mvImagePyramid[level], temp, EDGE_THRESHOLD, EDGE_THRESHOLD, EDGE_THRESHOLD, EDGE_THRESHOLD,
                           BORDER_REFLECT_101+BORDER_ISOLATED);//这里的BORDER_ISOLATED是判断是否超出有效区,参考文章9            
        }
        else
        {
            copyMakeBorder(image, temp, EDGE_THRESHOLD, EDGE_THRESHOLD, EDGE_THRESHOLD, EDGE_THRESHOLD,
                           BORDER_REFLECT_101);            
        }
    }
 
}
 
} //namespace ORB_SLAM

四、与opencv中ORB对比测试

    这里先介绍opencv中ORB的使用,构造函数如下:

 ORB(int nfeatures = 500, float scaleFactor = 1.2f, int nlevels = 8, int edgeThreshold = 31,int firstLevel = 0, int WTA_K=2, int scoreType=ORB::HARRIS_SCORE, int patchSize=31 );

可见opencv中是可以选角点评分的:scoreType=ORB::HARRIS_SCORE

下面给出简单的测试代码:

#include <iostream>
#include<opencv2/core/core.hpp>
#include<opencv2/imgproc/imgproc.hpp>
#include<opencv2/highgui/highgui.hpp>

  #include <opencv2/opencv.hpp>

 
#include "ORBextractor.h"
using namespace ORB_SLAM2;
using namespace std;
using namespace cv;

int main(int argc, char **argv) 
{
     VideoCapture cap; 
     cap.open(0); 
     if(!cap.isOpened())//如果视频不能正常打开则返回
         return -1;
     
     ORBextractor *pORBextractor= new ORBextractor(100,1.2,8,20,7);
    
     Mat frame,im;
     while(1)
     {
         cap>>frame;//等价于cap.read(frame);
         if(frame.empty())//如果某帧为空则退出循环
             break;
	 imshow("video", frame);
	 cvtColor(frame,im,CV_RGB2GRAY);
	 	 
//orb-salm orb/ 
         std::vector<cv::KeyPoint> vKeys;
         cv::Mat Descriptors;
         (*pORBextractor)(im,cv::Mat(),vKeys,Descriptors);
	 
         Mat outimg1;
         drawKeypoints( frame, vKeys, outimg1,Scalar::all(-1),DrawMatchesFlags::DRAW_RICH_KEYPOINTS );
	 //drawKeypoints( frame, vKeys, outimg1,Scalar::all(-1),DrawMatchesFlags::DEFAULT );
         imshow("ORB特征点",outimg1);
    
//opencv orb/
	 // ORB特征点检测
         ORB orb(100); 
         orb.detect(frame, vKeys); 
 
	// 绘制关键点
	Mat keypoint_img;
	drawKeypoints(frame,vKeys, keypoint_img, Scalar::all(-1), DrawMatchesFlags::DRAW_RICH_KEYPOINTS);
	//drawKeypoints(frame,vKeys, keypoint_img, Scalar::all(-1), DrawMatchesFlags::DEFAULT);
	imshow("opencv KeyPoints Image", keypoint_img);
	 
       if( 'q'==waitKey(20)){break;};//每帧延时20毫秒
    }
    
    cap.release();//释放资源
      
   return 0;
}

测试结果如下:


五、总结

1.orb-slam中的特征点更丰富,因为两次高低阈值,使得更灵活,同时低阈值带来了一些不稳定的点,动态视频流中有频繁跳动的点。

2.opencv中特征点更为密集,这是orb-slam中4叉树的功劳。


六、参考

  1. ORBSLAM2学习(一):ORB算法原理:https://blog.csdn.net/lwx309025167/article/details/80365075
  2. BRIEF特征点描述子:https://www.cnblogs.com/wyuzl/p/7838103.html
  3. 【特征检测】FAST特征点检测算法:https://blog.csdn.net/hujingshuang/article/details/46898007
  4. 【特征检测】Harris角点检测中的数学推导:https://blog.csdn.net/hujingshuang/article/details/46829627
  5. ORBSLAM2学习(二):ORB源码分析:https://blog.csdn.net/lwx309025167/article/details/80421077
  6. ORBSLAM2 特征点提取代码注释:https://blog.csdn.net/u014797790/article/details/79776339
  7. orb:https://www.cnblogs.com/wall-e2/p/8057448.html
  8. ORB SLAM2学习笔记之mono_kitti(三):https://blog.csdn.net/weixin_43527185/article/details/84029903
  9. OpenCV的BorderTypes:https://cloud.tencent.com/developer/article/1350371

  • 3
    点赞
  • 38
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值