ORBSLAM2 特征点提取代码注释
2018-04-01 11:00
2006 查看
#include <opencv2/core/core.hpp> #include <opencv2/highgui/highgui.hpp> #include <opencv2/features2d/features2d.hpp> #include <opencv2/imgproc/imgproc.hpp> #include <vector> #include <iterator> #include "ORBextractor.h" #include <iostream> using namespace cv; using namespace std; namespace ORB_SLAM2 { const int PATCH_SIZE = 31; const int HALF_PATCH_SIZE = 15; const int EDGE_THRESHOLD = 19; //边界阈值 //灰度质心法(IC)计算特征的旋转 static float IC_Angle(const Mat& image, Point2f pt, const vector<int> & u_max) { int m_01 = 0, m_10 = 0; const uchar* center = &image.at<uchar> (cvRound(pt.y), cvRound(pt.x)); //cvRound 返回跟参数最接近的整数值; //我们要在一个圆域中算出m10和m01,计算步骤是先算出中间红线的m10,然后在平行于x轴算出m10和m01,一次计算相当于图像中的同个颜色的两个line。 // Treat the center line differently, v=0 横坐标:-15-----+15 for (int u = -HALF_PATCH_SIZE; u <= HALF_PATCH_SIZE; ++u) m_10 += u * center[u]; // Go line by line in the circuI853lar patch int step = (int)image.step1(); //opencv中概念,计算每行的元素个数 for (int v = 1; v <= HALF_PATCH_SIZE; ++v) { // Proceed over the two lines int v_sum = 0; int d = u_max[v]; for (int u = -d; u <= d; ++u) { int val_plus = center[u + v*step], val_minus = center[u - v*step]; v_sum += (val_plus - val_minus); m_10 += u * (val_plus + val_minus); } m_01 += v * v_sum; } //返回计算的角度 return fastAtan2((float)m_01, (float)m_10); } //弧度制与角度的转换 const float factorPI = (float)(CV_PI/180.f); //计算描述子 static void computeOrbDescriptor(const KeyPoint& kpt, const Mat& img, const Point* pattern, uchar* desc) { float angle = (float)kpt.angle*factorPI; float a = (float)cos(angle), b = (float)sin(angle); const uchar* center = &img.at<uchar>(cvRound(kpt.pt.y), cvRound(kpt.pt.x)); const int step = (int)img.step; #define GET_VALUE(idx) \ center[cvRound(pattern[idx].x*b + pattern[idx].y*a)*step + \ cvRound(pattern[idx].x*a - pattern[idx].y*b)] for (int i = 0; i < 32; ++i, pattern += 16) { 。。。 } #undef GET_VALUE } static int bit_pattern_31_[256*4] = { 。。。。 }; ORBextractor::ORBextractor(int _nfeatures, float _scaleFactor, int _nlevels, int _iniThFAST, int _minThFAST): nfeatures(_nfeatures), scaleFactor(_scaleFactor), nlevels(_nlevels), iniThFAST(_iniThFAST), minThFAST(_minThFAST) { /***********确定每一层的特征点数,采用等比数列**************/ //定义每一层的尺度和逆尺度 mvScaleFactor.resize(nlevels); mvLevelSigma2.resize(nlevels); mvScaleFactor[0]=1.0f; mvLevelSigma2[0]=1.0f; for(int i=1; i<nlevels; i++) { mvScaleFactor[i]=mvScaleFactor[i-1]*scaleFactor; mvLevelSigma2[i]=mvScaleFactor[i]*mvScaleFactor[i]; } mvInvScaleFactor.resize(nlevels); mvInvLevelSigma2.resize(nlevels); for(int i=0; i<nlevels; i++) { mvInvScaleFactor[i]=1.0f/mvScaleFactor[i]; mvInvLevelSigma2[i]=1.0f/mvLevelSigma2[i]; } mvImagePyramid.resize(nlevels); mnFeaturesPerLevel.resize(nlevels); float factor = 1.0f / scaleFactor; //第一层特征点数,以后每一层成等比数列 float nDesiredFeaturesPerScale = nfeatures*(1 - factor)/(1 - (float)pow((double)factor, (double)nlevels)); //所有层数的特征点数量加起来是nfeatures int sumFeatures = 0; for( int level = 0; level < nlevels-1; level++ ) { mnFeaturesPerLevel[level] = cvRound(nDesiredFeaturesPerScale); //取整 sumFeatures += mnFeaturesPerLevel[level]; nDesiredFeaturesPerScale *= factor; } mnFeaturesPerLevel[nlevels-1] = std::max(nfeatures - sumFeatures, 0); //复制训练的模板 const int npoints = 512; const Point* pattern0 = (const Point*)bit_pattern_31_; std::copy(pattern0, pattern0 + npoints, std::back_inserter(pattern)); //This is for orientation // pre-compute the end of a row in a circular patch //计算方向时,每个v对应的最大的u坐标 umax.resize(HALF_PATCH_SIZE + 1); // 将v坐标划分为两部分进行计算,主要为了确保计算特征主方向的时候,x,y方向对称 int v, v0, vmax = cvFloor(HALF_PATCH_SIZE * sqrt(2.f) / 2 + 1);//cvFloor含义是取不大于参数的最大整数值 int vmin = cvCeil(HALF_PATCH_SIZE * sqrt(2.f) / 2); //cvCeil含义是取不小于参数的最小整数值 //利用勾股定理计算坐标 const double hp2 = HALF_PATCH_SIZE*HALF_PATCH_SIZE; //patch圆半径的平方 for (v = 0; v <= vmax; ++v) umax[v] = cvRound(sqrt(hp2 - v * v)); //每一个v坐标,最大的U坐标 // Make sure we are symmetric 确保是圆 for (v = HALF_PATCH_SIZE, v0 = 0; v >= vmin; --v) { while (umax[v0] == umax[v0 + 1]) ++v0; umax[v] = v0; ++v0; } } //计算每个关键点的角度 static void computeOrientation(const Mat& image, vector<KeyPoint>& keypoints, const vector<int>& umax) { for (vector<KeyPoint>::iterator keypoint = keypoints.begin(), keypointEnd = keypoints.end(); keypoint != keypointEnd; ++keypoint) { keypoint->angle = IC_Angle(image, keypoint->pt, umax); } } void ExtractorNode::DivideNode(ExtractorNode &n1, ExtractorNode &n2, ExtractorNode &n3, ExtractorNode &n4) { const int halfX = ceil(static_cast<float>(UR.x-UL.x)/2); const int halfY = ceil(static_cast<float>(BR.y-UL.y)/2); //Define boundaries of childs n1.UL = UL; n1.UR = cv::Point2i(UL.x+halfX,UL.y); n1.BL = cv::Point2i(UL.x,UL.y+halfY); n1.BR = cv::Point2i(UL.x+halfX,UL.y+halfY); n1.vKeys.reserve(vKeys.size()); n2.UL = n1.UR; n2.UR = UR; n2.BL = n1.BR; n2.BR = cv::Point2i(UR.x,UL.y+halfY); n2.vKeys.reserve(vKeys.size()); n3.UL = n1.BL; n3.UR = n1.BR; n3.BL = BL; n3.BR = cv::Point2i(n1.BR.x,BL.y); n3.vKeys.reserve(vKeys.size()); n4.UL = n3.UR; n4.UR = n2.BR; n4.BL = n3.BR; n4.BR = BR; n4.vKeys.reserve(vKeys.size()); //Associate points to childs for(size_t i=0;i<vKeys.size();i++) { const cv::KeyPoint &kp = vKeys[i]; if(kp.pt.x<n1.UR.x) { if(kp.pt.y<n1.BR.y) n1.vKeys.push_back(kp); else n3.vKeys.push_back(kp); } else if(kp.pt.y<n1.BR.y) n2.vKeys.push_back(kp); else n4.vKeys.push_back(kp); } if(n1.vKeys.size()==1) n1.bNoMore = true; if(n2.vKeys.size()==1) n2.bNoMore = true; if(n3.vKeys.size()==1) n3.bNoMore = true; if(n4.vKeys.size()==1) n4.bNoMore = true; } //计算fast特征点并进行筛选 //先用(maxX-minX)/(maxY-minY)来确定四叉数有几个初始节点,这里有 bug,如果输入的是一张宽高比小于0.5 的图像, //nIni 计算得到 0,下一步计算 hX 会报错,例如round(640/480)=1,所以只有一个初始节点,(UL,UR,BL,BR)就会分布到被裁掉边界后的图像的四个角。 //把所有的关键点分配给属于它的节点,当节点所分配到的关键点的个数为1时就不再进行分裂,当节点没有分配到关键点时就删除此节点。 //再根据兴趣点分布,利用四叉树方法对图像进行划分区域,当bFinish的值为true时就不再进行区域划分,首先对目前的区域进行划分, //把每次划分得到的有关键点的子区域设为新的节点,将nToExpand参数加一,并插入到节点列表的前边,删除掉其父节点。 //只要新节点中的关键点的个数超过一个,就继续划分,继续插入列表前面,继续删除父节点,直到划分的子区域中的关键点的个数是一个, //然后迭代器加以移动到下一个节点,继续划分区域。 //当划分的区域即节点的个数大于关键点的个数或者分裂过程没有增加节点的个数时就将bFinish的值设为true 4000 ,不再进行划分。 //如果以上条件没有满足,但是满足((int)lNodes.size()+nToExpand*3)>N,则说明将所有节点再分裂一次可以达到要求。 //vSizeAndPointerToNode 是前面分裂出来的子节点(n1, n2, n3, n4)中可以分裂的节点。 //按照它们特征点的排序,先从特征点多的开始分裂,分裂的结果继续存储在 lNodes 中。每分裂一个节点都会进行一次判断, //如果 lNodes 中的节点数量大于所需要的特征点数量,退出整个 while(!bFinish) 循环,如果进行了一次分裂, //并没有增加节点数量,不玩了,退出整个 while(!bFinish) 循环。取出每一个节点(每个区域)对应的最大响应点,即我们确定的特征点。 vector<cv::KeyPoint> ORBextractor::DistributeOctTree(const vector<cv::KeyPoint>& vToDistributeKeys, const int &minX, const int &maxX, const int &minY, const int &maxY, const int &N, const int &level) { // Compute how many initial nodes 计算有多少初始节点 const int nIni = round(static_cast<float>(maxX-minX)/(maxY-minY)); //round四舍五入取整,水平划分格子的数量 const float hX = static_cast<float>(maxX-minX)/nIni; //水平划分格子的宽度 list<ExtractorNode> lNodes; vector<ExtractorNode*> vpIniNodes; vpIniNodes.resize(nIni); for(int i=0; i<nIni; i++) { ExtractorNode ni; ni.UL = cv::Point2i(hX*static_cast<float>(i),0); ni.UR = cv::Point2i(hX*static_cast<float>(i+1),0); ni.BL = cv::Point2i(ni.UL.x,maxY-minY); ni.BR = cv::Point2i(ni.UR.x,maxY-minY); ni.vKeys.reserve(vToDistributeKeys.size()); lNodes.push_back(ni); vpIniNodes[i] = &lNodes.back(); } //Associate points to childs for(size_t i=0;i<vToDistributeKeys.size();i++) { const cv::KeyPoint &kp = vToDistributeKeys[i]; vpIniNodes[kp.pt.x/hX]->vKeys.push_back(kp); } list<ExtractorNode>::iterator lit = lNodes.begin(); while(lit!=lNodes.end()) { if(lit->vKeys.size()==1) { lit->bNoMore=true; lit++; } else if(lit->vKeys.empty()) lit = lNodes.erase(lit); else lit++; } bool bFinish = false; int iteration = 0; vector<pair<int,ExtractorNode*> > vSizeAndPointerToNode; vSizeAndPointerToNode.reserve(lNodes.size()*4); // 根据兴趣点分布,利用N叉树方法对图像进行划分区域 while(!bFinish) { iteration++; int prevSize = lNodes.size(); lit = lNodes.begin(); int nToExpand = 0; vSizeAndPointerToNode.clear(); // 将目前的子区域经行划分 while(lit!=lNodes.end()) { if(lit->bNoMore) { // If node only contains one point do not subdivide and continue lit++; continue; } else { // If more than one point, subdivide ExtractorNode n1,n2,n3,n4; lit->DivideNode(n1,n2,n3,n4); // 再细分成四个子区域 // Add childs if they contain points if(n1.vKeys.size()>0) { lNodes.push_front(n1); if(n1.vKeys.size()>1) { nToExpand++; vSizeAndPointerToNode.push_back(make_pair(n1.vKeys.size(),&lNodes.front())); lNodes.front().lit = lNodes.begin(); } } if(n2.vKeys.size()>0) { lNodes.push_front(n2); if(n2.vKeys.size()>1) { nToExpand++; vSizeAndPointerToNode.push_back(make_pair(n2.vKeys.size(),&lNodes.front())); lNodes.front().lit = lNodes.begin(); } } if(n3.vKeys.size()>0) { lNodes.push_front(n3); if(n3.vKeys.size()>1) { nToExpand++; vSizeAndPointerToNode.push_back(make_pair(n3.vKeys.size(),&lNodes.front())); lNodes.front().lit = lNodes.begin(); } } if(n4.vKeys.size()>0) { lNodes.push_front(n4); if(n4.vKeys.size()>1) { nToExpand++; vSizeAndPointerToNode.push_back(make_pair(n4.vKeys.size(),&lNodes.front())); lNodes.front().lit = lNodes.begin(); } } lit=lNodes.erase(lit); continue; } } // Finish if there are more nodes than required features // or all nodes contain just one point if((int)lNodes.size()>=N || (int)lNodes.size()==prevSize) { bFinish = true; } // 当再划分之后所有的Node数大于要求数目时 else if(((int)lNodes.size()+nToExpand*3)>N) { while(!bFinish) { prevSize = lNodes.size(); vector<pair<int,ExtractorNode*> > vPrevSizeAndPointerToNode = vSizeAndPointerToNode; vSizeAndPointerToNode.clear(); // 对需要划分的部分进行排序, 即对兴趣点数较多的区域进行划分 sort(vPrevSizeAndPointerToNode.begin(),vPrevSizeAndPointerToNode.end()); for(int j=vPrevSizeAndPointerToNode.size()-1;j>=0;j--) { ExtractorNode n1,n2,n3,n4; vPrevSizeAndPointerToNode[j].second->DivideNode(n1,n2,n3,n4); // Add childs if they contain points if(n1.vKeys.size()>0) { lNodes.push_front(n1); if(n1.vKeys.size()>1) { vSizeAndPointerToNode.push_back(make_pair(n1.vKeys.size(),&lNodes.front())); lNodes.front().lit = lNodes.begin(); } } if(n2.vKeys.size()>0) { lNodes.push_front(n2); if(n2.vKeys.size()>1) { vSizeAndPointerToNode.push_back(make_pair(n2.vKeys.size(),&lNodes.front())); lNodes.front().lit = lNodes.begin(); } } if(n3.vKeys.size()>0) { lNodes.push_front(n3); if(n3.vKeys.size()>1) { vSizeAndPointerToNode.push_back(make_pair(n3.vKeys.size(),&lNodes.front())); lNodes.front().lit = lNodes.begin(); } } if(n4.vKeys.size()>0) { lNodes.push_front(n4); if(n4.vKeys.size()>1) { vSizeAndPointerToNode.push_back(make_pair(n4.vKeys.size(),&lNodes.front())); lNodes.front().lit = lNodes.begin(); } } lNodes.erase(vPrevSizeAndPointerToNode[j].second->lit); if((int)lNodes.size()>=N) break; } if((int)lNodes.size()>=N || (int)lNodes.size()==prevSize) bFinish = true; } } } // Retain the best point in each node // 保留每个区域响应值最大的一个兴趣点 vector<cv::KeyPoint> vResultKeys; vResultKeys.reserve(nfeatures); for(list<ExtractorNode>::iterator lit=lNodes.begin(); lit!=lNodes.end(); lit++) { vector<cv::KeyPoint> &vNodeKeys = lit->vKeys; cv::KeyPoint* pKP = &vNodeKeys[0]; float maxResponse = pKP->response; for(size_t k=1;k<vNodeKeys.size();k++) { if(vNodeKeys[k].response>maxResponse) { pKP = &vNodeKeys[k]; maxResponse = vNodeKeys[k].response; } } vResultKeys.push_back(*pKP); } return vResultKeys; } //对影像金字塔中的每一层图像进行特征点的计算。具体计算过程是将影像网格分割成小区域,每一个小区域独立使用FAST角点检测 //检测完成之后使用DistributeOcTree函数对检测到所有的角点进行筛选,使得角点分布均匀 void ORBextractor::ComputeKeyPointsOctTree(vector<vector<KeyPoint> >& allKeypoints) { allKeypoints.resize(nlevels); const float W = 30;//窗口大小 // 对每一层图像做处理 for (int level = 0; level < nlevels; ++level) //计算边界 { const int minBorderX = EDGE_THRESHOLD-3; //裁边19-3=16, const int minBorderY = minBorderX; const int maxBorderX = mvImagePyramid[level].cols-EDGE_THRESHOLD+3; const int maxBorderY = mvImagePyramid[level].rows-EDGE_THRESHOLD+3; vector<cv::KeyPoint> vToDistributeKeys; vToDistributeKeys.reserve(nfeatures*10); //宽度和高度 const float width = (maxBorderX-minBorderX); const float height = (maxBorderY-minBorderY); //宽高方向的网格数量 const int nCols = width/W; const int nRows = height/W; //网格的宽与高 const int wCell = ceil(width/nCols); const int hCell = ceil(height/nRows); //遍历每一个窗口 for(int i=0; i<nRows; i++) { const float iniY =minBorderY+i*hCell; float maxY = iniY+hCell+6;//最后计算的 Cell 的宽或高不会小于7。因为 FAST 计算的邻域是直径为7的 BressenHam 圆。 //数字7与代码中出现的数字6对应。 //超出区域,进行下一个循环 if(iniY>=maxBorderY-3) continue; //最大Y超出边界就使用计算最宽的边界 if(maxY>maxBorderY) maxY = maxBorderY; //计算每列的位置 for(int j=0; j<nCols; j++) { const float iniX =minBorderX+j*wCell; float maxX = iniX+wCell+6; if(iniX>=maxBorderX-6) continue; if(maxX>maxBorderX) maxX = maxBorderX; // FAST提取兴趣点, 自适应阈值 vector<cv::KeyPoint> vKeysCell; FAST(mvImagePyramid[level].rowRange(iniY,maxY).colRange(iniX,maxX), vKeysCell,iniThFAST,true); //如果没有找到FAST特征点,就降低阈值重新计算 if(vKeysCell.empty()) { FAST(mvImagePyramid[level].rowRange(iniY,maxY).colRange(iniX,maxX), vKeysCell,minThFAST,true); } //找到特征点,就将其放到vToDistributeKeys if(!vKeysCell.empty()) { for(vector<cv::KeyPoint>::iterator vit=vKeysCell.begin(); vit!=vKeysCell.end();vit++) { (*vit).pt.x+=j*wCell; (*vit).pt.y+=i*hCell; vToDistributeKeys.push_back(*vit); } } } } vector<KeyPoint> & keypoints = allKeypoints[level]; keypoints.reserve(nfeatures); // 根据mnFeaturesPerLevel,即该层的兴趣点数,对特征点进行剔除,采用Harris角点的score进行排序 keypoints = DistributeOctTree(vToDistributeKeys, minBorderX, maxBorderX, minBorderY, maxBorderY,mnFeaturesPerLevel[level], level); const int scaledPatchSize = PATCH_SIZE*mvScaleFactor[level]; // Add border to coordinates and scale information const int nkps = keypoints.size(); for(int i=0; i<nkps ; i++) { keypoints[i].pt.x+=minBorderX; keypoints[i].pt.y+=minBorderY; keypoints[i].octave=level; keypoints[i].size = scaledPatchSize; } } // compute orientations for (int level = 0; level < nlevels; ++level) computeOrientation(mvImagePyramid[level], allKeypoints[level], umax); } //计算关键点 void ORBextractor::ComputeKeyPointsOld(std::vector<std::vector<KeyPoint> > &allKeypoints) { allKeypoints.resize(nlevels); float imageRatio = (float)mvImagePyramid[0].cols/mvImagePyramid[0].rows;//图像纵横比 for (int level = 0; level < nlevels; ++level) { const int nDesiredFeatures = mnFeaturesPerLevel[level]; const int levelCols = sqrt((float)nDesiredFeatures/(5*imageRatio)); //论文中提到的每个网格5个点吗? const int levelRows = imageRatio*levelCols; //得到每一层图像进行特征检测区域的上下两个坐标 const int minBorderX = EDGE_THRESHOLD; const int minBorderY = minBorderX; const int maxBorderX = mvImagePyramid[level].cols-EDGE_THRESHOLD; const int maxBorderY = mvImagePyramid[level].rows-EDGE_THRESHOLD; //将待检测区域划分为格子的行列个数 const int W = maxBorderX - minBorderX; const int H = maxBorderY - minBorderY; const int cellW = ceil((float)W/levelCols); const int cellH = ceil((float)H/levelRows); const int nCells = levelRows*levelCols; const int nfeaturesCell = ceil((float)nDesiredFeatures/nCells);//每一个cell中特征点的个数 //Vector<T> v(n,i),向量V中含有n个值为i 的元素 means cellKeypoint has levelRows层,每一层中又有levelCols层,均初始化为0 vector<vector<vector<KeyPoint> > > cellKeyPoints(levelRows, vector<vector<KeyPoint> >(levelCols)); vector<vector<int> > nToRetain(levelRows,vector<int>(levelCols,0)); vector<vector<int> > nTotal(levelRows,vector<int>(levelCols,0)); vector<vector<bool> > bNoMore(levelRows,vector<bool>(levelCols,false)); vector<int> iniXCol(levelCols); vector<int> iniYRow(levelRows); int nNoMore = 0; int nToDistribute = 0; float hY = cellH + 6; for(int i=0; i<levelRows; i++) { const float iniY = minBorderY + i*cellH - 3;//第i个cell的第一个Y iniYRow[i] = iniY;// vector<int> iniYRow(levelRows) if(i == levelRows-1)//如果循环到最后一个 { hY = maxBorderY+3-iniY;//hY=3+Ymax-iniY=3+Ymax-(Ymin+(levelRows-1)*cellH -3)=6+Ymax-Ymin-H+cellH=cellH+6 if(hY<=0) //hY牵扯到后面cellimage的大小 范围从iniY到 iniY+hY,不可能为负值 continue; //continue 只管for、while,不看if,不管多少if都直接无视;如果小于直接跳出本次循环,根据上一个注释的式子,正常是不会小于的 } float hX = cellW + 6; for(int j=0; j<levelCols; j++) { float iniX; if(i==0) { iniX = minBorderX + j*cellW - 3; //和上面计算的y其实是关于某条线对称的 iniXCol[j] = iniX; } else { iniX = iniXCol[j];// 和第一行的x值对齐 } if(j == levelCols-1) a4bf { hX = maxBorderX+3-iniX; if(hX<=0) continue; } Mat cellImage = mvImagePyramid[level].rowRange(iniY,iniY+hY).colRange(iniX,iniX+hX); cellKeyPoints[i][j].reserve(nfeaturesCell*5);//论文中至少5个点 FAST(cellImage,cellKeyPoints[i][j],iniThFAST,true);//FAST检测关键子 if(cellKeyPoints[i][j].size()<=3) { cellKeyPoints[i][j].clear(); FAST(cellImage,cellKeyPoints[i][j],minThFAST,true);//降低阈值重新检测关键子 } const int nKeys = cellKeyPoints[i][j].size();//网格中到底有多少关键点 nTotal[i][j] = nKeys; if(nKeys>nfeaturesCell) //网格中的关键点比需要的要多 { nToRetain[i][j] = nfeaturesCell; //保存预先计算好的关键点 bNoMore[i][j] = false; } else { nToRetain[i][j] = nKeys; nToDistribute += nfeaturesCell-nKeys; bNoMore[i][j] = true; nNoMore++; } } } // Retain by score 如果 总共的离散点数大于0并且 未达到阈值的cell数目比总共的格网数小;直到不需要离散 不需要加点为止 while(nToDistribute>0 && nNoMore<nCells) { int nNewFeaturesCell = nfeaturesCell + ceil((float)nToDistribute/(nCells-nNoMore)); //不够的cell需要加入后 新的点的数目(旧的加上均分的新的) nToDistribute = 0; for(int i=0; i<levelRows; i++) { for(int j=0; j<levelCols; j++) { if(!bNoMore[i][j]) //有足够点数的cell { if(nTotal[i][j]>nNewFeaturesCell) //总数目甚至比新的要求的点数还要多(当所有cell都执行这个条件语句,while循环就可以终止了) { nToRetain[i][j] = nNewFeaturesCell;//只保存新要求的点的数目即可 bNoMore[i][j] = false; } else { nToRetain[i][j] = nTotal[i][j]; nToDistribute += nNewFeaturesCell-nTotal[i][j];//还要离散的点的数目 bNoMore[i][j] = true; //还需要在加点 nNoMore++; } } } } } vector<KeyPoint> & keypoints = allKeypoints[level]; keypoints.reserve(nDesiredFeatures*2); const int scaledPatchSize = PATCH_SIZE*mvScaleFactor[level]; // Retain by score and transform coordinates 换算特征点真实位置(添加边界值),添加特征点的尺度信息 for(int i=0; i<levelRows; i++) { for(int j=0; j<levelCols; j++) { vector<KeyPoint> &keysCell = cellKeyPoints[i][j]; KeyPointsFilter::retainBest(keysCell,nToRetain[i][j]);//保存最佳点 if((int)keysCell.size()>nToRetain[i][j]) keysCell.resize(nToRetain[i][j]); for(size_t k=0, kend=keysCell.size(); k<kend; k++) { keysCell[k].pt.x+=iniXCol[j]; keysCell[k].pt.y+=iniYRow[i]; keysCell[k].octave=level; keysCell[k].size = scaledPatchSize; keypoints.push_back(keysCell[k]); } } } //特征点还多的话,在进行一次滤波 if((int)keypoints.size()>nDesiredFeatures) { KeyPointsFilter::retainBest(keypoints,nDesiredFeatures); keypoints.resize(nDesiredFeatures); } } // and compute orientations for (int level = 0; level < nlevels; ++level) computeOrientation(mvImagePyramid[level], allKeypoints[level], umax); } //计算描述子 static void computeDescriptors(const Mat& image, vector<KeyPoint>& keypoints, Mat& descriptors, const vector<Point>& pattern) { descriptors = Mat::zeros((int)keypoints.size(), 32, CV_8UC1); for (size_t i = 0; i < keypoints.size(); i++) computeOrbDescriptor(keypoints[i], image, &pattern[0], descriptors.ptr((int)i)); } //输入的变量 // _image:获取的灰度图像 // _mask:掩码 // _keypoints:关键点 // _descriptors:描述子 //括号运算符输入图像,并且传入引用参数_keypoints,_descriptors用于计算得到的特征点及其描述子 // 这种设计使得只需要构造一次ORBextractor就可以为为所有图像生成特征点 void ORBextractor::operator()( InputArray _image, InputArray _mask, vector<KeyPoint>& _keypoints, OutputArray _descriptors) { if(_image.empty()) return; Mat image = _image.getMat(); assert(image.type() == CV_8UC1 ); //若错误则终止程序 // Pre-compute the scale pyramid // 构建图像金字塔 ComputePyramid(image); // 计算每层图像的兴趣点 vector < vector<KeyPoint> > allKeypoints; // vector<vector<KeyPoint>> ComputeKeyPointsOctTree(allKeypoints); //ComputeKeyPointsOld(allKeypoints); Mat descriptors; int nkeypoints = 0; for (int level = 0; level < nlevels; ++level) nkeypoints += (int)allKeypoints[level].size(); if( nkeypoints == 0 ) _descriptors.release(); else { _descriptors.create(nkeypoints, 32, CV_8U); descriptors = _descriptors.getMat(); } _keypoints.clear(); _keypoints.reserve(nkeypoints); int offset = 0; for (int level = 0; level < nlevels; ++level) { vector<KeyPoint>& keypoints = allKeypoints[level]; int nkeypointsLevel = (int)keypoints.size(); if(nkeypointsLevel==0) continue; // preprocess the resized image 对图像进行高斯模糊 Mat workingMat = mvImagePyramid[level].clone(); GaussianBlur(workingMat, workingMat, Size(7, 7), 2, 2, BORDER_REFLECT_101); // Compute the descriptors 计算描述子,采用高斯分布取点,就是上面的patten Mat desc = descriptors.rowRange(offset, offset + nkeypointsLevel); computeDescriptors(workingMat, keypoints, desc, pattern); offset += nkeypointsLevel; // Scale keypoint coordinates对关键点的位置坐做尺度恢复,恢复到原图的位置 if (level != 0) { float scale = mvScaleFactor[level]; //getScale(level, firstLevel, scaleFactor); for (vector<KeyPoint>::iterator keypoint = keypoints.begin(), keypointEnd = keypoints.end(); keypoint != keypointEnd; ++keypoint) keypoint->pt *= scale; } //在_keypoints.end()前面插入区间keypoints.begin(), keypoints.end()的所有元素 // And add the keypoints to the output _keypoints.insert(_keypoints.end(), keypoints.begin(), keypoints.end()); } } /** * 构建图像金字塔 * @param image 输入图像 */ void ORBextractor::ComputePyramid(cv::Mat image) { for (int level = 0; level < nlevels; ++level) { float scale = mvInvScaleFactor[level]; Size sz(cvRound((float)image.cols*scale), cvRound((float)image.rows*scale)); Size wholeSize(sz.width + EDGE_THRESHOLD*2, sz.height + EDGE_THRESHOLD*2); Mat temp(wholeSize, image.type()), masktemp; mvImagePyramid[level] = temp(Rect(EDGE_THRESHOLD, EDGE_THRESHOLD, sz.width, sz.height)); // Compute the resized image if( level != 0 ) { resize(mvImagePyramid[level-1], mvImagePyramid[level], sz, 0, 0, cv::INTER_LINEAR); copyMakeBorder(mvImagePyramid[level], temp, EDGE_THRESHOLD, EDGE_THRESHOLD, EDGE_THRESHOLD, EDGE_THRESHOLD, BORDER_REFLECT_101+BORDER_ISOLATED); } else { copyMakeBorder(image, temp, EDGE_THRESHOLD, EDGE_THRESHOLD, EDGE_THRESHOLD, EDGE_THRESHOLD, BORDER_REFLECT_101); } } } } //namespace ORB_SLAM
相关文章推荐
- ORB-SLAM2跟踪之提取ORB特征
- ORB_SLAM 中特征提取代码介绍
- SLAM——ORB特征提取、对极约束(E\F)、单应矩阵和三角测量(一)
- 一起学ORBSLAM(2)ORB特征点提取
- ORB_SLAM ORBextractor 特征点提取 opencv源码学习
- ORB特征提取与匹配
- opencv 特征点提取算法 SIFT SURF ORB FAST LBP学习(二)
- 【AR实验室】mulberryAR:并行提取ORB特征
- opencv库图像特征提取与匹配--参考SLAM十四讲7.2
- ORB-SLAM2:基于可识别特征的自主导航与地图构建
- 一步步实现slam2-orb特征检测
- SLAM学习笔记4: ORB-SLAM中BOW特征匹配
- 【图像特征提取5】ORB特征提取-----FAST角点检测
- 图像特征点提取及描述子汇总:FIST、SURF、FAST、BRIEF、ORB、BRISK、FREAK
- 特征提取算法--ORB
- opencv上gpu版surf特征点与orb特征点提取及匹配实例
- OpenCV2学习笔记(十四):ORB特征提取
- SIFT,SURF,ORB,FAST 特征提取算法比较
- 基于ORB特征提取算法图像匹配 python实现
- ORB特征提取