CN112115953A - Optimized ORB algorithm based on RGB-D camera combined with plane detection and random sampling consistency algorithm - Google Patents

Optimized ORB algorithm based on RGB-D camera combined with plane detection and random sampling consistency algorithm Download PDF

Info

Publication number
CN112115953A
CN112115953A CN202010985540.5A CN202010985540A CN112115953A CN 112115953 A CN112115953 A CN 112115953A CN 202010985540 A CN202010985540 A CN 202010985540A CN 112115953 A CN112115953 A CN 112115953A
Authority
CN
China
Prior art keywords
point
points
feature
algorithm
image data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010985540.5A
Other languages
Chinese (zh)
Other versions
CN112115953B (en
Inventor
程明
司雨晨
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Tech University
Original Assignee
Nanjing Tech University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Tech University filed Critical Nanjing Tech University
Priority to CN202010985540.5A priority Critical patent/CN112115953B/en
Publication of CN112115953A publication Critical patent/CN112115953A/en
Application granted granted Critical
Publication of CN112115953B publication Critical patent/CN112115953B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides an optimized ORB algorithm based on an RGB-D camera combined with a plane detection and random sampling consistency algorithm, which comprises the following steps: s1: acquiring image data by using an RGB-D camera, wherein the image data comprises a color image and a depth image; s2: using an ORB algorithm to extract the feature points of the image data, and using a feature point uniformity evaluation method to judge the distribution uniformity of the feature points; s3: generating point clouds and performing down-sampling on the point clouds for the image data part with uniformly distributed characteristic points; s4: performing plane detection extraction on the point cloud subjected to down-sampling, and eliminating mismatching by using a random sampling consistency algorithm; s5: and for the image data part with non-uniform distribution of the feature points, using a set threshold value to extract the feature points and removing overlapped feature points by a non-maximum value inhibition method. The method can reduce the calculation amount, improve the accuracy of feature point extraction and reduce the mismatching, thereby realizing the requirements of the mobile robot on the accuracy and the real-time performance.

Description

Optimized ORB algorithm based on RGB-D camera combined with plane detection and random sampling consistency algorithm
Technical Field
The invention discloses an optimized ORB algorithm based on an RGB-D camera combined with a plane detection and random sampling consistency algorithm, and belongs to the technical field of indoor mobile robot path planning navigation.
Background
In recent years, smart mobile robot technology has been rapidly developed and has been widely used in the fields of industry, military, logistics, office, home service, and the like. With the advent of RGB-D sensors, research into mobile robot positioning or SLAM using RGB-D sensors has rapidly progressed. The improvement of image processing, point cloud processing and other related technologies, and the advantages of rich acquired information, non-contact measurement, easy installation and use, low cost and the like of the RGB-D sensor, so that the RGB-D sensor is widely applied to the fields of target recognition, tracking and the like. The first problem in the robot navigation problem is how to determine a scene model, and many solutions have relied on two-dimensional sensors such as lasers, radars, etc. for map construction and robot pose estimation over the past decade. With the advent of RGB-D cameras, more and more researchers have been concerned about solving the problem of robot indoor environment model building using RGB-D cameras, and have produced many influential research results.
At present, the visual SLAM technology gradually becomes a mainstream positioning scheme, however, the monocular SLAM cannot acquire depth information of a pixel point from one image, and the depth of the pixel point needs to be estimated by a triangulation or inverse depth method. And the depth information estimated by monocular SLAM has scale uncertainty, and along with the accumulation of positioning error, the phenomenon of 'scale drift' easily occurs. The binocular SLAM obtains matched feature points by matching images of the left camera and the right camera, and then estimates the depth information of the feature points according to a parallax method. The binocular SLAM has the advantages of large measurement range and the like, but the calculation amount is large, the precision requirement on the camera is high, and the real-time requirement can be met by accelerating the GPU. An RGB-D camera is a new type of camera that has emerged in recent years, and the camera can actively acquire depth information of pixel points in an image through physical hardware. Compared with a monocular camera and a binocular camera, the RGB-D camera does not need to consume a large amount of computing resources to compute the depth of the pixel point, can directly carry out three-dimensional measurement on the surrounding environment and the obstacle, generates a dense point cloud map through the RGB-D SLAM technology, and provides convenience for subsequent navigation planning.
The feature point extraction and matching method based on the RGB-D camera commonly used at present comprises an ORB algorithm and an ICP algorithm. ORB (organized FAST and rotaed BRIEF) is an algorithm for FAST feature point extraction and description. The ORB algorithm is divided into two parts, namely feature point extraction and feature point description. The feature extraction is developed by the fast (features from accessed Segment test) algorithm, and the feature point description is improved according to the brief (binary Robust Independent feature features) feature description algorithm. The ORB feature is to combine the detection method of FAST feature points with BRIEF feature descriptors and make improvements and optimization on the original basis. The ORB algorithm is characterized by fast calculation speed. This first benefits from using FAST to detect feature points. And again, the BRIEF algorithm is used for calculating the descriptor, and the expression form of the 2-system string specific to the descriptor not only saves the storage space, but also greatly shortens the matching time. The ICP algorithm is proposed by Besl and McKay 1992, Method for registration of 3-D maps. The basic principle of the ICP algorithm is: and respectively finding out the nearest points (pi, qi) in the target point cloud P and the source point cloud Q with matching according to a certain constraint condition, and then calculating optimal matching parameters R and t to ensure that an error function is minimum. This method can improve the efficiency of producing an optimal solution, but is less practical for a mobile robot due to a large amount of calculation, resulting in an increase in cost.
Disclosure of Invention
In view of this, the present invention provides an optimized ORB algorithm based on an RGB-D camera combined with a plane detection and random sampling consensus algorithm, which is used to reduce the amount of computation, improve the accuracy of feature point extraction, and reduce mismatching, thereby meeting the requirements of a mobile robot on accuracy and real-time performance.
In order to solve the technical problems, the technical scheme adopted by the invention is as follows:
an optimized ORB algorithm based on an RGB-D camera combined with a plane detection and random sampling consensus algorithm, the method comprising the steps of:
s1: acquiring image data by using an RGB-D camera, wherein the image data comprises a color image and a depth image;
s2: using an ORB algorithm to extract the feature points of the image data, and using a feature point uniformity evaluation method to judge the distribution uniformity of the feature points;
s3: generating point clouds and performing down-sampling on the point clouds for the image data part with uniformly distributed characteristic points;
s4: performing plane detection extraction on the point cloud subjected to down-sampling, and eliminating mismatching by using a random sampling consistency algorithm;
s5: for the image data part with non-uniform distribution of the feature points, using a set threshold value to extract the feature points and remove overlapped feature points by a non-maximum value inhibition method;
s6: and (5) for the point cloud after the mismatching is eliminated in the step S4 and the feature points after the overlapped feature points are eliminated in the step S5, projecting the point cloud and the feature points back to a two-dimensional image plane, and reconstructing and equalizing the gray level image.
Preferably, in S1, the RGB-D camera includes a Kinect camera.
Preferably, S2 specifically includes the following steps:
s21: using the feature point to extract an Oriented FAST algorithm, judging whether the feature point x is a feature point, when the feature point x is judged to be a feature point, calculating the main direction of the feature point, and naming the feature point as a key point, so that the key point has directionality;
the method for judging whether the feature point x is a feature point comprises the following steps: drawing a circle by taking the feature point x as a center, wherein the circle passes through n pixel points, and whether the distance between at least m continuous pixel points in the n pixel points arranged on the circumference and the feature point x satisfies a mean ratio lx+ t is greater or equal to lx-t is small, and if such a requirement is met, the feature point x is judged to be a feature point;
wherein lxRepresenting the distance between the feature point x and the pixel point on the circle; l represents a distance; t represents a threshold value, which is an adjustment amount of the range; n is 16; m is more than or equal to 9 and less than or equal to 12;
s22: using a characteristic point description rBRIEF algorithm to perform Gaussian smoothing processing on image data, selecting pixel points y in a neighborhood by taking key points as centers, forming n point pairs (xi, yi), wherein I (x, y) are compared with gray values, I represents the gray value, x > y takes 1, otherwise, 0 is taken, generating an n-dimensional characteristic descriptor, defining the n point pairs (xi, yi) as a 2 n matrix S,
Figure BDA0002689061160000031
the rotation of S is performed by means of theta,
Sθ=RθS (2)
in the formula (2), SθRepresenting a matrix with the rotation angle theta, wherein theta is the angle of rotating theta along the main direction of the characteristic point;
the neighborhood is a circle which takes the key point as the center and passes k pixel points to select a pixel point y; wherein 0< k < n;
s23: the characteristic point distribution uniformity evaluation method comprises the steps of segmenting image data through different dividing modes, and obtaining an image data part with uniformly distributed characteristic points and an image data part with non-uniformly distributed characteristic points after segmentation.
Further preferably, the method for evaluating the distribution uniformity of the feature points comprises the following steps: firstly, image data is preliminarily divided into a plurality of sub-regions SiFor each sub-region SiIs divided into several secondary subregions S againijSecond sub-region SijComprising Si1To SijIndividual region according to secondary sub-region SijEvaluating whether the characteristic points in the area are uniformly distributed or not according to the number of the characteristic points in the area; if Si1To SijThe number of the characteristic points is similar, and the similar calculation method comprises the following steps: by calculating the variance value of the feature points statistically distributed in the secondary subareas and judging according to the variance value, when the variance value is less than 15, S can be judgediAnd the characteristic points of the subareas are uniformly distributed, otherwise, the judgment is not uniform.
Further preferably, the image data segmentation method includes segmentation into center and peripheral directions, segmentation from top left to bottom right, or segmentation from bottom left to top right.
Preferably, S3 specifically includes the following steps:
s31: according to the color image and the depth image with uniformly distributed characteristic points, the point cloud is obtained by adopting the following formula (3),
Figure BDA0002689061160000041
equation (3) uses the pixel coordinate system o-u-v; c. CX,cY,fX,fYS is camera internal reference; u and v are pixel coordinates of the characteristic points; d being a characteristic pointDepth; the coordinates of the characteristic points are (X, Y, Z), and a plurality of points defined by the coordinates form a point cloud;
s32: and (4) processing by using a grid filter, and extracting a plane from the point cloud.
Preferably, S4 specifically includes the following steps:
s41: the formula for plane extraction is as follows:
aX+bY+cZ+d=0 (4)
in the formula (4), a, b and c represent constants;
s42: extracting a plane from the point cloud data with noise by using a random sampling consistency method, wherein the extraction judgment conditions are as follows: when the number of the remaining point clouds is larger than a threshold value g of the total number of the point clouds or the extraction plane is smaller than a threshold value h, extracting feature points, and after extracting the feature points, performing surface extraction on the remaining points again until the number of the remaining points reaches the threshold value h of the number of the extraction planes or the number of the remaining points is smaller than the threshold value g;
wherein g is more than or equal to 20 percent and less than or equal to 40 percent; h is more than or equal to 3 and less than or equal to 5.
Preferably, S5 specifically includes the following steps:
s51: changing the threshold t according to the oFAST algorithm of S2, wherein the changed threshold t is t', and reducing lx+ t' and lxThe range between t 'is adjusted according to the characteristic point extraction result, and the changed value of t' is determined to be optimal;
s52: if a plurality of key points exist in the neighborhood of a certain key point, comparing the values J of the feature points, wherein the value J is reserved in the largest mode, and the rest are deleted, wherein the value J is defined as follows:
Figure BDA0002689061160000051
in the formula (5), lxy-lxAnd lx-lxyRepresent the distance between the keypoint and the known feature point.
Preferably, in S6, the projection formula is:
Figure BDA0002689061160000052
in the formula (6), s is a scale factor, after projection, the gray level image of each plane is reconstructed, and after gray level histogram equalization is performed on the gray level image, the image is clearer and clearer, and noise brought by depth can be reduced.
The invention has the beneficial effects that: the accuracy of the algorithm is obviously higher than that of the ORB algorithm and the ICP algorithm which are used independently, and although parameters such as calibration errors of a camera have a little influence on a result, the accuracy of the algorithm is still far higher than that of other algorithms. In addition, the algorithm of the invention combines the ORB algorithm, and inserts the uniformity judgment of the distribution of the regional characteristic points and combines the plane extraction and random sampling consistency method, so the running time is slightly longer than that of the ORB algorithm which is used alone, but the practicability of the algorithm of the invention is not influenced. The ICP algorithm, however, has too much computation to run for its practical use and increases the hardware requirements of the mobile robot. The algorithm only sacrifices negligible extra running time, and can greatly improve the accuracy of extracting the characteristic points of the picture. Therefore, the invention can improve the accuracy of feature point extraction under the condition of keeping less running time, thereby improving the positioning precision and the real-time performance of the robot.
Drawings
FIG. 1 is a flow chart of the present invention;
FIG. 2 is a drawing of a freiburg1_ desk picture before (left) and after (right) extraction in the present invention;
FIG. 3 is a front (left) and back (right) extracted pictures from the cherry 1_ room picture in the present invention;
fig. 4 is a picture before (left) and after (right) extraction of the freiburg1_ teddy picture in the present invention.
Detailed Description
The following describes in detail a machine learning-based SCR denitration system prediction model optimization method according to the present invention with reference to the accompanying drawings and specific embodiments.
Example 1
As shown in fig. 1, an optimized ORB algorithm based on RGB-D camera combined with plane detection and random sampling consensus algorithm, the method comprises the following steps:
s1: acquiring image data by using an RGB-D camera, wherein the image data comprises a color image and a depth image;
s2: using an ORB algorithm to extract the feature points of the image data, and using a feature point uniformity evaluation method to judge the distribution uniformity of the feature points;
s3: generating point clouds and performing down-sampling on the point clouds for the image data part with uniformly distributed characteristic points;
s4: performing plane detection extraction on the point cloud subjected to down-sampling, and eliminating mismatching by using a random sampling consistency algorithm;
s5: for the image data part with non-uniform distribution of the feature points, using a set threshold value to extract the feature points and remove overlapped feature points by a non-maximum value inhibition method;
s6: and (5) for the point cloud after the mismatching is eliminated in the step S4 and the feature points after the overlapped feature points are eliminated in the step S5, projecting the point cloud and the feature points back to a two-dimensional image plane, and reconstructing and equalizing the gray level image.
Preferably, in S1, the RGB-D camera includes a Kinect camera.
Preferably, S2 specifically includes the following steps:
s21: using the feature point to extract an organized FAST algorithm (namely an oFAST algorithm), judging whether the feature point x is a feature point, calculating the main direction of the feature point when the feature point x is judged to be a feature point, and naming the feature point as a key point to ensure that the key point (namely a detector) has directionality;
the method for judging whether the feature point x is a feature point comprises the following steps: drawing a circle by taking the feature point x as a center, wherein the circle passes through 16 pixel points, and whether the distance between at least 12 continuous pixel points in the 16 pixel points arranged on the circumference and the feature point x satisfies the average ratio lx+ t is greater or equal to lx-t is small, and if such a requirement is met, the feature point x is judged to be a feature point;
wherein lxStand for special purposeDistances between the characteristic point x and 12 continuous pixel points on the circle respectively; l represents a distance; t represents a threshold value, is an adjustment quantity of a range, sets a t value to screen out pixel points with the distance close to an x point, if the t value is not set or is too small, the pixel points meeting the judgment condition are too many to cause too many characteristic points, and if the t value is too large, the judged characteristic points are too few;
s22: using a characteristic point description rBRIEF algorithm to perform Gaussian smoothing processing on image data, selecting pixel points y in a neighborhood by taking key points as centers, and forming 6 point pairs (xi, yi), wherein the point pairs are randomly selected, I (x, y) compares gray values with each other, I represents the gray value, x > y takes 1, otherwise 0 is taken, an n-dimensional characteristic descriptor is generated, n point pairs (xi, yi) are defined as a 2 & lt6 & gt matrix S,
Figure BDA0002689061160000071
the rotation of S is performed by means of theta,
Sθ=RθS (2)
in the formula (2), SθRepresenting a matrix with the rotation angle theta, wherein theta is the angle of rotating theta along the main direction of the characteristic point;
the neighborhood is a circle which takes the key point as the center and passes k pixel points to select a pixel point y; wherein 0< k < n;
s23: the characteristic point distribution uniformity evaluation method comprises the steps of segmenting image data through different dividing modes, and obtaining an image data part with uniformly distributed characteristic points and an image data part with non-uniformly distributed characteristic points after segmentation.
Further preferably, the method for evaluating the distribution uniformity of the feature points comprises the following steps: firstly, image data is preliminarily divided into a plurality of sub-regions SiFor each sub-region SiIs divided into several secondary subregions S againijSecond sub-region SijComprising Si1To SijIndividual region according to secondary sub-region SijEvaluating whether the characteristic points in the area are uniformly distributed or not according to the number of the characteristic points in the area; if Si1To SijThe number of the characteristic points are all close to each otherThe calculation method comprises the following steps: by calculating the variance value of the feature points statistically distributed in the secondary subareas and judging according to the variance value, when the variance value is less than 15, S can be judgediAnd the characteristic points of the subareas are uniformly distributed, otherwise, the judgment is not uniform.
Further preferably, the image data segmentation method includes segmentation into center and peripheral directions, segmentation from top left to bottom right, or segmentation from bottom left to top right.
In the embodiment, the feature point distribution condition of each region is judged by image segmentation, so that different algorithms are applied according to different distribution conditions.
Preferably, S3 specifically includes the following steps:
s31: according to the color image and the depth image with uniformly distributed characteristic points, the point cloud is obtained by adopting the following formula (3),
Figure BDA0002689061160000072
equation (3) uses the pixel coordinate system o-u-v; c. CX,cY,fX,fYS is camera internal reference; u and v are pixel coordinates of the characteristic points; d is the depth of the feature point; the coordinates of the characteristic points are (X, Y, Z), and a plurality of points defined by the coordinates form a point cloud;
s32: a grid filter is used for processing, a plane is extracted from the point cloud, and points with longer distances are filtered out by using a z-direction interval filter. Points that are further away are planes that are too far away from other points, resulting in only one point in the extracted plane if not filtered out, thereby increasing the number of non-conforming feature points.
The present embodiment obtains color information (i.e., gray scale) through a color image and distance information through a depth image, so that 3D camera coordinates of pixels can be calculated to generate a point cloud. The present embodiment generates a point cloud from an RGB image and a depth image.
Preferably, S4 specifically includes the following steps:
s41: the formula for plane extraction is as follows:
aX+bY+cZ+d=0 (4)
in the formula (4), a, b and c represent constants;
s42: extracting a plane from the point cloud data with noise by using a random sampling consistency method, wherein the extraction judgment conditions are as follows: and extracting the feature points when the number of the remaining points of the point cloud is more than 30% of the total number or the extraction plane is less than 3, and after extracting the feature points, performing surface extraction on the remaining points again until the number of the extracted planes is 3 or the number of the remaining points is less than 30% of the total number.
Preferably, S5 specifically includes the following steps:
s51: changing the threshold t according to the oFAST algorithm of S2, wherein the changed threshold t is t', and reducing lx+ t' and lxThe range between t 'is adjusted according to the characteristic point extraction result, and the changed value of t' is determined to be optimal;
s52: if a plurality of key points exist in the neighborhood of a certain key point, comparing the values J of the feature points, wherein the value J is reserved in the largest mode, and the rest are deleted, wherein the value J is defined as follows:
Figure BDA0002689061160000081
in the formula (5), lxy-lxAnd lx-lxyRepresent the distance between the keypoint and the known feature point.
Preferably, in S6, the projection formula is:
Figure BDA0002689061160000082
in the formula (6), s is a scale factor and can be selected according to actual conditions, after projection, the gray level image of each plane is reconstructed, and after gray level histogram equalization is performed on the gray level image, the image is clearer and clearer, and noise brought by depth can be reduced.
And (3) comparing the extraction accuracy of the feature points of different algorithms and the algorithm with the running time according to the values in the tables 1 and 2.
As shown in fig. 2 to 4, the invention respectively extracts feature points of the desk, the room and the teddy bear, and the extracted feature points are distributed uniformly and have representativeness and accuracy, so that the images reconstructed subsequently are clearer.
Table 1 is a comparison table of feature point extraction accuracy rates of 3 pictures of the algorithm, the ORB algorithm and the ICP algorithm
Image classification freiburg1_desk freiburg1_room freiburg1_teddy
ORB algorithm 83.45% 76.09% 85.64%
ICP algorithm 83.18% 83.56% 85.08%
Algorithm of the invention 93.44% 90.28% 97.21%
Table 2 shows the run-time comparison table of 3 pictures of the algorithm of the present invention, the ORB algorithm and the ICP algorithm
Image classification freiburg1_desk freiburg1_room freiburg1_teddy
ORB algorithm 0.7523 0.9735 0.6285
ICP algorithm 1.0028 1.5236 0.9658
Algorithm of the invention 0.7886 1.0032 0.6422
From table 1, it can be seen that, regardless of the feature point extraction of a single object or the feature point extraction of a plurality of objects, the accuracy of the algorithm of the present invention is significantly higher than that of the ORB algorithm and the ICP algorithm used alone, and although parameters such as calibration error of the camera have some influence on the result, the accuracy of the algorithm of the present invention is still much higher than that of other algorithms.
As can be seen from table 2, since the algorithm of the present invention combines the ORB algorithm, and inserts the uniformity determination for the distribution of the region feature points and combines the plane extraction and random sampling consistency method, the running time is slightly longer than that of the ORB algorithm used alone, but the practicability of the algorithm of the present invention is not affected. The ICP algorithm, however, has too much computation to run for its practical use and increases the hardware requirements of the mobile robot. The algorithm only sacrifices negligible extra running time, and can greatly improve the accuracy of extracting the characteristic points of the picture.
Example 2
This example differs from example 1 only in that: in S21, m is 9; and in S42, extracting feature points when the number of the remaining point cloud is more than 40% of the total number or the extraction plane is less than 5.
Example 3
This example differs from example 1 only in that: in S21, m is 10; in S42, if the remaining point number of the point cloud is more than 20% of the total number or the extraction plane is less than 4, extracting the feature points.
In the description provided herein, numerous specific details are set forth. It is understood, however, that embodiments of the invention may be practiced without these specific details. In some instances, well-known methods, structures and techniques have not been shown in detail in order not to obscure an understanding of this description.
In the description of the present invention, it is to be understood that the terms "upper", "lower", "front", "rear", "left", "right", and the like indicate orientations or positional relationships based on those shown in the drawings, and are only for convenience of description and simplicity of description, but do not indicate or imply that the referred device or element must have a specific orientation, be constructed in a specific orientation, and be operated, and thus, should not be construed as limiting the present invention.
Similarly, it should be appreciated that in the foregoing description of exemplary embodiments of the invention, various features of the invention are sometimes grouped together in a single embodiment, figure, or description thereof for the purpose of streamlining the disclosure and aiding in the understanding of one or more of the various inventive aspects. However, the disclosed method should not be interpreted as reflecting an intention that: that the invention as claimed requires more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive aspects lie in less than all features of a single foregoing disclosed embodiment. Thus, the claims following the detailed description are hereby expressly incorporated into this detailed description, with each claim standing on its own as a separate embodiment of this invention.
As used herein, unless otherwise specified the use of the ordinal adjectives "first", "second", "third", etc., to describe a common object, merely indicate that different instances of like objects are being referred to, and are not intended to imply that the objects so described must be in a given sequence, either temporally, spatially, in ranking, or in any other manner.
While the invention has been described with respect to a limited number of embodiments, those skilled in the art, having benefit of this description, will appreciate that other embodiments can be devised which do not depart from the scope of the invention as described herein. Furthermore, it should be noted that the language used in the specification has been principally selected for readability and instructional purposes, and may not have been selected to delineate or circumscribe the inventive subject matter. Accordingly, many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the appended claims. The present invention has been disclosed in an illustrative rather than a restrictive sense, and the scope of the present invention is defined by the appended claims.
The above description is only of the preferred embodiments of the present invention, and it should be noted that: it will be apparent to those skilled in the art that various modifications and adaptations can be made without departing from the principles of the invention and these are intended to be within the scope of the invention.

Claims (8)

1. An optimized ORB algorithm based on an RGB-D camera combined with a plane detection and random sampling consensus algorithm is characterized by comprising the following steps:
s1: acquiring image data by using an RGB-D camera, wherein the image data comprises a color image and a depth image;
s2: using an ORB algorithm to extract the feature points of the image data, and using a feature point uniformity evaluation method to judge the distribution uniformity of the feature points;
s3: generating point clouds and performing down-sampling on the point clouds for the image data part with uniformly distributed characteristic points;
s4: performing plane detection extraction on the point cloud subjected to down-sampling, and eliminating mismatching by using a random sampling consistency algorithm;
s5: for the image data part with non-uniform distribution of the feature points, using a set threshold value to extract the feature points and remove overlapped feature points by a non-maximum value inhibition method;
s6: and (5) for the point cloud after the mismatching is eliminated in the step S4 and the feature points after the overlapped feature points are eliminated in the step S5, projecting the point cloud and the feature points back to a two-dimensional image plane, and reconstructing and equalizing the gray level image.
2. The optimized ORB algorithm based on the combination of RGB-D camera with planar detection and random sampling consensus algorithm as claimed in claim 1, wherein in S1, the RGB-D camera comprises a Kinect camera.
3. The optimized ORB algorithm based on the RGB-D camera combined with the planar detection and random sampling consensus algorithm as claimed in claim 1, wherein S2 comprises the following steps:
s21: using the feature point to extract an Oriented FAST algorithm, judging whether the feature point x is a feature point, when the feature point x is judged to be a feature point, calculating the main direction of the feature point, and naming the feature point as a key point, so that the key point has directionality;
the method for judging whether the feature point x is a feature point comprises the following steps: drawing a circle by taking the feature point x as a center, wherein the circle passes through n pixel points, and whether the distance between at least m continuous pixel points in the n pixel points arranged on the circumference and the feature point x satisfies a mean ratio lx+ t is greater or equal to lx-t is small, and if such a requirement is met, the feature point x is judged to be a feature point; wherein lxRepresenting the distance between the feature point x and the pixel point on the circle; l represents a distance; t represents a threshold value, which is an adjustment amount of the range; n is 16; m is more than or equal to 9 and less than or equal to 12;
s22: using a characteristic point description rBRIEF algorithm to perform Gaussian smoothing processing on image data, selecting pixel points y in a neighborhood by taking key points as centers, forming n point pairs (xi, yi), wherein I (x, y) are compared with gray values, I represents the gray value, x > y takes 1, otherwise, 0 is taken, generating an n-dimensional characteristic descriptor, defining the n point pairs (xi, yi) as a 2 n matrix S,
Figure FDA0002689061150000021
the rotation of S is performed by means of theta,
Sθ=RθS (2)
in the formula (2), SθRepresenting a matrix with the rotation angle theta, wherein theta is the angle of rotating theta along the main direction of the characteristic point;
the neighborhood is a circle which takes the key point as the center and passes k pixel points to select a pixel point y; wherein 0< k < n;
s23: the characteristic point distribution uniformity evaluation method comprises the steps of segmenting image data through different dividing modes, and obtaining an image data part with uniformly distributed characteristic points and an image data part with non-uniformly distributed characteristic points after segmentation.
4. The optimized ORB algorithm based on the RGB-D camera combined with the planar detection and random sampling consensus algorithm as claimed in claim 3, wherein the evaluation method of the feature point distribution uniformity is as follows: firstly, image data is preliminarily divided into a plurality of sub-regions SiFor each sub-region SiIs divided into several secondary subregions S againijSecond sub-region SijComprising Si1To SijIndividual region according to secondary sub-region SijEvaluating whether the characteristic points in the area are uniformly distributed or not according to the number of the characteristic points in the area; if it isSi1To SijThe number of the characteristic points is similar, and the similar calculation method comprises the following steps: by calculating the variance value of the feature points statistically distributed in the secondary subareas and judging according to the variance value, when the variance value is less than 15, S can be judgediAnd the characteristic points of the subareas are uniformly distributed, otherwise, the judgment is not uniform.
5. The optimized ORB algorithm based on the RGB-D camera combined with the flat panel detection and random sampling consensus algorithm as claimed in claim 3, wherein the image data segmentation method comprises segmentation into center and peripheral directions, from top left to bottom right or from bottom left to top right.
6. The optimized ORB algorithm based on the RGB-D camera combined with the planar detection and random sampling consensus algorithm as claimed in claim 5, wherein S3 comprises the following steps:
s31: according to the color image and the depth image with uniformly distributed characteristic points, the point cloud is obtained by adopting the following formula (3),
Figure FDA0002689061150000022
equation (3) uses the pixel coordinate system o-u-v; c. CX,cY,fX,fYS is camera internal reference; u and v are pixel coordinates of the characteristic points; d is the depth of the feature point; the coordinates of the characteristic points are (X, Y, Z), and a plurality of points defined by the coordinates form a point cloud;
s32: and (4) processing by using a grid filter, and extracting a plane from the point cloud.
7. The optimized ORB algorithm based on the RGB-D camera combined with the planar detection and random sampling consensus algorithm as claimed in claim 6, wherein S4 comprises the following steps:
s41: the formula for plane extraction is as follows:
aX+bY+cZ+d=0 (4)
in the formula (4), a, b and c represent constants;
s42: extracting a plane from the point cloud data with noise by using a random sampling consistency method, wherein the extraction judgment conditions are as follows: when the number of the remaining point clouds is larger than a threshold value g of the total number of the point clouds or the extraction plane is smaller than a threshold value h, extracting feature points, and after extracting the feature points, performing surface extraction on the remaining points again until the number of the remaining points reaches the threshold value h of the number of the extraction planes or the number of the remaining points is smaller than the threshold value g;
wherein g is more than or equal to 20 percent and less than or equal to 40 percent; h is more than or equal to 3 and less than or equal to 5.
8. The optimized ORB algorithm based on the RGB-D camera combined with the planar detection and random sampling consensus algorithm as claimed in claim 7, wherein S5 comprises the following steps:
s51: changing the threshold t according to the oFAST algorithm of S2, wherein the changed threshold t is t', and reducing lx+ t' and lxThe range between t 'is adjusted according to the characteristic point extraction result, and the changed value of t' is determined to be optimal;
s52: if a plurality of key points exist in the neighborhood of a certain key point, comparing the values J of the feature points, wherein the value J is reserved in the largest mode, and the rest are deleted, wherein the value J is defined as follows:
Figure FDA0002689061150000031
in the formula (5), lxy-lxAnd lx-lxyRepresent the distance between the keypoint and the known feature point.
The optimized ORB algorithm based on the RGB-D camera combined with the flat panel detection and random sampling consensus algorithm as claimed in claim 1, wherein in S6, the projection formula is:
Figure FDA0002689061150000032
in the formula (6), s is a scale factor, after projection, the gray level image of each plane is reconstructed, and after gray level histogram equalization is performed on the gray level image, the image is clearer and clearer, and noise brought by depth can be reduced.
CN202010985540.5A 2020-09-18 2020-09-18 Optimized ORB algorithm based on RGB-D camera combined plane detection and random sampling coincidence algorithm Active CN112115953B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010985540.5A CN112115953B (en) 2020-09-18 2020-09-18 Optimized ORB algorithm based on RGB-D camera combined plane detection and random sampling coincidence algorithm

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010985540.5A CN112115953B (en) 2020-09-18 2020-09-18 Optimized ORB algorithm based on RGB-D camera combined plane detection and random sampling coincidence algorithm

Publications (2)

Publication Number Publication Date
CN112115953A true CN112115953A (en) 2020-12-22
CN112115953B CN112115953B (en) 2023-07-11

Family

ID=73800133

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010985540.5A Active CN112115953B (en) 2020-09-18 2020-09-18 Optimized ORB algorithm based on RGB-D camera combined plane detection and random sampling coincidence algorithm

Country Status (1)

Country Link
CN (1) CN112115953B (en)

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112752028A (en) * 2021-01-06 2021-05-04 南方科技大学 Pose determination method, device and equipment of mobile platform and storage medium
CN112783995A (en) * 2020-12-31 2021-05-11 杭州海康机器人技术有限公司 V-SLAM map checking method, device and equipment
US11403069B2 (en) 2017-07-24 2022-08-02 Tesla, Inc. Accelerated mathematical engine
US11409692B2 (en) 2017-07-24 2022-08-09 Tesla, Inc. Vector computational unit
US11487288B2 (en) 2017-03-23 2022-11-01 Tesla, Inc. Data synthesis for autonomous control systems
US11537811B2 (en) 2018-12-04 2022-12-27 Tesla, Inc. Enhanced object detection for autonomous vehicles based on field view
US11562231B2 (en) 2018-09-03 2023-01-24 Tesla, Inc. Neural networks for embedded devices
US11561791B2 (en) 2018-02-01 2023-01-24 Tesla, Inc. Vector computational unit receiving data elements in parallel from a last row of a computational array
US11567514B2 (en) 2019-02-11 2023-01-31 Tesla, Inc. Autonomous and user controlled vehicle summon to a target
US11610117B2 (en) 2018-12-27 2023-03-21 Tesla, Inc. System and method for adapting a neural network model on a hardware platform
US11636333B2 (en) 2018-07-26 2023-04-25 Tesla, Inc. Optimizing neural network structures for embedded systems
US11665108B2 (en) 2018-10-25 2023-05-30 Tesla, Inc. QoS manager for system on a chip communications
US11681649B2 (en) 2017-07-24 2023-06-20 Tesla, Inc. Computational array microprocessor system using non-consecutive data formatting
US11734562B2 (en) 2018-06-20 2023-08-22 Tesla, Inc. Data pipeline and deep learning system for autonomous driving
US11748620B2 (en) 2019-02-01 2023-09-05 Tesla, Inc. Generating ground truth for machine learning from time series elements
US11790664B2 (en) 2019-02-19 2023-10-17 Tesla, Inc. Estimating object properties using visual image data
US11816585B2 (en) 2018-12-03 2023-11-14 Tesla, Inc. Machine learning models operating at different frequencies for autonomous vehicles
US11841434B2 (en) 2018-07-20 2023-12-12 Tesla, Inc. Annotation cross-labeling for autonomous control systems
US11893393B2 (en) 2017-07-24 2024-02-06 Tesla, Inc. Computational array microprocessor system with hardware arbiter managing memory requests
US11893774B2 (en) 2018-10-11 2024-02-06 Tesla, Inc. Systems and methods for training machine models with augmented data
US12014553B2 (en) 2019-02-01 2024-06-18 Tesla, Inc. Predicting three-dimensional features for autonomous driving
US12136030B2 (en) 2023-03-16 2024-11-05 Tesla, Inc. System and method for adapting a neural network model on a hardware platform

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107220995A (en) * 2017-04-21 2017-09-29 西安交通大学 A kind of improved method of the quick point cloud registration algorithms of ICP based on ORB characteristics of image
CN110414533A (en) * 2019-06-24 2019-11-05 东南大学 A kind of feature extracting and matching method for improving ORB
US20190362178A1 (en) * 2017-11-21 2019-11-28 Jiangnan University Object Symmetry Axis Detection Method Based on RGB-D Camera

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107220995A (en) * 2017-04-21 2017-09-29 西安交通大学 A kind of improved method of the quick point cloud registration algorithms of ICP based on ORB characteristics of image
US20190362178A1 (en) * 2017-11-21 2019-11-28 Jiangnan University Object Symmetry Axis Detection Method Based on RGB-D Camera
CN110414533A (en) * 2019-06-24 2019-11-05 东南大学 A kind of feature extracting and matching method for improving ORB

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
JIANGYING QIN等: "Accumulative Errors Optimizaton for Visual Odometry od ORB-SLAM2 Based on RGB-D Cameras", ISPRS INTERNATIONAL JOURNAL OF GEO-INFORMATION *

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11487288B2 (en) 2017-03-23 2022-11-01 Tesla, Inc. Data synthesis for autonomous control systems
US12020476B2 (en) 2017-03-23 2024-06-25 Tesla, Inc. Data synthesis for autonomous control systems
US11403069B2 (en) 2017-07-24 2022-08-02 Tesla, Inc. Accelerated mathematical engine
US11681649B2 (en) 2017-07-24 2023-06-20 Tesla, Inc. Computational array microprocessor system using non-consecutive data formatting
US11409692B2 (en) 2017-07-24 2022-08-09 Tesla, Inc. Vector computational unit
US12086097B2 (en) 2017-07-24 2024-09-10 Tesla, Inc. Vector computational unit
US11893393B2 (en) 2017-07-24 2024-02-06 Tesla, Inc. Computational array microprocessor system with hardware arbiter managing memory requests
US11561791B2 (en) 2018-02-01 2023-01-24 Tesla, Inc. Vector computational unit receiving data elements in parallel from a last row of a computational array
US11797304B2 (en) 2018-02-01 2023-10-24 Tesla, Inc. Instruction set architecture for a vector computational unit
US11734562B2 (en) 2018-06-20 2023-08-22 Tesla, Inc. Data pipeline and deep learning system for autonomous driving
US11841434B2 (en) 2018-07-20 2023-12-12 Tesla, Inc. Annotation cross-labeling for autonomous control systems
US12079723B2 (en) 2018-07-26 2024-09-03 Tesla, Inc. Optimizing neural network structures for embedded systems
US11636333B2 (en) 2018-07-26 2023-04-25 Tesla, Inc. Optimizing neural network structures for embedded systems
US11562231B2 (en) 2018-09-03 2023-01-24 Tesla, Inc. Neural networks for embedded devices
US11983630B2 (en) 2018-09-03 2024-05-14 Tesla, Inc. Neural networks for embedded devices
US11893774B2 (en) 2018-10-11 2024-02-06 Tesla, Inc. Systems and methods for training machine models with augmented data
US11665108B2 (en) 2018-10-25 2023-05-30 Tesla, Inc. QoS manager for system on a chip communications
US11816585B2 (en) 2018-12-03 2023-11-14 Tesla, Inc. Machine learning models operating at different frequencies for autonomous vehicles
US11537811B2 (en) 2018-12-04 2022-12-27 Tesla, Inc. Enhanced object detection for autonomous vehicles based on field view
US11908171B2 (en) 2018-12-04 2024-02-20 Tesla, Inc. Enhanced object detection for autonomous vehicles based on field view
US11610117B2 (en) 2018-12-27 2023-03-21 Tesla, Inc. System and method for adapting a neural network model on a hardware platform
US12014553B2 (en) 2019-02-01 2024-06-18 Tesla, Inc. Predicting three-dimensional features for autonomous driving
US11748620B2 (en) 2019-02-01 2023-09-05 Tesla, Inc. Generating ground truth for machine learning from time series elements
US11567514B2 (en) 2019-02-11 2023-01-31 Tesla, Inc. Autonomous and user controlled vehicle summon to a target
US11790664B2 (en) 2019-02-19 2023-10-17 Tesla, Inc. Estimating object properties using visual image data
CN112783995B (en) * 2020-12-31 2022-06-03 杭州海康机器人技术有限公司 V-SLAM map checking method, device and equipment
CN112783995A (en) * 2020-12-31 2021-05-11 杭州海康机器人技术有限公司 V-SLAM map checking method, device and equipment
CN112752028A (en) * 2021-01-06 2021-05-04 南方科技大学 Pose determination method, device and equipment of mobile platform and storage medium
US12136030B2 (en) 2023-03-16 2024-11-05 Tesla, Inc. System and method for adapting a neural network model on a hardware platform

Also Published As

Publication number Publication date
CN112115953B (en) 2023-07-11

Similar Documents

Publication Publication Date Title
CN112115953B (en) Optimized ORB algorithm based on RGB-D camera combined plane detection and random sampling coincidence algorithm
US8199977B2 (en) System and method for extraction of features from a 3-D point cloud
CN111210429B (en) Point cloud data partitioning method and device and obstacle detection method and device
CN107093205B (en) A kind of three-dimensional space building window detection method for reconstructing based on unmanned plane image
US9754377B2 (en) Multi-resolution depth estimation using modified census transform for advanced driver assistance systems
CN111582054B (en) Point cloud data processing method and device and obstacle detection method and device
CN106338733B (en) Forward-Looking Sonar method for tracking target based on frogeye visual characteristic
CN106033621B (en) A kind of method and device of three-dimensional modeling
Chen et al. Transforming a 3-d lidar point cloud into a 2-d dense depth map through a parameter self-adaptive framework
EP2166375B1 (en) System and method of extracting plane features
CN111768447A (en) Monocular camera object pose estimation method and system based on template matching
CN114549549B (en) Dynamic target modeling tracking method based on instance segmentation in dynamic environment
CN111553946A (en) Method and device for removing ground point cloud and obstacle detection method and device
EP2054835A1 (en) Target orientation
Zhou et al. Monet3d: Towards accurate monocular 3d object localization in real time
CN112800966B (en) Sight tracking method and electronic equipment
CN107122782B (en) Balanced semi-dense stereo matching method
Chen et al. Multi-stage matching approach for mobile platform visual imagery
Sakai et al. Phase-based window matching with geometric correction for multi-view stereo
WO2020197495A1 (en) Method and system for feature matching
CN113409242A (en) Intelligent monitoring method for point cloud of rail intersection bow net
CN116740160A (en) Millisecond level multi-plane real-time extraction method and device in complex traffic scene
CN111179327A (en) Depth map calculation method
Xu et al. Hybrid plane fitting for depth estimation
Stentoumis et al. Implementing an adaptive approach for dense stereo-matching

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant