CN108682026B - Binocular vision stereo matching method based on multi-matching element fusion - Google Patents

Binocular vision stereo matching method based on multi-matching element fusion Download PDF

Info

Publication number
CN108682026B
CN108682026B CN201810249294.XA CN201810249294A CN108682026B CN 108682026 B CN108682026 B CN 108682026B CN 201810249294 A CN201810249294 A CN 201810249294A CN 108682026 B CN108682026 B CN 108682026B
Authority
CN
China
Prior art keywords
window
matching
pixel
algorithm
adaptive
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810249294.XA
Other languages
Chinese (zh)
Other versions
CN108682026A (en
Inventor
孙福明
杜仁鹏
蔡希彪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Applied Technology Co Ltd
Original Assignee
Jiang Dabai
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiang Dabai filed Critical Jiang Dabai
Priority to CN201810249294.XA priority Critical patent/CN108682026B/en
Publication of CN108682026A publication Critical patent/CN108682026A/en
Application granted granted Critical
Publication of CN108682026B publication Critical patent/CN108682026B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The stereo matching method based on the fusion of multiple matching elements for binocular vision comprises a computer, two cameras and a stereo matching method based on the fusion of multiple matching elements. The camera is a ZED binocular camera. The stereo matching method of multi-matching element fusion comprises three stages: an initial matching cost stage, a cost function aggregation stage and a parallax post-processing stage. In the initial matching cost stage, a cost function containing fusion of a color primitive and a gradient primitive is designed, and self-adaptive adjustment is carried out through a Kalman coefficient alpha; in a cost function aggregation stage, a self-adaptive matching window is designed based on RGB color and distance relation, and aggregation is carried out through correlation between a reference image and a cost function in the window; and finally, performing parallax postprocessing through LRC left-right consistency and sub-pixel level-based adaptive weight median filtering. Compared with the classical adaptive weight algorithm, the method improves the accuracy and the real-time performance of the algorithm and shows higher robustness in a complex area.

Description

Binocular vision stereo matching method based on multi-matching element fusion
Technical Field
The invention relates to a binocular vision stereo matching method based on multi-matching element fusion, which can be widely applied to the technical fields of unmanned driving, robot navigation and the like.
Background
The stereo matching is widely applied to the fields of unmanned driving, robot navigation and the like as a core technology in binocular vision. The basic principle is that a two-dimensional image of a scene is obtained through a binocular camera, the parallax of points to be matched is obtained through a matching algorithm, and then the depth information of the scene is obtained. The stereo matching algorithm is widely concerned by scholars due to its high accuracy characteristic. However, due to the complexity of texture in natural scenes and the discontinuity of scene depth, the practical application of stereo matching algorithm is restricted.
Currently, mainstream stereo matching algorithms can be divided into global stereo matching algorithms and local stereo matching algorithms. The global stereo matching method is complex in calculation and low in efficiency due to the fact that operation of an energy function is involved, and actual requirements are difficult to meet. Based on a local stereo matching method, only local data items are operated in a cost function aggregation stage, the real-time performance is high, but the accuracy is relatively low, until Yoon proposes a classical adaptive support weight Algorithm (ASW), the accuracy is qualitatively improved, but the characteristic and texture information of an image cannot be reflected due to the fixed window.
The local stereo matching algorithm can be subdivided into stereo matching methods based on regions, characteristics and phases according to different matching primitives. The self-adaptive support weight method has the advantages that the accuracy of the algorithm in a complex texture region is low due to the fact that the matching primitive is single and the window is fixed, and the real-time performance of the algorithm is low due to the fact that the weight calculation is complex.
Aiming at the problems of the classical algorithm, some scholars express the initial matching cost of the image based on various matching primitive information such as gradient and color, and the accuracy of the algorithm is further improved. However, complex weight calculation is not abandoned, so that the real-time performance of the algorithm is still low and the robustness is weak.
Lin et al use linear functions to fit Gaussian weight functions in the original algorithm, down-sample the original image to obtain Gaussian pyramid samples, and use hierarchical clustering algorithm to calculate cost aggregation functions, thereby improving algorithm instantaneity and accuracy. However, due to the limitation of the sampling times, the sampled image is too blurred, and the robustness of the algorithm is poor.
Researchers have studied adaptive windows all the time, and abandon some apparent wrong parallax values by utilizing the coincidence degree of a reference image and an image to be matched, and abandon the operation of complex weight values in a cost aggregation stage. However, this makes the accuracy of the algorithm no better than the classical one, since the correlation between the reference image and the matching cost function is not taken into account.
Aiming at the problems, in order to improve the robustness of the algorithm in the practical scene application, the color matching primitive and the gradient matching primitive are fused in the initial matching cost stage, and the proportionality coefficient alpha of the color and the gradient matching primitive is adaptively adjusted by utilizing the Cones image which conforms to the natural scene condition of the Middlebury platform; in a cost function aggregation stage, aiming at the problem that the accuracy of the algorithm is low under a complex texture condition caused by a fixed window in the traditional method, a central pixel point is expanded according to color information and spatial distance information between pixels, and then the traditional complex energy function and weight calculation are replaced by correlation calculation between an initial matching cost function and a reference image pixel for cost aggregation, so that the accuracy of the algorithm is improved, and the time complexity of the algorithm is reduced; and finally, performing LRC left-right consistency detection on the obtained disparity map and performing sub-pixel level-based adaptive weight median filtering operation, thereby improving the accuracy of the algorithm again.
Disclosure of Invention
The invention aims to provide a binocular vision stereo matching method based on multi-matching element fusion, aiming at the problems in the prior art, the average mismatching rate of a disparity map obtained by adopting the method is greatly improved, the image is clear and accurate, and the precision of the running track of a robot or an unmanned aerial vehicle is greatly improved.
The technical scheme is as follows:
a binocular vision stereo matching method based on multi-matching element fusion comprises a computer, two cameras and the stereo matching method based on multi-matching element fusion, and is characterized in that:
the two cameras adopt ZED binocular cameras produced by Astrolabes company, and the ZED binocular cameras are set as follows:
1) image acquisition:
the ZED SDK and the CUDA are downloaded and connected with a computer through a USB3.0 interface. Connecting a ZED binocular camera through a webcam function in an MATLAB, and acquiring pictures through a snapshot function;
2) calibrating a camera:
the aim of camera calibration is to obtain accurate internal and external parameters of a camera. The internal parameters mainly comprise the focal lengths of the left lens and the right lens and the baseline distance; the extrinsic parameters mainly include a rotation matrix of the two cameras relative to a world coordinate system and a relative translation matrix of the left camera and the right camera. The invention obtains the default internal and external parameters of the camera from the official manual;
3) setting image parameters:
and performing epipolar correction through the calibration parameters to enable the acquired left and right images to meet epipolar constraint conditions. Resetting parameters by embedding a ZED Explorer plug-in into the ZED SDK;
4) stereo matching:
stereo matching is a core part of the binocular vision system. The purpose of stereo matching is to perform imaging point matching on the acquired left and right images, obtain a parallax value through the matching points, and obtain depth information of a scene.
The ZED binocular camera not only can acquire high-quality image information, but also has the advantages of small size and low power consumption. Especially, because the ZED SDK function library and the CUDA are embedded for parallel processing, the processing time of the system is greatly reduced. The ZED binocular camera can be installed on a robot or an unmanned plane. The actual scene shot by the two eyes can achieve the purpose of a real scene through the stereo matching processing of the fusion of multiple matching primitives, and then a navigation instruction is sent to a robot or unmanned aerial vehicle control and driving system through the processing of a computer arranged on the robot or the unmanned aerial vehicle; or the navigation instruction is sent to the robot or the unmanned aerial vehicle control and drive system through a computer arranged outside the robot or the unmanned aerial vehicle, connected with the computer through a local area network, an Ethernet or a wire and processed by the computer. The precision of the running track of the robot or the unmanned aerial vehicle is greatly improved.
The stereo matching module method for multi-matching element fusion comprises the following processes:
the invention adopts an improved ASW stereo matching method.
The currently known ASW stereo matching algorithm is:
in the stereo matching algorithm, it is generally defaulted that two images satisfy epipolar constraint conditions, that is, matching points corresponding to left and right images are located at the same row position of the two images. The core of the ASW algorithm is that when the support weights are used to measure the similarity between image pixels, the support of neighboring pixels is only valid if they come from the same depth, which has the same disparity as the pixel to be matched. Therefore, the support weight w of the pixels around the window is proportional to the window disparity probability Pr, as shown in equation (1):
w(p,q)∝Pr(dp=dq) (1)
wherein p is the pixel to be matched, q is the other pixels except the pixel to be matched in the window, and d is the required parallax. w (p, q) is related to the color and distance of the image, as shown in equation (2):
w(p,q)=k·f(Δcpq,Δgpq) (2)
wherein, Δ cpq,ΔgpqRespectively representing the distances of two points p and q in an LAB color space and a geometric space, wherein k is a proportionality coefficient, a specific numerical value is obtained through experiments, f is a Laplace kernel function, and the two are independent from each other, as shown in a formula (3):
f(Δcpq,Δgpq)=f(Δcpq)·f(Δgpq) (3)
wherein Δ cpq,ΔgpqThe calculation is shown in equation (4) (5):
Figure GDA0003001593790000031
Figure GDA0003001593790000032
cp=|Lp,ap,bp|,cq=|Lq,aq,bql is Lab three-channel color space chroma value, L component in Lab color space is used for representing pixel brightness, and the value range is [0, 100 ]]From pure black to pure white; a represents the range from red to green, and the value range is [127, -128 ]](ii) a b represents the range from yellow to blue, and the value range is [127 to 128 ]]。(px,py) And (q)x,qy) Are coordinate values of the geometric space. The grouping strength is defined using a laplacian kernel, as shown in equations (6) (7).
Figure GDA0003001593790000033
Figure GDA0003001593790000034
γc,γpObtaining gamma through experimentsc=7,γp=36
(the value is typically related to the window size). And then performing cost aggregation using equation (8).
Figure GDA0003001593790000035
Wherein the initial matching cost
Figure GDA0003001593790000036
As shown in formula (9):
Figure GDA0003001593790000037
Ic(q) and
Figure GDA0003001593790000038
for fixing reference image and image to be matchedAnd determining the gray values of two pixels with the parallax of d in the window. And finally, determining a final disparity map by a WTA (Winner-Takes-All) method.
The invention adopts an improved ASW matching method which mainly comprises the following steps: the method comprises a left reference image reading stage, a right reference image reading stage, a left initial matching cost stage, a right initial matching cost stage, a left cost function aggregation stage and a right cost function aggregation stage, and 4 parallax post-processing stages, wherein the 4 parallax post-processing stages mainly comprise LRC left and right consistency detection and filtering operation. Wherein:
firstly, the calculation of the initial matching cost is as follows:
the ASW algorithm uses the gray information of the image as a primitive for matching cost calculation. The method is improved by setting a truncation threshold value for the mean value of the gradient primitive and the R, G, B three-channel color primitive, fusing R, G, B color and gradient information of a pixel through the thought of Kalman filtering, and performing adaptive adjustment through a control coefficient alpha. The specific process is as follows:
(1) setting color and gradient thresholds t1 and t2 respectively, and calculating initial cost eg,ecAs shown in equations (10) (11).
Figure GDA0003001593790000041
Figure GDA0003001593790000042
(2) And (3) adaptively adjusting the alpha coefficient to calculate the final initial matching cost, wherein the final initial matching cost is shown in formula (12).
Figure GDA0003001593790000043
Secondly, an improved adaptive window expansion algorithm:
the traditional ASW matching algorithm has poor robustness in processing complex texture regions due to the fixed window. The invention adopts a self-adaptive window method according to the color and the space distance between pixels. And knowing a central pixel point p (x, y) to be matched, wherein the neighborhood pixel points in the x and y directions are p (x-1, y), p (x +1, y), p (x, y-1) and p (x, y + 1). Different from the traditional adaptive window expansion method which carries out pixel expansion by using the gray information of the central pixel point, the invention takes the central points R, G and B as expansion elements, and carries out window expansion when the three-channel information of the neighborhood pixels and the central pixel point simultaneously meets the condition of the following formula (13).
Ir,g,b(x,y)-Ir,g,b(x-1,y)<t (13)
t is a preset color threshold, and t ∈ (0, 1). When the same region pixel jumps due to discontinuous textures in the image, it is difficult to make three channels of pixel information of the neighboring pixels satisfy formula (13) at the same time. Based on this characteristic, the present invention makes an improvement over the conventional fixed window. When the neighborhood pixels are subjected to window adaptive expansion under the condition of satisfying the formula (13), if a texture repetition region exists in a scene, the window is too large, so that the calculation is too complex when the cost is aggregated, and the method does not meet the real-time requirement of the algorithm. The invention sets the truncation arm length for the adaptive window according to the geometric characteristics of the image. The window arm length is truncated when the following equation (14) is satisfied.
Figure GDA0003001593790000044
Wherein, p (x), p (y) are horizontal and vertical coordinate values of the central pixel, and q (x), q (y) are coordinate values of the adjacent pixels. The minimum arm length L is set through experiments on four test images of tsukuba, teddy, cons and venus under a midlebury platformminThreshold L5max=11。
Adaptive window, as shown in fig. 3. First, the central pixel I (x, y) is expanded laterally, and if the final window arm length is less than the shortest arm length, the minimum arm length L is usedminInstead of the original length. In the window expansion process, when the spatial distance between the neighborhood pixel and the central pixel is greater than a truncation threshold, the window is truncated. The final result is shown in FIG. 4, which is based on the center point, and has four arm lengths L with different lengthsu,Ld,Ll,Lr. The values of the row and column sizes of the self-adaptive window are respectively shown in the following formulas (15) and (16):
Rows=Lu+Ld (15)
Cows=Ll+Lr (16)
thirdly, matching cost aggregation algorithm:
after the initial matching cost is calculated, cost function aggregation is carried out, which is different from the complex weight operation of a classical algorithm. According to the method, the variance in the self-adaptive window of the reference image and the covariance of the reference image and the cost function are calculated through the correlation between the cost function and the reference image and the correlation of the reference image, so that the correlation function is obtained, and cost aggregation is performed after the correlation function is subtracted from the initial matching cost function, so that the accuracy and the real-time performance of the algorithm are greatly improved. The specific process is as follows:
first, the mean value of the convolution of the R, G, B three-channel pixel value of the pixel to be matched and the matching cost function within the matching window is calculated, as shown in equation (17).
Figure GDA0003001593790000051
Wherein, p is the pixel coordinate of the center point of the window, I (x, y) is the pixel coordinate of the center point in the window, I (x, y) is the self-adaptive window, and n is the total number of pixels in the window. After the mean of the convolution functions of the channels is calculated, the covariance function is calculated, as shown in equation (18):
Figure GDA0003001593790000052
in the above formula, mc,meThe average values of the three channel elements and the matching cost function in the adaptive window are respectively shown as the following formulas (19) and (20):
Figure GDA0003001593790000053
Figure GDA0003001593790000054
then, a variance matrix θ of the three channel element composition in the reference image adaptation window is calculated, which is described in detail as shown in the following equation (21).
Figure GDA0003001593790000055
Wherein, each element of the matrix theta is calculated as shown in formula (22):
Figure GDA0003001593790000061
the coefficient matrix obtained by equations (17) to (20) is shown in equation (23):
Figure GDA0003001593790000062
due to vce(c ∈ { r, g, b }) is a vector of 1 x 3, and γ can be obtainedknContains a vector of three channels. When the correlation coefficient gamma is calculatedknAnd then, subtracting the convolution of three-channel pixels of the reference image and the correlation coefficient from the mean value of each point cost function, so that the matching cost of the left image and the right image is more independent. Finally, the initial cost function is shown in equation (24) below:
Figure GDA0003001593790000063
due to gammaknIs a three channel vector, so n ∈ (1,2, 3). After the final matching cost is calculated, matching cost aggregation is performed through an adaptive window, and a specific formula is shown as a formula (25).
Figure GDA0003001593790000064
The invention only carries out aggregation on the basis of the reference image,
the image to be matched is not added. The experimental result shows that the accuracy of the algorithm is not reduced while the real-time performance of the algorithm is improved. And finally, selecting the parallax value with the minimum cost function as the pixel value of the parallax image by adopting a formula (26) through a WTA (Winner-Takes-A11) algorithm.
Figure GDA0003001593790000065
Four, parallax post-processing
1, LRC left-right consistency detection
In the stereo matching algorithm, the problem of occlusion is inevitable because of the parallax of the left and right images. Before the final disparity map is obtained, the invention firstly adopts an LRC left-right consistency algorithm to carry out disparity postprocessing operation.
When the left image is taken as a reference image, the parallax d is calculatedlObtaining the parallax d by using the right image as the reference imager. When the following condition of equation (27) is satisfied:
|dl-dr|>δ (27)
δ is a threshold, δ ∈ (0, 1). In the present invention, δ is 1. And when the absolute value of the left-right parallax difference is larger than delta, the occlusion point is considered. And taking the smaller parallax value in the left-right parallax of the occlusion point for parallax filling.
2, adaptive weighted median filtering
After the cost aggregation algorithm is carried out, the obtained disparity map has more salt-pepper noise, and the median filtering is necessary for the image. Conventional filtering, however, tends to ignore the correlation between pixels. The invention endows different weights to the pixels in the window based on the difference of the color and the distance between the pixels in the space, and the specific weight calculation is shown as a formula (28).
Figure GDA0003001593790000071
γc,γdIs constant, gamma is obtained by experimentc=0.1,γd9. k1, k2 is obtained by the difference between the central pixel and the surrounding pixel points in the color space and the distance space,
obtained from the following formulas (29) and (30), respectively.
Figure GDA0003001593790000072
Figure GDA0003001593790000073
The window size is 19 x 19. And after the weight value of each pixel in the window is obtained, the self-adaptive median filtering is carried out. The specific process is as follows:
(1) and multiplying the gray value of each pixel except the central point in the window by the respective weight value to obtain a new gray value, and calculating by using the formula (31).
I′(q)=w·I(q) (31)
(2) The new values for the pixels within the window, including the center point, are sorted by taking the 2 pixel values I' (q) nearest the center point near the median value1),I′(q2) And taking the average value to obtain a new gray value of the sub-pixel level to replace the gray value of the original central point pixel, and calculating by the formula (32).
Figure GDA0003001593790000074
Description of the Algorithm
The detailed implementation of the algorithm is described in table 1. TABLE 1
Figure GDA0003001593790000075
Sixth, experimental results
The experimental environment of the algorithm is Intel core i 7-67003.5 HZ CPU, 12G memory and Matlab2016a platform. In order to verify the effectiveness of the algorithm provided by the invention, the invention firstly compares a disparity map of a classical adaptive weight algorithm under a midlebury platform, for example, fig. 6 is a comparison view of the mismatching rates of all regions and depth discontinuity regions in a cons image mismatching rate comparison map.
The maximum and minimum arm lengths of the self-adaptive window in the experiment are respectively Lmin=5,LmaxThe minimum gradient and color threshold are t 1-2, t 2-7, and kalman coefficient α -0.11, respectively. Self-adaptive weight filtering weight coefficient gammac=0.1,γdThe window is 19 x 19, 9. The invention has three test indexes of the obtained disparity map under a midlebury platform: the nonocc (non-occlusion region mismatching rate), all (all region mismatching rate), disc (depth discontinuity region mismatching rate) test compares the conventional adaptive stereo matching algorithm and the De-Maeztu gradient-based ASW improved algorithm, and the specific ratio is shown in table 2 below.
TABLE 2 evaluation of matching algorithm Table (unit:%)
Figure GDA0003001593790000081
From table 1, it can be seen that compared with the classical adaptive weight and its improved algorithm, the stereo matching algorithm based on multi-matching primitive fusion has obvious advantages in Venus, Teddy, cons images, no matter in all regions, non-occlusion regions or depth discontinuity region scenes. Under the Tsukuba image with darker light, a certain promotion is also provided in the non-blocking area. Compared with the GradAdaptWgt algorithm based on single gradient primitives, Tsukuba, Venus and Teddy have lower mismatching rate in all regions, and the mismatching rate of the Conses image is equivalent to that of the original GradAdaptWgt algorithm. The algorithm of the invention performs average mismatching rate comparison under middleb ury platform, as shown in table 3, and obtains considerable effect.
TABLE 3 average mismatch ratio comparison
Figure GDA0003001593790000082
In order to improve the robustness of the algorithm, the Kalman coefficient ratio of the color and gradient information is compared. Selecting the Cones images conforming to the natural scene for comparison, and comparing mismatching rates of matching conditions of noccc, all and disc under different alpha coefficients, wherein alpha belongs to (0.11: 0.02: 0.29).
From the above results, we set the value of α to 0.23 in practical applications to achieve the robustness requirement of the stereo matching algorithm. Compared with the classical self-adaptive support weight algorithm, the method abandons the operation of complex weight and greatly improves the real-time property of the algorithm. The invention contrasts with the classic adaptive support weight algorithm in time complexity as shown in table 4.
TABLE 4 comparison of time complexity (unit: ms) between the algorithm of the present invention and the classical ASW algorithm
Figure GDA0003001593790000091
The experimental results show that the algorithm greatly reduces the time complexity of the algorithm.
The invention has the advantages that:
the invention provides a novel stereo matching method based on multiple matching primitives by fusing color matching primitives and gradient matching primitives and introducing the correlation between a cost function and a reference image. The algorithm greatly improves the accuracy of scene information under the unmanned condition, contributes to obtaining a more accurate depth map, greatly reduces the time complexity of the algorithm, and provides a theoretical basis for the practical application of unmanned driving. And the navigation precision of the robot or the unmanned aerial vehicle is further improved.
Drawings
Fig. 1 is left and right images acquired by the zed binocular camera after calibration and image setting.
Fig. 2 is a flow chart of the improved ASW algorithm.
Fig. 3 is a lateral expansion diagram in an adaptive window map.
Fig. 4 is a diagram of the window expansion final effect in the adaptive window map.
FIG. 5 is a graph of the Cones image mismatch rate versus the non-occluded area mismatch rate in the plot.
Fig. 6 is a cross-sectional view of the cross-sectional area and depth discontinuity area cross-matching rate in the cons image cross-matching rate comparison map.
Detailed Description
A binocular vision stereo matching method based on multi-matching element fusion comprises a computer, two cameras and the stereo matching method based on multi-matching element fusion, and is characterized in that:
1. the camera is a ZED binocular camera manufactured by Astrolabes company.
The resolution of the video stream collected by the ZED binocular camera is divided into four categories,
the specific parameters are shown in the following table 1:
table 1: ZED output video stream mode
Figure GDA0003001593790000092
Figure GDA0003001593790000101
2. Image acquisition:
the ZED SDK and the CUDA are downloaded and connected with a computer through a USB3.0 interface. Connecting a ZED binocular camera through a webcam function in an MATLAB, and acquiring pictures through a snapshot function;
3. calibrating a camera:
the aim of camera calibration is to obtain accurate internal and external parameters of a camera. The internal parameters mainly comprise the focal lengths of the left lens and the right lens and the baseline distance; the extrinsic parameters mainly include a rotation matrix of the two cameras relative to a world coordinate system and a relative translation matrix of the left camera and the right camera. The invention obtains the default internal and external parameters of the camera from the official manual;
4. setting image parameters:
and performing epipolar correction through the calibration parameters to enable the acquired left and right images to meet epipolar constraint conditions. Resetting parameters by embedding a ZED Explorer plug-in into the ZED SDK;
5. stereo matching:
stereo matching is a core part of the binocular vision system. The purpose of stereo matching is to perform imaging point matching on the acquired left and right images, obtain a parallax value through the matching points, and obtain depth information of a scene.
The ZED binocular camera can be installed on a robot or an unmanned plane. The actual scene shot by the two eyes can achieve the purpose of a real scene through the three-dimensional matching processing of the fusion of the multiple matching elements, and then a navigation instruction is sent to a robot or unmanned aerial vehicle control and driving system through the processing of a computer arranged on the robot or the unmanned aerial vehicle.
The stereo matching method of the multi-matching primitive fusion comprises the following processes:
with the improved invention, an improved ASW matching method is adopted:
the improved ASW algorithm flow provided by the invention mainly comprises the following steps: the method comprises a left reference image reading stage, a right reference image reading stage, a left initial matching cost stage, a right initial matching cost stage, a left cost function aggregation stage and a right cost function aggregation stage, and 4 parallax post-processing stages, wherein the 4 parallax post-processing stages mainly comprise LRC left and right consistency detection and filtering operation. Wherein:
firstly, the calculation of the initial matching cost is as follows:
the ASW algorithm uses the gray information of the image as a primitive for matching cost calculation. The method is improved by setting a truncation threshold value for the mean value of the gradient primitive and the R, G, B three-channel color primitive, fusing R, G, B color and gradient information of a pixel through the thought of Kalman filtering, and performing adaptive adjustment through a control coefficient alpha. The specific process is as follows:
(1) setting color and gradient thresholds t1 and t2 respectively, and calculating initial cost eg,ecAs shown in equations (10) (11).
Figure GDA0003001593790000102
Figure GDA0003001593790000103
(2) And (3) adaptively adjusting the alpha coefficient to calculate the final initial matching cost, wherein the final initial matching cost is shown in formula (12).
Figure GDA0003001593790000104
Secondly, an improved adaptive window expansion algorithm: the traditional ASW matching algorithm has poor robustness in processing complex texture regions due to the fixed window. The invention adopts a self-adaptive window method according to the color and the space distance between pixels. And knowing a central pixel point p (x, y) to be matched, wherein the neighborhood pixel points in the x and y directions are p (x-1, y), p (x +1, y), p (x, y-1) and p (x, y + 1). Different from the traditional adaptive window expansion method which carries out pixel expansion by using the gray information of the central pixel point, the invention takes the central points R, G and B as expansion elements, and carries out window expansion when the three-channel information of the neighborhood pixels and the central pixel point simultaneously meets the condition of the following formula (13).
Ir,g,b(x,y)-Ir,g,b(x-1,y)<t (13)
t is a preset color threshold, and t ∈ (0, 1). When the same region pixel jumps due to discontinuous textures in the image, it is difficult to make three channels of pixel information of the neighboring pixels satisfy formula (13) at the same time. Based on this characteristic, the present invention makes an improvement over the conventional fixed window. When the neighborhood pixels are subjected to window adaptive expansion under the condition of satisfying the formula (13), if a texture repetition region exists in a scene, the window is too large, so that the calculation is too complex when the cost is aggregated, and the method does not meet the real-time requirement of the algorithm. The invention sets the truncation arm length for the adaptive window according to the geometric characteristics of the image. The window arm length is truncated when the following equation (14) is satisfied.
Figure GDA0003001593790000111
Wherein, p (x), p (y) are horizontal and vertical coordinate values of the central pixel, and q (x), q (y) are coordinate values of the adjacent pixels. The minimum arm length L is set through experiments on four test images of tsukuba, teddy, cons and venus under a midlebury platformminThreshold L5max=11。
Adaptive window, as shown in fig. 3. First, the central pixel I (x, y) is expanded laterally, and if the final window arm length is less than the shortest arm length, the minimum arm length L is usedminInstead of the original length. In the window expansion process, when the spatial distance between the neighborhood pixel and the central pixel is greater than a truncation threshold, the window is truncated. The final result is shown in FIG. 4, which is based on the center point, and has four arm lengths L with different lengthsu,Ld,Ll,Lr. The values of the row and column sizes of the self-adaptive window are respectively shown in the following formulas (15) and (16):
Rows=Lu+Ld (15)
Cows=Ll+Lr (16)
thirdly, matching cost aggregation algorithm:
after the initial matching cost is calculated, cost function aggregation is carried out, which is different from the complex weight operation of a classical algorithm. According to the method, through the correlation between the cost function and the reference image and the correlation between the reference image and the self-adaptive window of the reference image in the document [11], the variance in the adaptive window of the reference image and the covariance between the reference image and the cost function are calculated to obtain the correlation function of the reference image, and cost aggregation is performed after the correlation function is subtracted from the initial matching cost function, so that the accuracy and the real-time performance of the algorithm are greatly improved. The specific process is as follows:
first, the mean value of the convolution of the R, G, B three-channel pixel value of the pixel to be matched and the matching cost function within the matching window is calculated, as shown in equation (17).
Figure GDA0003001593790000112
Wherein p is the pixel coordinate of the center point of the window, q is the pixel coordinate of the center point in the window, and NpFor an adaptive window, n is the total number of pixels in the window. After the mean of the convolution functions of the channels is calculated, the covariance function is calculated, as shown in equation (18):
Figure GDA0003001593790000121
in the above formula, mc,meThe average values of the three channel elements and the matching cost function in the adaptive window are respectively shown as the following formulas (19) and (20):
Figure GDA0003001593790000122
Figure GDA0003001593790000123
then, a variance matrix θ of the three channel element composition in the reference image adaptation window is calculated, which is described in detail as shown in the following equation (21).
Figure GDA0003001593790000124
Wherein, each element of the matrix theta is calculated as shown in formula (22):
Figure GDA0003001593790000125
the coefficient matrix obtained by equations (17) to (20) is shown in equation (23):
Figure GDA0003001593790000126
due to vce(c ∈ { r, g, b }) is a vector of 1 x 3, and γ can be obtainedknContains a vector of three channels. When the correlation coefficient gamma is calculatedknAnd then, subtracting the convolution of three-channel pixels of the reference image and the correlation coefficient from the mean value of each point cost function, so that the matching cost of the left image and the right image is more independent. Finally, the initial cost function is shown in equation (24) below:
Figure GDA0003001593790000127
due to gammaknIs a three channel vector, so n ∈ (1,2, 3). After the final matching cost is calculated, matching cost aggregation is performed through an adaptive window, and a specific formula is shown as a formula (25).
Figure GDA0003001593790000128
The invention only carries out aggregation on the basis of the reference image,
the image to be matched is not added. The experimental result shows that the accuracy of the algorithm is not reduced while the real-time performance of the algorithm is improved. And finally, selecting the parallax value with the minimum cost function as the pixel value of the parallax image by adopting a formula (26) through a WTA (Winner-Takes-All) algorithm.
Figure GDA0003001593790000131
Four, parallax post-processing
1, LRC left-right consistency detection
In the stereo matching algorithm, the problem of occlusion is inevitable because of the parallax of the left and right images. Before the final disparity map is obtained, the invention firstly adopts an LRC left-right consistency algorithm to carry out disparity postprocessing operation.
When the left image is taken as a reference image, the parallax d is calculatedlObtaining the parallax d by using the right image as the reference imager. When the following are satisfiedThe conditions of equation (27):
|dl-dr|>δ (27)
δ is a threshold, δ ∈ (0, 1). In the present invention, δ is 1. And when the absolute value of the left-right parallax difference is larger than delta, the occlusion point is considered. And taking the smaller parallax value in the left-right parallax of the occlusion point for parallax filling.
2, adaptive weighted median filtering
After the cost aggregation algorithm is carried out, the obtained disparity map has more salt-pepper noise, and the median filtering is necessary for the image. Conventional filtering, however, tends to ignore the correlation between pixels. The invention endows different weights to the pixels in the window based on the difference of the color and the distance between the pixels in the space, and the specific weight calculation is shown as a formula (28).
Figure GDA0003001593790000132
γc,γdIs constant, gamma is obtained by experimentc=0.1,γd9. k1, k2 is obtained by the difference between the central pixel and the surrounding pixel points in the color space and the distance space,
obtained from the following formulas (29) and (30), respectively.
Figure GDA0003001593790000133
Figure GDA0003001593790000134
The window size is 19 x 19. And after the weight value of each pixel in the window is obtained, the self-adaptive median filtering is carried out. The specific process is as follows:
(1) and multiplying the gray value of each pixel except the central point in the window by the respective weight value to obtain a new gray value, and calculating by using the formula (31).
I′(q)=w·I(q) (31)
(2) The new values for the pixels within the window, including the center point, are sorted by taking the 2 pixel values I' (q) nearest the center point near the median value1),I′(q2) And taking the average value to obtain a new gray value of the sub-pixel level to replace the gray value of the original central point pixel, and calculating by the formula (32).
Figure GDA0003001593790000135

Claims (1)

1. A binocular vision stereo matching method based on multi-matching element fusion comprises a computer, two cameras and the stereo matching method based on multi-matching element fusion, and is characterized in that:
the camera selects a ZED binocular camera and is set as follows:
(1) and image acquisition:
the ZED SDK and the CUDA are downloaded and connected with a computer through a USB3.0 interface; connecting a ZED binocular camera through a webcam function in an MATLAB, and acquiring pictures through a snapshot function;
(2) calibrating a camera:
the aim of camera calibration is to obtain accurate internal and external parameters of a camera; the internal parameters mainly comprise the focal lengths of the left lens and the right lens and the baseline distance; the external parameters mainly comprise a rotation matrix of the two cameras relative to a world coordinate system and a relative translation matrix of the left camera and the right camera; acquiring default internal and external parameters of the camera;
(3) and image parameter setting:
performing epipolar line correction through calibration parameters so that the acquired left and right images meet epipolar constraint conditions; resetting parameters by embedding a ZED Explorer plug-in into the ZED SDK;
(4) and stereo matching:
stereo matching is used as a core part of a binocular vision system; the purpose of stereo matching is to perform imaging point matching on the acquired left and right images, obtain a parallax value through the matching points and obtain depth information of a scene;
the ZED binocular camera can be arranged on a robot or an unmanned aerial vehicle, the actual scene shot by the two eyes can achieve the purpose of a real scene through the stereo matching processing of the fusion of multiple matching primitives, and then a navigation instruction is sent to a robot or unmanned aerial vehicle control and driving system through the processing of a computer arranged on the robot or the unmanned aerial vehicle;
the stereo matching method of the multi-matching primitive fusion comprises the following processes:
an improved ASW matching method is adopted, and the method mainly comprises a left reference image reading-in stage, a left initial matching cost stage, a right initial matching cost stage, a left cost function aggregation stage and a right cost function aggregation stage, and 4 stages of parallax post-processing; the parallax post-processing stage mainly comprises LRC left-right consistency detection and filtering operation, wherein:
1) the initial matching cost calculation is:
the ASW algorithm utilizes the gray information of the image as a primitive of matching cost calculation; the method is improved by setting a truncation threshold value for the mean value of a gradient element and an R, G, B three-channel color element, fusing R, G, B color and gradient information of a pixel through the idea of Kalman filtering, and performing adaptive adjustment through a control coefficient alpha; the specific process is as follows:
(1) setting color and gradient thresholds t1 and t2 respectively, and calculating initial cost
Figure 963207DEST_PATH_IMAGE002
As shown in equations (10) (11);
Figure 818031DEST_PATH_IMAGE004
Figure 456823DEST_PATH_IMAGE006
(2) adaptively adjusting the alpha coefficient to calculate the final initial matching cost, wherein the final initial matching cost is shown in a formula (12);
Figure 773403DEST_PATH_IMAGE008
2) the improved adaptive window expansion algorithm comprises the following steps:
according to the color and the space distance between pixels, an adaptive window method is adopted, the central pixel point p (x, y) to be matched is known, and the neighborhood pixel points in the x and y directions are p (x-1, y), p (x +1, y), p (x, y-1) and p (x, y +1) respectively; the method is different from the traditional self-adaptive window expansion method which carries out pixel expansion by using the gray information of the central pixel point; taking the central points R, G and B as extension elements, and carrying out window expansion when three-channel information of the neighborhood pixels and the central pixel simultaneously meets the condition of the following formula (13);
Figure 525459DEST_PATH_IMAGE010
t is a preset color threshold, and t belongs to (0, 1); when the same region pixel jumps due to discontinuous textures in the image, three-channel pixel information of adjacent pixels hardly satisfies formula (13); based on the characteristic, the traditional fixed window is improved; when the neighborhood pixels are subjected to window self-adaptive expansion under the condition of satisfying the formula (13), if a texture repetition region exists in a scene, the window is too large, so that the calculation is too complex when the cost is aggregated, and the real-time requirement of the algorithm is not met; setting a truncation arm length for the self-adaptive window according to the geometric characteristics of the image; truncating the window arm length when the following formula (14) is satisfied;
Figure 537802DEST_PATH_IMAGE012
wherein, p (x), p (y) are horizontal and vertical coordinate values of the central pixel, q (x), q (y) are coordinate values of the neighborhood pixels; the minimum arm length L is set through experiments on four test images of tsukuba, teddy, cons and venus under a midlebury platformminThreshold L5max11; the adaptive window is first adapted by, for a central pixel I (x,y) performing transverse expansion, and if the final window arm length is less than the shortest arm length, using the minimum arm length LminReplacing the original length; in the window expansion process, when the space distance between the neighborhood pixel and the central pixel is greater than a truncation threshold, the window is truncated; four arms with different lengths L at upper, lower, left and right sides by taking a central point as referenceu,Ld,Ll,Lr(ii) a The values of the row and column sizes of the adaptive window are respectively shown in the following formulas (15) and (16):
Rows=Lu+Ld (15)
Cows=Ll+Lr (16)
3) and matching cost aggregation algorithm:
calculating the mean value of the convolution of the R, G, B three-channel pixel value of the pixel to be matched and the matching cost function in the matching window, as shown in the formula (17);
Figure 742518DEST_PATH_IMAGE014
wherein, p is the pixel coordinate of the center point of the window, q is the pixel coordinate of the center point in the window, Np is the self-adaptive window, and n is the total number of pixels in the window; after the mean of the convolution functions of each channel is calculated,
the covariance function is calculated as shown in equation (18):
Figure 128369DEST_PATH_IMAGE016
in the above formula, mc,meThe average values of the three channel elements and the matching cost function in the adaptive window are respectively shown as the following formulas (19) and (20):
Figure 734931DEST_PATH_IMAGE018
then, calculating a variance matrix theta formed by three channel elements in a reference image adaptive window, wherein the variance matrix theta is specifically described as shown in a formula (21);
Figure 915245DEST_PATH_IMAGE020
wherein, each element of the matrix theta is calculated as shown in formula (22):
Figure 403995DEST_PATH_IMAGE022
the coefficient matrix obtained by equations (17) to (20) is shown in equation (23):
Figure 531220DEST_PATH_IMAGE024
due to vce(c ∈ { r, g, b }) is a vector of 1 x 3, and γ can be obtainedknContains a vector of three channels; when the correlation coefficient gamma is calculatedknThen, the convolution of three-channel pixels of the reference image and the correlation coefficient is subtracted from the average value of each point cost function, so that the matching cost of the left image and the right image is more independent; finally, the initial cost function is shown in equation (24) below:
Figure 789026DEST_PATH_IMAGE026
due to gammaknIs a three channel vector, so n ∈ (1,2, 3); after the final matching cost is calculated, performing matching cost aggregation through a self-adaptive window, wherein a specific formula is shown as a formula (25);
Figure 606154DEST_PATH_IMAGE028
only the reference image is aggregated, and the image to be matched is not added; the experimental result shows that the accuracy of the algorithm is not reduced while the real-time performance of the algorithm is improved; finally, selecting the parallax value with the minimum cost function as the pixel value of the parallax image by adopting a formula (26) through a WTA (Winner-Takes-All) algorithm;
Figure 785462DEST_PATH_IMAGE030
4) parallax post-processing
(1) LRC left and right consistency detection
In the stereo matching algorithm, the problem of occlusion is inevitable all the time because of the parallax of the left and right images; before obtaining a final disparity map, performing disparity postprocessing operation by adopting an LRC left-right consistency algorithm;
when the left image is taken as a reference image, the parallax d is calculatedlObtaining a parallax dr by taking the right image as a reference image; when the following condition of equation (27) is satisfied:
Figure 247537DEST_PATH_IMAGE032
delta is a threshold value, delta belongs to (0, 1); herein δ takes the value 1; when the absolute value of the left-right parallax difference is larger than delta, the left-right parallax difference is considered as a shielding point; selecting a smaller parallax value in left-right parallax from the shielding points to perform parallax filling;
(2) adaptive weighted median filtering
After the cost aggregation algorithm is carried out, the obtained disparity map usually has more salt and pepper noises, and the median filtering is necessary to be carried out on the image; however, conventional filtering tends to ignore the correlation between pixels; different weights are given to the pixels in the window based on the difference between the colors and the distances of the pixels in the space, and the specific weight calculation is shown as a formula (28);
Figure 625428DEST_PATH_IMAGE034
γc,γdis constant, gamma is obtained by experimentc=0.1,γd9; k1, k2 are derived from the difference between the central pixel and the surrounding pixel points in color space and distance space, and are obtained by the following formulas (29) and (30), respectively;
Figure 350808DEST_PATH_IMAGE036
window size 19 x 19; after obtaining the weight of each pixel in the window, the self-adaptive median filtering is performed, and the specific process is as follows:
(1) multiplying the gray value of each pixel except the central point in the window by the respective weight value to obtain a new gray value, and calculating by using the formula (31);
Figure 814150DEST_PATH_IMAGE038
(2) sorting the new values of the pixels including the central point in the window, taking the 2 pixel values I '(q 1) and I' (q2) which are positioned near the median and are closest to the central point, taking the average value to obtain the new gray value of the sub-pixel level to replace the gray value of the pixel at the original central point, and obtaining the gray value by the calculation of the formula (32)
Figure 283178DEST_PATH_IMAGE040
CN201810249294.XA 2018-03-22 2018-03-22 Binocular vision stereo matching method based on multi-matching element fusion Active CN108682026B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810249294.XA CN108682026B (en) 2018-03-22 2018-03-22 Binocular vision stereo matching method based on multi-matching element fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810249294.XA CN108682026B (en) 2018-03-22 2018-03-22 Binocular vision stereo matching method based on multi-matching element fusion

Publications (2)

Publication Number Publication Date
CN108682026A CN108682026A (en) 2018-10-19
CN108682026B true CN108682026B (en) 2021-08-06

Family

ID=63800451

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810249294.XA Active CN108682026B (en) 2018-03-22 2018-03-22 Binocular vision stereo matching method based on multi-matching element fusion

Country Status (1)

Country Link
CN (1) CN108682026B (en)

Families Citing this family (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109544611B (en) * 2018-11-06 2021-05-14 深圳市爱培科技术股份有限公司 Binocular vision stereo matching method and system based on bit characteristics
CN109741385A (en) * 2018-12-24 2019-05-10 浙江大华技术股份有限公司 A kind of image processing system, method, apparatus, electronic equipment and storage medium
CN109978934B (en) * 2019-03-04 2023-01-10 北京大学深圳研究生院 Binocular vision stereo matching method and system based on matching cost weighting
WO2020177060A1 (en) * 2019-03-04 2020-09-10 北京大学深圳研究生院 Binocular visual stereoscopic matching method based on extreme value checking and weighted voting
WO2020177061A1 (en) * 2019-03-04 2020-09-10 北京大学深圳研究生院 Binocular stereo vision matching method and system based on extremum verification
CN109978928B (en) * 2019-03-04 2022-11-04 北京大学深圳研究生院 Binocular vision stereo matching method and system based on weighted voting
CN110148168B (en) * 2019-05-23 2023-03-24 南京大学 Three-eye camera depth image processing method based on size double baselines
CN112802114B (en) * 2019-11-13 2024-07-02 浙江舜宇智能光学技术有限公司 Multi-vision sensor fusion device, method thereof and electronic equipment
CN111553296B (en) * 2020-04-30 2021-08-03 中山大学 Two-value neural network stereo vision matching method based on FPGA
CN111754588B (en) * 2020-06-30 2024-03-29 江南大学 Binocular vision matching method for self-adaptive window size based on variance
CN112184833B (en) * 2020-09-29 2023-09-15 南京蕴智科技有限公司 Hardware implementation system and method of binocular stereo matching algorithm
CN112200852B (en) * 2020-10-09 2022-05-20 西安交通大学 Stereo matching method and system for space-time hybrid modulation
CN112308897B (en) * 2020-10-30 2024-07-12 江苏大学 Stereo matching method based on neighborhood information constraint and self-adaptive window
CN112348871B (en) * 2020-11-16 2023-02-10 长安大学 Local stereo matching method
CN113052862B (en) * 2021-04-12 2024-06-18 北京机械设备研究所 Multi-level optimization-based stereo matching method, device and equipment in outdoor scene
CN113516699A (en) * 2021-05-18 2021-10-19 哈尔滨理工大学 Stereo matching system based on super-pixel segmentation
CN113933315B (en) * 2021-10-13 2024-04-05 深圳市中纬智能有限公司 Method and system for detecting collinearity of chip circuit board
CN116188558B (en) * 2023-04-27 2023-07-11 华北理工大学 Stereo photogrammetry method based on binocular vision
CN116258759B (en) * 2023-05-15 2023-09-22 北京爱芯科技有限公司 Stereo matching method, device and equipment

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102611904A (en) * 2012-02-15 2012-07-25 山东大学 Stereo matching method based on image partitioning in three-dimensional television system
CN103985128A (en) * 2014-05-23 2014-08-13 南京理工大学 Three-dimensional matching method based on color intercorrelation and self-adaptive supporting weight
CN103996202A (en) * 2014-06-11 2014-08-20 北京航空航天大学 Stereo matching method based on hybrid matching cost and adaptive window
CN103996201A (en) * 2014-06-11 2014-08-20 北京航空航天大学 Stereo matching method based on improved gradient and adaptive window
CN104867135A (en) * 2015-05-04 2015-08-26 中国科学院上海微系统与信息技术研究所 High-precision stereo matching method based on guiding image guidance
CN106355570A (en) * 2016-10-21 2017-01-25 昆明理工大学 Binocular stereoscopic vision matching method combining depth characteristics
CN107301664A (en) * 2017-05-25 2017-10-27 天津大学 Improvement sectional perspective matching process based on similarity measure function
CN107301642A (en) * 2017-06-01 2017-10-27 中国人民解放军国防科学技术大学 A kind of full-automatic prospect background segregation method based on binocular vision
CN107392950A (en) * 2017-07-28 2017-11-24 哈尔滨理工大学 A kind of across yardstick cost polymerization solid matching method based on weak skin texture detection

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102214934B1 (en) * 2014-07-18 2021-02-10 삼성전자주식회사 Stereo matching apparatus and method using unary confidences learning and pairwise confidences learning
KR101622344B1 (en) * 2014-12-16 2016-05-19 경북대학교 산학협력단 A disparity caculation method based on optimized census transform stereo matching with adaptive support weight method and system thereof

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102611904A (en) * 2012-02-15 2012-07-25 山东大学 Stereo matching method based on image partitioning in three-dimensional television system
CN103985128A (en) * 2014-05-23 2014-08-13 南京理工大学 Three-dimensional matching method based on color intercorrelation and self-adaptive supporting weight
CN103996202A (en) * 2014-06-11 2014-08-20 北京航空航天大学 Stereo matching method based on hybrid matching cost and adaptive window
CN103996201A (en) * 2014-06-11 2014-08-20 北京航空航天大学 Stereo matching method based on improved gradient and adaptive window
CN104867135A (en) * 2015-05-04 2015-08-26 中国科学院上海微系统与信息技术研究所 High-precision stereo matching method based on guiding image guidance
CN106355570A (en) * 2016-10-21 2017-01-25 昆明理工大学 Binocular stereoscopic vision matching method combining depth characteristics
CN107301664A (en) * 2017-05-25 2017-10-27 天津大学 Improvement sectional perspective matching process based on similarity measure function
CN107301642A (en) * 2017-06-01 2017-10-27 中国人民解放军国防科学技术大学 A kind of full-automatic prospect background segregation method based on binocular vision
CN107392950A (en) * 2017-07-28 2017-11-24 哈尔滨理工大学 A kind of across yardstick cost polymerization solid matching method based on weak skin texture detection

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
A New Stereo Matching Algorithm Based on Adaptive Window;Yingyun Yang等;《2012 International Conference on Systems and Informatics (ICSAI 2012)》;20120625;第1815-1819页 *
Region Based Stereo Matching Method with Gradient and Distance Information;Yong-Jun Chang等;《2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)》;20170119;第1-4页 *
双目视觉立体匹配技术研究现状和展望;林森等;《科学技术与工程》;20171031;第17卷(第30期);第135-147页 *
多种相似性测度结合的局部立体匹配算法;卢迪等;《机器人》;20160131;第38卷(第1期);第1-7页 *

Also Published As

Publication number Publication date
CN108682026A (en) 2018-10-19

Similar Documents

Publication Publication Date Title
CN108682026B (en) Binocular vision stereo matching method based on multi-matching element fusion
CN110569704B (en) Multi-strategy self-adaptive lane line detection method based on stereoscopic vision
CN105374019B (en) A kind of more depth map fusion methods and device
CN111066065B (en) System and method for hybrid depth regularization
CN106408513B (en) Depth map super resolution ratio reconstruction method
CN108596975B (en) Stereo matching algorithm for weak texture region
CN110148181A (en) A kind of general binocular solid matching process
CN102665086B (en) Method for obtaining parallax by using region-based local stereo matching
CN106780590A (en) The acquisition methods and system of a kind of depth map
CN109961417B (en) Image processing method, image processing apparatus, and mobile apparatus control method
CN103426200B (en) Tree three-dimensional reconstruction method based on unmanned aerial vehicle aerial photo sequence image
CN111027415B (en) Vehicle detection method based on polarization image
CN111553862B (en) Defogging and binocular stereoscopic vision positioning method for sea and sky background image
CN110853151A (en) Three-dimensional point set recovery method based on video
CN104504688A (en) Method and system based on binocular stereoscopic vision for passenger flow density estimation
CN102903098A (en) Depth estimation method based on image definition difference
CN110349249B (en) Real-time dense reconstruction method and system based on RGB-D data
CN111354077B (en) Binocular vision-based three-dimensional face reconstruction method
CN109859249B (en) Scene flow estimation method based on automatic layering in RGBD sequence
CN108460792B (en) Efficient focusing stereo matching method based on image segmentation
CN106530336B (en) Stereo matching method based on color information and graph cut theory
CN106023189B (en) A kind of light field data depth reconstruction method based on matching optimization
CN114996814A (en) Furniture design system based on deep learning and three-dimensional reconstruction
CN104038752B (en) Multi-view point video rectangular histogram color correction based on three-dimensional Gaussian mixed model
CN108742495A (en) A kind of three-dimensional stereo laparoscope system and solid matching method for medical field

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20210708

Address after: 230031 102-f, Hefei R & D building, Chinese Academy of Sciences, northwest corner of intersection of Xiyou road and Shilian South Road, high tech Zone, Hefei City, Anhui Province

Applicant after: Hefei Jinren Technology Co.,Ltd.

Address before: 121000 Shiying street, Guta District, Jinzhou City, Liaoning Province

Applicant before: LIAONING University OF TECHNOLOGY

TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20210722

Address after: 230000, 12 / F, building e, intelligent technology park, economic development zone, Hefei, Anhui Province

Applicant after: Jiang Dabai

Address before: 230031 102-f, Hefei R & D building, Chinese Academy of Sciences, northwest corner of intersection of Xiyou road and Shilian South Road, high tech Zone, Hefei City, Anhui Province

Applicant before: Hefei Jinren Technology Co.,Ltd.

GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20221205

Address after: 230000, 12 / F, building e, intelligent technology park, economic development zone, Hefei, Anhui Province

Patentee after: CHINA APPLIED TECHNOLOGY Co.,Ltd.

Address before: 230000, 12 / F, building e, intelligent technology park, economic development zone, Hefei, Anhui Province

Patentee before: Jiang Dabai