CN105966314A - Lane departure pre-warning method based on double low-cost cameras - Google Patents
Lane departure pre-warning method based on double low-cost cameras Download PDFInfo
- Publication number
- CN105966314A CN105966314A CN201610423005.4A CN201610423005A CN105966314A CN 105966314 A CN105966314 A CN 105966314A CN 201610423005 A CN201610423005 A CN 201610423005A CN 105966314 A CN105966314 A CN 105966314A
- Authority
- CN
- China
- Prior art keywords
- image
- lane line
- point
- lane
- avg
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/14—Means for informing the driver, warning the driver or prompting a driver intervention
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/181—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/10—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used
- B60R2300/105—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used using multiple cameras
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/80—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
- B60R2300/804—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for lane monitoring
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/14—Means for informing the driver, warning the driver or prompting a driver intervention
- B60W2050/146—Display means
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Mechanical Engineering (AREA)
- Automation & Control Theory (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Transportation (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention belongs to the field of aided driving, and relates to a lane departure pre-warning method based on double low-cost cameras. A lane departure pre-warning algorithm is achieved by using the double low-cost cameras. The method includes the steps that real-time images are obtained through the cameras below a left vehicle lug and a right vehicle lug of an intelligent vehicle; pretreatment, IPM and feature point extraction are carried out on the two images; feature points of candidate lane lines are clustered to obtain feature points of lane lines; Hoff straight line conversion detection is carried out on the images; the rightmost lane line of the left camera and the leftmost lane line of the right camera are extracted to calculate departure distance. The departure distance is calculated based on a model set up on the IPM images obtained after image inverse perspective conversion, and the problem that detection of the camera at one side fails is effectively solved. The double low-cost cameras are used for replacing a traditional single high-definition industrial camera, cost is greatly reduced, and the algorithm is low in operation cost and high in accuracy and real-time performance.
Description
Technical field
The present invention is lane departure warning methods based on double low cost photographic head, belongs to intelligence auxiliary driving technology neck
Territory.
Background technology
Lane departure warning is a critical aspects of intelligence DAS (Driver Assistant System) research, along with safe driving, intelligence are handed over
Logical increasingly by people's concern and attention, lane departure warning method research has become as the hot issue of research.Deviation
The design of method for early warning is according to road ahead environment and this truck position relation, it is judged that automotive run-off-road distance is also carried out in time
Remind, prevent owing to driver neglects the generation of the Lane Departure caused.Show according to research, lane departure warning method
Realization and application can avoid the generation of deviation vehicle accident of 30-70%, to improving traffic safety, reduce and hand over
Interpreter thus cause be significant.Its design head-up position under monocular-camera of traditional lane departure warning method enters
OK, including image capture module, lane detection module and warning module.Typically first pass through imageing sensor and obtain vehicle row
Road image real-time during sailing, then the run-off-road line that lane detection is obtained vehicle on hardware processing platform
Whether distance parameters etc., then have the probability that deviation occurs to be estimated by method for early warning to current running state,
Finally show early warning information.
Summary of the invention
The present invention utilizes the deviation that the photographic head of two low costs immediately below vehicle car ear (rearview mirror) is carried out
Method for early warning, it is possible to lane departure warning information is provided correctly, in real time.Its advantage be the cost requirement to photographic head very
Low, and ensure that computing overhead minimum and real-time.The design of algorithm is photographic head based on double low costs, it is possible to adapt to car
Produce family to cost, real-time and requirement accurately, market prospect will be the best.
To achieve these goals, this invention takes following technical scheme:
Step 1: sensor is installed;
Video camera is separately mounted to the underface of intelligent vehicle sided mirror unit parallel with the longitudinal coordinate axle of car body,
When video camera is installed, should ensure that two video cameras must be positioned in same plane, and vehicle both sides lane line image can be collected;
Step 2: camera calibration;
Demarcating two photographic head respectively, the field range of demarcation is: wide 600cm, remote 1000cm.Adjust photographic head
Left and right, upper lower angle, after enabling demarcation, the position of left and right photographic head (to require that the picture obtained is same in the same plane
Process in plane);
Step 3: the pretreatment of image;
According to lane line feature, first image being carried out gray proces, the formula of gray proces is
Gray=R*0.5+G*0.5, wherein R, G represent red blue channel component value, G respectivelyrayRepresent the pixel after conversion
Gray value.Then image is carried out medium filtering, this experiment uses the square field of 3 × 3 be filtered image processing.
Step 4:IPM;
Gray level image carrying out inverse perspective process and obtains birds-eye view picture, perspective matrix is calibrated H-matrix.
Step 5: feature point extraction;
The gray value of lane line is higher than the value on its both sides, forms a crest;Present and from left to right fall after rising
Trend;When in lane line region, average is higher, average differs bigger with its summit value.We utilize these characteristics, by calculating
The change of adjacent image pixels judges the edge of lane line.
Step 5-1: calculate the average of adjacent image pixels;
If certain point be (x, y), meet y ∈ [0, h) and x ∈ [2, w-2).X, y are the columns and rows of pixel respectively, and w is figure
The width of picture, h is the height of image.The average then having an adjacent image pixels is:
Wherein t ∈ [1,3,5,7 ...], t=5 can obtain good effect.
Step 5-2: calculating edge extracting threshold value T, its computing formula is expressed as follows:
Wherein avg(x,y)For point (x, y) average near horizontal line.
Step 5-3: calculate liter height e at edgepWith fall height ev。
ep∈{f(x+2,y)-f(x,y)>T}
ev∈{f(x+2,y)-f(x,y)<-T}
(x is y) that ((x+2 y) is the gray value of spaced points of this horizontal direction to f to current point for x, gray value y) to f.
Step 5-4: the liter height of lane line and fall height are to occur in pairs in the picture, and between meet certain
Distance.Relatively rise height and the width of fall height, reject ungratified point.
Δ w=ep(x)-ev(x)
If Δ w > W, then it is assumed that be it is unlikely that lane line, then to give up.Wherein, ep(x) and evX () represents respectively
Rising height and the row pixel coordinate of fall height, W is the maximum number of pixels that lane line occupies in the picture.
Step 6: clustering method based on lane line.
The lane line candidate domain obtained after feature point extraction, clusters lane line characteristic point.Assuming that starting point is (x0,
y0), set the square window of a m*n, i.e. m row n row, the most from bottom to top, scanning characteristic point from left to right.If
The characteristic point that in window, often row runs into for the first time is validity feature point, and other characteristic points of this row will be not scanned;If this
Not having characteristic point in individual window, window will translate n-1 row, not have characteristic point, cluster to terminate if accumulative more than 2n row;If y > h
Or x > w, cluster terminates, and wherein w is the width of image, and h is the height of image.
Step 7: the characteristic point generated after cluster is carried out Hough transform respectively, and extends straight line, take y=nearby respectively
180pix, at a distance y=120pix.
Step 8: calculate abscissa distance D of these 2 isolated edges at distance respectivelyrAnd Dl, then have W=Dr+Dl, so
There is Δ D=Dr-Dl, as Δ D > 0, inclined left lane line;When Δ D < when 0, inclined right lane line;As Δ D=0, do not occur
Skew.
Step 9: the result of detection shown or is sent to policymaker, in order to being adjusted in time;
When wherein calculating at distance, may have a relative error, this error utilizes the seriality of intra-frame trunk,
Can reduce.I.e. as continuous 5 frame images above error delta D ' > TD, TDTake 30pix, then reinitialize, and by up-to-date detection
Result is issued or is shown to policymaker;Pix represents pixel.
Further illustrate:
1. two video cameras installed in step 1 must be positioned in same plane, and can collect vehicle both sides lane line
Image;Do not require that there is automatic exposure, the function such as area-of-interest can be arranged, AWB;
2. the field range demarcated in step 2 to be determined according to the ultimate resolution of photographic head, and in experiment, we use
Sieve skill C170, ultimate resolution is 640*480, therefore, is designed as wide 600cm, and the visual field of remote 1000cm is the most suitable.
3. step 3, step 4 are carried out on a hardware platform, it is desirable to the platform of work at least meets more than internal memory 2G, place
Reason more than device 2.1Hz.
4. the feature point extraction of step 5 is the important module of lane detection, when calculating edge extracting threshold value T, and basis
Real road is adjusted, and under super expressway, the threshold value that we set will be most suitable.Additionally, it is unconspicuous at lane line
In the case of, may be greatly reduced even empty, for this situation, Wo Men by lane line information after feature point extraction
Algorithm devises the increase reliable point module of lane line.This requires when being IPM, and a image of reservation is as backup, in feature
When point extracts less, we contrast two parts of images, increase lane line information in region.Formula is as follows:
{(x1,y1),…,(xn,yn)}∈{(x1,y1),…,(xm,ym),(xm+1,ym+1)…,(xm+k,ym+k)}
5. the clustering method in step 6 is can to determine for the result in step 5, for left and right photographic head, it is known that
The lane line information of the left-hand lane line that left photographic head obtains, what right photographic head obtained is the lane line information on right side.At algorithm
In design, we are it is of concern that current lane line information, and therefore, the IPM image of left photographic head is most concerned is proximate to right side
The lane line characteristic point at edge, the most concerned lane line characteristic point being proximate to left side edge of the IPM image of right photographic head.In choosing
Take starting point (x0,y0) time it is noted that belong to the image of which photographic head.
6. in step 7, the lane line of all bends and straight way is contrasted by we, finds under our model, institute
Having bend the most all can be approximately straight line, 120-180 on hand is the most suitable for selection range.
7. the computational methods of the ratio k of the actual range in step 8 and pixel distance: intelligent vehicle is stopped in track
And parallel with lane line, from birds-eye view picture, calculate horizontal pixel distance P (unit pixel) in two adjacent lane lines, so
After measure the width W (unit cm), then k=W/p in a track;.
The invention has the beneficial effects as follows:
The present invention, by feasible technical scheme, not only can meet real-time during actual application, it can also be ensured that identifies
Accuracy is more than 95%, additionally significantly reduces cost.
Accompanying drawing explanation
The schematic flow sheet of Fig. 1 present invention
Fig. 2 camera calibration figure
Fig. 3-(a) bend turnout gray-scale map
Fig. 3-(b) straight way gray-scale map
Fig. 3-(c) bend gray-scale map
Fig. 4-(a) bend IMP figure-1
Fig. 4-(b) bend IMP figure-2
Fig. 4-(c) straight way IMP figure-1
Fig. 4-(d) straight way IMP figure-2
Fig. 5-(a) bend feature point extraction-1
Fig. 5-(b) bend feature point extraction-2
Fig. 5-(c) straight way I feature point extraction-1
Fig. 5-(d) straight way feature point extraction-2
Fig. 6 lane line dendrogram picture
Fig. 7 Hough transform image (extending)
Fig. 8 deviation distance result figure
Detailed description of the invention
The method using the present invention, provides the example of an indefiniteness, concrete real to the present invention further in conjunction with Fig. 1
The process of executing illustrates.The present invention realizes at intelligent vehicle platform, intelligent vehicle test site, in order to ensure driving intelligent vapour
Car and personal security, platform used and place are intelligent driving technology specialty experiment porch and test site.Used
Some current techiques such as image acquisition, image conversion etc. are not in narration in detail.
Embodiments of the present invention are as follows:
1. requiring to install video camera according to step 1, required device installed by platform used by this example, it is only necessary to slightly adjust
Just can test.
2. according to step 2, photographic head is demarcated, photographic head is finely adjusted whole.
3. realize according to the detailed step of 3,4,5,6,7,8,9.
Claims (1)
1. lane departure warning methods based on double low cost photographic head, it is characterised in that comprise the following steps:
Step 1: sensor is installed;
Video camera is separately mounted to the underface of intelligent vehicle sided mirror unit parallel with the longitudinal coordinate axle of car body, installs
During video camera, should ensure that two video cameras must be positioned in same plane, and the lane line image of vehicle both sides can be collected;
Step 2: camera calibration;
Demarcating two photographic head respectively, the field range of demarcation is: wide 600cm, remote 1000cm;Adjust a left side for photographic head
Lower angle right, upper, makes after demarcation that the position of left and right photographic head will be in the same plane;
Step 3: the pretreatment of image;
First image being carried out gray proces, the formula of gray proces is Gray=R*0.5+G*0.5, wherein R, G represent red indigo plant respectively
Channel components value, GrayRepresent the gray value of the pixel after conversion;Then image is carried out medium filtering, use 3 × 3 square
Image is filtered processing by field;
Step 4:IPM;
Gray level image carrying out inverse perspective process and obtains birds-eye view picture, perspective matrix is calibrated H-matrix;
Step 5-1: calculate the average of adjacent image pixels;
If certain point be (x, y), meet y ∈ [0, h) and x ∈ [2, w-2);X, y are the columns and rows of pixel respectively, and w is image
Width, h is the height of image;The average then having an adjacent image pixels is:
Wherein t=5;
Step 5-2: calculating edge extracting threshold value T, its computing formula is expressed as follows:
Wherein avg(x,y)For point (x, y) average near horizontal line;
Step 5-3: calculate liter height e at edgepWith fall height ev;
ep∈ { f (x+2, y)-f (x, y) > T}
ev∈ { f (x+2, y)-f (x, y) <-T}
(x is y) that ((x+2 y) is the gray value of spaced points of this horizontal direction to f to current point for x, gray value y) to f;
Step 5-4: compare liter height and the width of fall height, reject ungratified point;
Δ w=ep(x)-ev(x)
If Δ w > W, then it is assumed that be it is unlikely that lane line, then to give up;Wherein, ep(x) and evX () represents that liter becomes respectively
Point and the row pixel coordinate of fall height, W is the maximum number of pixels that lane line occupies in the picture;
Step 6: clustering method based on lane line;
The lane line candidate domain obtained after feature point extraction, clusters lane line characteristic point;Assuming that starting point is (x0,y0), if
The square window of a fixed m*n, i.e. m row n row, the most from bottom to top, scanning characteristic point from left to right;If in window
The characteristic point that often row runs into for the first time is validity feature point, and other characteristic points of this row will be not scanned;If this window
Inside not having characteristic point, window will translate n-1 row, not have characteristic point, cluster to terminate if accumulative more than 2n row;If y > h or x >
W, cluster terminates, and wherein w is the width of image, and h is the height of image;
Step 7: the characteristic point generated after cluster is carried out Hough transform respectively, and extends straight line, take y=nearby respectively
180pix, at a distance y=120pix;
Step 8: calculate abscissa distance D of these 2 isolated edges at distance respectivelyrAnd Dl, W=Dr+Dl, then have Δ D=Dr-
Dl, as Δ D > 0, inclined left lane line;As Δ D < 0, inclined right lane line;As Δ D=0, do not offset;
Step 9: the result of detection shown or is sent to policymaker, in order to being adjusted in time;When continuous 5 frame images above are missed
Difference Δ D ' > TD, TDTake 30pix, then reinitialize, and up-to-date testing result is issued or is shown to policymaker.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610423005.4A CN105966314B (en) | 2016-06-15 | 2016-06-15 | Lane departure warning method based on double inexpensive cameras |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610423005.4A CN105966314B (en) | 2016-06-15 | 2016-06-15 | Lane departure warning method based on double inexpensive cameras |
Publications (2)
Publication Number | Publication Date |
---|---|
CN105966314A true CN105966314A (en) | 2016-09-28 |
CN105966314B CN105966314B (en) | 2017-11-21 |
Family
ID=57020981
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610423005.4A Active CN105966314B (en) | 2016-06-15 | 2016-06-15 | Lane departure warning method based on double inexpensive cameras |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105966314B (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106529488A (en) * | 2016-11-18 | 2017-03-22 | 北京联合大学 | Lane line detection method based on ORB feature extraction |
CN106682586A (en) * | 2016-12-03 | 2017-05-17 | 北京联合大学 | Method for real-time lane line detection based on vision under complex lighting conditions |
CN109883432A (en) * | 2019-02-21 | 2019-06-14 | 百度在线网络技术(北京)有限公司 | Location determining method, device, equipment and computer readable storage medium |
CN110239436A (en) * | 2018-03-07 | 2019-09-17 | 松下知识产权经营株式会社 | Display control unit, vehicle-surroundings display system and display control method |
TWI819928B (en) * | 2022-12-20 | 2023-10-21 | 鴻海精密工業股份有限公司 | Method for detecting skewing of vehicle and related devices |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7551103B2 (en) * | 2001-07-31 | 2009-06-23 | Donnelly Corporation | Alert system for a vehicle |
JP2010218528A (en) * | 2009-03-18 | 2010-09-30 | Hyundai Motor Co Ltd | Lane departure warning method and system using virtual lane |
CN104859563A (en) * | 2015-05-28 | 2015-08-26 | 北京汽车股份有限公司 | Lane departure early warning method and lane departure early warning system |
CN104899858A (en) * | 2015-02-15 | 2015-09-09 | 北京联合大学 | Lane line identification method based on double-camera image fusion |
CN104951790A (en) * | 2015-02-15 | 2015-09-30 | 北京联合大学 | Lane line identification method based on seamless multi-source inverse perspective image splicing |
-
2016
- 2016-06-15 CN CN201610423005.4A patent/CN105966314B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7551103B2 (en) * | 2001-07-31 | 2009-06-23 | Donnelly Corporation | Alert system for a vehicle |
JP2010218528A (en) * | 2009-03-18 | 2010-09-30 | Hyundai Motor Co Ltd | Lane departure warning method and system using virtual lane |
CN104899858A (en) * | 2015-02-15 | 2015-09-09 | 北京联合大学 | Lane line identification method based on double-camera image fusion |
CN104951790A (en) * | 2015-02-15 | 2015-09-30 | 北京联合大学 | Lane line identification method based on seamless multi-source inverse perspective image splicing |
CN104859563A (en) * | 2015-05-28 | 2015-08-26 | 北京汽车股份有限公司 | Lane departure early warning method and lane departure early warning system |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106529488A (en) * | 2016-11-18 | 2017-03-22 | 北京联合大学 | Lane line detection method based on ORB feature extraction |
CN106682586A (en) * | 2016-12-03 | 2017-05-17 | 北京联合大学 | Method for real-time lane line detection based on vision under complex lighting conditions |
CN110239436A (en) * | 2018-03-07 | 2019-09-17 | 松下知识产权经营株式会社 | Display control unit, vehicle-surroundings display system and display control method |
CN109883432A (en) * | 2019-02-21 | 2019-06-14 | 百度在线网络技术(北京)有限公司 | Location determining method, device, equipment and computer readable storage medium |
CN109883432B (en) * | 2019-02-21 | 2020-12-29 | 百度在线网络技术(北京)有限公司 | Position determination method, device, equipment and computer readable storage medium |
TWI819928B (en) * | 2022-12-20 | 2023-10-21 | 鴻海精密工業股份有限公司 | Method for detecting skewing of vehicle and related devices |
Also Published As
Publication number | Publication date |
---|---|
CN105966314B (en) | 2017-11-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105966314A (en) | Lane departure pre-warning method based on double low-cost cameras | |
US9904856B2 (en) | Method and apparatus for detecting target object in blind area of vehicle | |
CN104008645B (en) | One is applicable to the prediction of urban road lane line and method for early warning | |
CN101630360B (en) | Method for identifying license plate in high-definition image | |
CN108038416B (en) | Lane line detection method and system | |
CN103324930B (en) | A kind of registration number character dividing method based on grey level histogram binaryzation | |
CN201427553Y (en) | Alarm system of vehicle departure from lane | |
Li et al. | Nighttime lane markings recognition based on Canny detection and Hough transform | |
CN111563412B (en) | Rapid lane line detection method based on parameter space voting and Bessel fitting | |
CN109886175B (en) | Method for detecting lane line by combining straight line and circular arc | |
Guo et al. | Image-based seat belt detection | |
CN105005771A (en) | Method for detecting full line of lane based on optical flow point locus statistics | |
CN108764108A (en) | A kind of Foregut fermenters method based on Bayesian inference | |
CN105809184A (en) | Vehicle real-time identification tracking and parking space occupancy determining method suitable for gas station | |
CN106887004A (en) | A kind of method for detecting lane lines based on Block- matching | |
CN104036246A (en) | Lane line positioning method based on multi-feature fusion and polymorphism mean value | |
He et al. | A lane detection method for lane departure warning system | |
CN104881661B (en) | Vehicle checking method based on structural similarity | |
CN103440785B (en) | One is traffic lane offset warning method fast | |
CN103313039B (en) | A kind of freeway tunnel entrance security prompt device on daytime and reminding method | |
CN105654073A (en) | Automatic speed control method based on visual detection | |
CN106485248A (en) | A kind of slag-soil truck license plate sloped correcting method | |
CN103593981A (en) | Vehicle model identification method based on video | |
CN107610506A (en) | The detection method and system of parking position state | |
CN102902951A (en) | System and method for vehicle target location and event detection on basis of high-definition video monitoring images |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |