CN105966314A - Lane departure pre-warning method based on double low-cost cameras - Google Patents

Lane departure pre-warning method based on double low-cost cameras Download PDF

Info

Publication number
CN105966314A
CN105966314A CN201610423005.4A CN201610423005A CN105966314A CN 105966314 A CN105966314 A CN 105966314A CN 201610423005 A CN201610423005 A CN 201610423005A CN 105966314 A CN105966314 A CN 105966314A
Authority
CN
China
Prior art keywords
image
lane line
point
lane
avg
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610423005.4A
Other languages
Chinese (zh)
Other versions
CN105966314B (en
Inventor
刘宏哲
袁家政
李超
宣寒宇
牛小宁
门晓杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Union University
Original Assignee
Beijing Union University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Union University filed Critical Beijing Union University
Priority to CN201610423005.4A priority Critical patent/CN105966314B/en
Publication of CN105966314A publication Critical patent/CN105966314A/en
Application granted granted Critical
Publication of CN105966314B publication Critical patent/CN105966314B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R1/00Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/14Means for informing the driver, warning the driver or prompting a driver intervention
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/181Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R2300/00Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
    • B60R2300/10Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used
    • B60R2300/105Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used using multiple cameras
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R2300/00Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
    • B60R2300/80Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
    • B60R2300/804Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for lane monitoring
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/14Means for informing the driver, warning the driver or prompting a driver intervention
    • B60W2050/146Display means

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Mechanical Engineering (AREA)
  • Automation & Control Theory (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Transportation (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention belongs to the field of aided driving, and relates to a lane departure pre-warning method based on double low-cost cameras. A lane departure pre-warning algorithm is achieved by using the double low-cost cameras. The method includes the steps that real-time images are obtained through the cameras below a left vehicle lug and a right vehicle lug of an intelligent vehicle; pretreatment, IPM and feature point extraction are carried out on the two images; feature points of candidate lane lines are clustered to obtain feature points of lane lines; Hoff straight line conversion detection is carried out on the images; the rightmost lane line of the left camera and the leftmost lane line of the right camera are extracted to calculate departure distance. The departure distance is calculated based on a model set up on the IPM images obtained after image inverse perspective conversion, and the problem that detection of the camera at one side fails is effectively solved. The double low-cost cameras are used for replacing a traditional single high-definition industrial camera, cost is greatly reduced, and the algorithm is low in operation cost and high in accuracy and real-time performance.

Description

Lane departure warning methods based on double low cost photographic head
Technical field
The present invention is lane departure warning methods based on double low cost photographic head, belongs to intelligence auxiliary driving technology neck Territory.
Background technology
Lane departure warning is a critical aspects of intelligence DAS (Driver Assistant System) research, along with safe driving, intelligence are handed over Logical increasingly by people's concern and attention, lane departure warning method research has become as the hot issue of research.Deviation The design of method for early warning is according to road ahead environment and this truck position relation, it is judged that automotive run-off-road distance is also carried out in time Remind, prevent owing to driver neglects the generation of the Lane Departure caused.Show according to research, lane departure warning method Realization and application can avoid the generation of deviation vehicle accident of 30-70%, to improving traffic safety, reduce and hand over Interpreter thus cause be significant.Its design head-up position under monocular-camera of traditional lane departure warning method enters OK, including image capture module, lane detection module and warning module.Typically first pass through imageing sensor and obtain vehicle row Road image real-time during sailing, then the run-off-road line that lane detection is obtained vehicle on hardware processing platform Whether distance parameters etc., then have the probability that deviation occurs to be estimated by method for early warning to current running state, Finally show early warning information.
Summary of the invention
The present invention utilizes the deviation that the photographic head of two low costs immediately below vehicle car ear (rearview mirror) is carried out Method for early warning, it is possible to lane departure warning information is provided correctly, in real time.Its advantage be the cost requirement to photographic head very Low, and ensure that computing overhead minimum and real-time.The design of algorithm is photographic head based on double low costs, it is possible to adapt to car Produce family to cost, real-time and requirement accurately, market prospect will be the best.
To achieve these goals, this invention takes following technical scheme:
Step 1: sensor is installed;
Video camera is separately mounted to the underface of intelligent vehicle sided mirror unit parallel with the longitudinal coordinate axle of car body, When video camera is installed, should ensure that two video cameras must be positioned in same plane, and vehicle both sides lane line image can be collected;
Step 2: camera calibration;
Demarcating two photographic head respectively, the field range of demarcation is: wide 600cm, remote 1000cm.Adjust photographic head Left and right, upper lower angle, after enabling demarcation, the position of left and right photographic head (to require that the picture obtained is same in the same plane Process in plane);
Step 3: the pretreatment of image;
According to lane line feature, first image being carried out gray proces, the formula of gray proces is
Gray=R*0.5+G*0.5, wherein R, G represent red blue channel component value, G respectivelyrayRepresent the pixel after conversion Gray value.Then image is carried out medium filtering, this experiment uses the square field of 3 × 3 be filtered image processing.
Step 4:IPM;
Gray level image carrying out inverse perspective process and obtains birds-eye view picture, perspective matrix is calibrated H-matrix.
Step 5: feature point extraction;
The gray value of lane line is higher than the value on its both sides, forms a crest;Present and from left to right fall after rising Trend;When in lane line region, average is higher, average differs bigger with its summit value.We utilize these characteristics, by calculating The change of adjacent image pixels judges the edge of lane line.
Step 5-1: calculate the average of adjacent image pixels;
If certain point be (x, y), meet y ∈ [0, h) and x ∈ [2, w-2).X, y are the columns and rows of pixel respectively, and w is figure The width of picture, h is the height of image.The average then having an adjacent image pixels is:
avg ( x , y ) = 1 t Σ i = - t / 2 i = t / 2 f ( i + x , y )
Wherein t ∈ [1,3,5,7 ...], t=5 can obtain good effect.
Step 5-2: calculating edge extracting threshold value T, its computing formula is expressed as follows:
T = avg ( x , y ) 12 avg ( x , y ) > 200 avg ( x , y ) 5 50 < avg ( x , y ) &le; 200 avg ( x , y ) 10 o t h e r
Wherein avg(x,y)For point (x, y) average near horizontal line.
Step 5-3: calculate liter height e at edgepWith fall height ev
ep∈{f(x+2,y)-f(x,y)>T}
ev∈{f(x+2,y)-f(x,y)<-T}
(x is y) that ((x+2 y) is the gray value of spaced points of this horizontal direction to f to current point for x, gray value y) to f.
Step 5-4: the liter height of lane line and fall height are to occur in pairs in the picture, and between meet certain Distance.Relatively rise height and the width of fall height, reject ungratified point.
Δ w=ep(x)-ev(x)
If Δ w > W, then it is assumed that be it is unlikely that lane line, then to give up.Wherein, ep(x) and evX () represents respectively Rising height and the row pixel coordinate of fall height, W is the maximum number of pixels that lane line occupies in the picture.
Step 6: clustering method based on lane line.
The lane line candidate domain obtained after feature point extraction, clusters lane line characteristic point.Assuming that starting point is (x0, y0), set the square window of a m*n, i.e. m row n row, the most from bottom to top, scanning characteristic point from left to right.If The characteristic point that in window, often row runs into for the first time is validity feature point, and other characteristic points of this row will be not scanned;If this Not having characteristic point in individual window, window will translate n-1 row, not have characteristic point, cluster to terminate if accumulative more than 2n row;If y > h Or x > w, cluster terminates, and wherein w is the width of image, and h is the height of image.
Step 7: the characteristic point generated after cluster is carried out Hough transform respectively, and extends straight line, take y=nearby respectively 180pix, at a distance y=120pix.
Step 8: calculate abscissa distance D of these 2 isolated edges at distance respectivelyrAnd Dl, then have W=Dr+Dl, so There is Δ D=Dr-Dl, as Δ D > 0, inclined left lane line;When Δ D < when 0, inclined right lane line;As Δ D=0, do not occur Skew.
Step 9: the result of detection shown or is sent to policymaker, in order to being adjusted in time;
When wherein calculating at distance, may have a relative error, this error utilizes the seriality of intra-frame trunk, Can reduce.I.e. as continuous 5 frame images above error delta D ' > TD, TDTake 30pix, then reinitialize, and by up-to-date detection Result is issued or is shown to policymaker;Pix represents pixel.
Further illustrate:
1. two video cameras installed in step 1 must be positioned in same plane, and can collect vehicle both sides lane line Image;Do not require that there is automatic exposure, the function such as area-of-interest can be arranged, AWB;
2. the field range demarcated in step 2 to be determined according to the ultimate resolution of photographic head, and in experiment, we use Sieve skill C170, ultimate resolution is 640*480, therefore, is designed as wide 600cm, and the visual field of remote 1000cm is the most suitable.
3. step 3, step 4 are carried out on a hardware platform, it is desirable to the platform of work at least meets more than internal memory 2G, place Reason more than device 2.1Hz.
4. the feature point extraction of step 5 is the important module of lane detection, when calculating edge extracting threshold value T, and basis Real road is adjusted, and under super expressway, the threshold value that we set will be most suitable.Additionally, it is unconspicuous at lane line In the case of, may be greatly reduced even empty, for this situation, Wo Men by lane line information after feature point extraction Algorithm devises the increase reliable point module of lane line.This requires when being IPM, and a image of reservation is as backup, in feature When point extracts less, we contrast two parts of images, increase lane line information in region.Formula is as follows:
{(x1,y1),…,(xn,yn)}∈{(x1,y1),…,(xm,ym),(xm+1,ym+1)…,(xm+k,ym+k)}
5. the clustering method in step 6 is can to determine for the result in step 5, for left and right photographic head, it is known that The lane line information of the left-hand lane line that left photographic head obtains, what right photographic head obtained is the lane line information on right side.At algorithm In design, we are it is of concern that current lane line information, and therefore, the IPM image of left photographic head is most concerned is proximate to right side The lane line characteristic point at edge, the most concerned lane line characteristic point being proximate to left side edge of the IPM image of right photographic head.In choosing Take starting point (x0,y0) time it is noted that belong to the image of which photographic head.
6. in step 7, the lane line of all bends and straight way is contrasted by we, finds under our model, institute Having bend the most all can be approximately straight line, 120-180 on hand is the most suitable for selection range.
7. the computational methods of the ratio k of the actual range in step 8 and pixel distance: intelligent vehicle is stopped in track And parallel with lane line, from birds-eye view picture, calculate horizontal pixel distance P (unit pixel) in two adjacent lane lines, so After measure the width W (unit cm), then k=W/p in a track;.
The invention has the beneficial effects as follows:
The present invention, by feasible technical scheme, not only can meet real-time during actual application, it can also be ensured that identifies Accuracy is more than 95%, additionally significantly reduces cost.
Accompanying drawing explanation
The schematic flow sheet of Fig. 1 present invention
Fig. 2 camera calibration figure
Fig. 3-(a) bend turnout gray-scale map
Fig. 3-(b) straight way gray-scale map
Fig. 3-(c) bend gray-scale map
Fig. 4-(a) bend IMP figure-1
Fig. 4-(b) bend IMP figure-2
Fig. 4-(c) straight way IMP figure-1
Fig. 4-(d) straight way IMP figure-2
Fig. 5-(a) bend feature point extraction-1
Fig. 5-(b) bend feature point extraction-2
Fig. 5-(c) straight way I feature point extraction-1
Fig. 5-(d) straight way feature point extraction-2
Fig. 6 lane line dendrogram picture
Fig. 7 Hough transform image (extending)
Fig. 8 deviation distance result figure
Detailed description of the invention
The method using the present invention, provides the example of an indefiniteness, concrete real to the present invention further in conjunction with Fig. 1 The process of executing illustrates.The present invention realizes at intelligent vehicle platform, intelligent vehicle test site, in order to ensure driving intelligent vapour Car and personal security, platform used and place are intelligent driving technology specialty experiment porch and test site.Used Some current techiques such as image acquisition, image conversion etc. are not in narration in detail.
Embodiments of the present invention are as follows:
1. requiring to install video camera according to step 1, required device installed by platform used by this example, it is only necessary to slightly adjust Just can test.
2. according to step 2, photographic head is demarcated, photographic head is finely adjusted whole.
3. realize according to the detailed step of 3,4,5,6,7,8,9.

Claims (1)

1. lane departure warning methods based on double low cost photographic head, it is characterised in that comprise the following steps:
Step 1: sensor is installed;
Video camera is separately mounted to the underface of intelligent vehicle sided mirror unit parallel with the longitudinal coordinate axle of car body, installs During video camera, should ensure that two video cameras must be positioned in same plane, and the lane line image of vehicle both sides can be collected;
Step 2: camera calibration;
Demarcating two photographic head respectively, the field range of demarcation is: wide 600cm, remote 1000cm;Adjust a left side for photographic head Lower angle right, upper, makes after demarcation that the position of left and right photographic head will be in the same plane;
Step 3: the pretreatment of image;
First image being carried out gray proces, the formula of gray proces is Gray=R*0.5+G*0.5, wherein R, G represent red indigo plant respectively Channel components value, GrayRepresent the gray value of the pixel after conversion;Then image is carried out medium filtering, use 3 × 3 square Image is filtered processing by field;
Step 4:IPM;
Gray level image carrying out inverse perspective process and obtains birds-eye view picture, perspective matrix is calibrated H-matrix;
Step 5-1: calculate the average of adjacent image pixels;
If certain point be (x, y), meet y ∈ [0, h) and x ∈ [2, w-2);X, y are the columns and rows of pixel respectively, and w is image Width, h is the height of image;The average then having an adjacent image pixels is:
avg ( x , y ) = 1 t &Sigma; i = - t / 2 i = t / 2 f ( i + x , y )
Wherein t=5;
Step 5-2: calculating edge extracting threshold value T, its computing formula is expressed as follows:
T = avg ( x , y ) 12 avg ( x , y ) > 200 avg ( x , y ) 5 50 < avg ( x , y ) &le; 200 avg ( x , y ) 10 o t h e r
Wherein avg(x,y)For point (x, y) average near horizontal line;
Step 5-3: calculate liter height e at edgepWith fall height ev
ep∈ { f (x+2, y)-f (x, y) > T}
ev∈ { f (x+2, y)-f (x, y) <-T}
(x is y) that ((x+2 y) is the gray value of spaced points of this horizontal direction to f to current point for x, gray value y) to f;
Step 5-4: compare liter height and the width of fall height, reject ungratified point;
Δ w=ep(x)-ev(x)
If Δ w > W, then it is assumed that be it is unlikely that lane line, then to give up;Wherein, ep(x) and evX () represents that liter becomes respectively Point and the row pixel coordinate of fall height, W is the maximum number of pixels that lane line occupies in the picture;
Step 6: clustering method based on lane line;
The lane line candidate domain obtained after feature point extraction, clusters lane line characteristic point;Assuming that starting point is (x0,y0), if The square window of a fixed m*n, i.e. m row n row, the most from bottom to top, scanning characteristic point from left to right;If in window The characteristic point that often row runs into for the first time is validity feature point, and other characteristic points of this row will be not scanned;If this window Inside not having characteristic point, window will translate n-1 row, not have characteristic point, cluster to terminate if accumulative more than 2n row;If y > h or x > W, cluster terminates, and wherein w is the width of image, and h is the height of image;
Step 7: the characteristic point generated after cluster is carried out Hough transform respectively, and extends straight line, take y=nearby respectively 180pix, at a distance y=120pix;
Step 8: calculate abscissa distance D of these 2 isolated edges at distance respectivelyrAnd Dl, W=Dr+Dl, then have Δ D=Dr- Dl, as Δ D > 0, inclined left lane line;As Δ D < 0, inclined right lane line;As Δ D=0, do not offset;
Step 9: the result of detection shown or is sent to policymaker, in order to being adjusted in time;When continuous 5 frame images above are missed Difference Δ D ' > TD, TDTake 30pix, then reinitialize, and up-to-date testing result is issued or is shown to policymaker.
CN201610423005.4A 2016-06-15 2016-06-15 Lane departure warning method based on double inexpensive cameras Active CN105966314B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610423005.4A CN105966314B (en) 2016-06-15 2016-06-15 Lane departure warning method based on double inexpensive cameras

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610423005.4A CN105966314B (en) 2016-06-15 2016-06-15 Lane departure warning method based on double inexpensive cameras

Publications (2)

Publication Number Publication Date
CN105966314A true CN105966314A (en) 2016-09-28
CN105966314B CN105966314B (en) 2017-11-21

Family

ID=57020981

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610423005.4A Active CN105966314B (en) 2016-06-15 2016-06-15 Lane departure warning method based on double inexpensive cameras

Country Status (1)

Country Link
CN (1) CN105966314B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106529488A (en) * 2016-11-18 2017-03-22 北京联合大学 Lane line detection method based on ORB feature extraction
CN106682586A (en) * 2016-12-03 2017-05-17 北京联合大学 Method for real-time lane line detection based on vision under complex lighting conditions
CN109883432A (en) * 2019-02-21 2019-06-14 百度在线网络技术(北京)有限公司 Location determining method, device, equipment and computer readable storage medium
CN110239436A (en) * 2018-03-07 2019-09-17 松下知识产权经营株式会社 Display control unit, vehicle-surroundings display system and display control method
TWI819928B (en) * 2022-12-20 2023-10-21 鴻海精密工業股份有限公司 Method for detecting skewing of vehicle and related devices

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7551103B2 (en) * 2001-07-31 2009-06-23 Donnelly Corporation Alert system for a vehicle
JP2010218528A (en) * 2009-03-18 2010-09-30 Hyundai Motor Co Ltd Lane departure warning method and system using virtual lane
CN104859563A (en) * 2015-05-28 2015-08-26 北京汽车股份有限公司 Lane departure early warning method and lane departure early warning system
CN104899858A (en) * 2015-02-15 2015-09-09 北京联合大学 Lane line identification method based on double-camera image fusion
CN104951790A (en) * 2015-02-15 2015-09-30 北京联合大学 Lane line identification method based on seamless multi-source inverse perspective image splicing

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7551103B2 (en) * 2001-07-31 2009-06-23 Donnelly Corporation Alert system for a vehicle
JP2010218528A (en) * 2009-03-18 2010-09-30 Hyundai Motor Co Ltd Lane departure warning method and system using virtual lane
CN104899858A (en) * 2015-02-15 2015-09-09 北京联合大学 Lane line identification method based on double-camera image fusion
CN104951790A (en) * 2015-02-15 2015-09-30 北京联合大学 Lane line identification method based on seamless multi-source inverse perspective image splicing
CN104859563A (en) * 2015-05-28 2015-08-26 北京汽车股份有限公司 Lane departure early warning method and lane departure early warning system

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106529488A (en) * 2016-11-18 2017-03-22 北京联合大学 Lane line detection method based on ORB feature extraction
CN106682586A (en) * 2016-12-03 2017-05-17 北京联合大学 Method for real-time lane line detection based on vision under complex lighting conditions
CN110239436A (en) * 2018-03-07 2019-09-17 松下知识产权经营株式会社 Display control unit, vehicle-surroundings display system and display control method
CN109883432A (en) * 2019-02-21 2019-06-14 百度在线网络技术(北京)有限公司 Location determining method, device, equipment and computer readable storage medium
CN109883432B (en) * 2019-02-21 2020-12-29 百度在线网络技术(北京)有限公司 Position determination method, device, equipment and computer readable storage medium
TWI819928B (en) * 2022-12-20 2023-10-21 鴻海精密工業股份有限公司 Method for detecting skewing of vehicle and related devices

Also Published As

Publication number Publication date
CN105966314B (en) 2017-11-21

Similar Documents

Publication Publication Date Title
CN105966314A (en) Lane departure pre-warning method based on double low-cost cameras
US9904856B2 (en) Method and apparatus for detecting target object in blind area of vehicle
CN104008645B (en) One is applicable to the prediction of urban road lane line and method for early warning
CN101630360B (en) Method for identifying license plate in high-definition image
CN108038416B (en) Lane line detection method and system
CN103324930B (en) A kind of registration number character dividing method based on grey level histogram binaryzation
CN201427553Y (en) Alarm system of vehicle departure from lane
Li et al. Nighttime lane markings recognition based on Canny detection and Hough transform
CN111563412B (en) Rapid lane line detection method based on parameter space voting and Bessel fitting
CN109886175B (en) Method for detecting lane line by combining straight line and circular arc
Guo et al. Image-based seat belt detection
CN105005771A (en) Method for detecting full line of lane based on optical flow point locus statistics
CN108764108A (en) A kind of Foregut fermenters method based on Bayesian inference
CN105809184A (en) Vehicle real-time identification tracking and parking space occupancy determining method suitable for gas station
CN106887004A (en) A kind of method for detecting lane lines based on Block- matching
CN104036246A (en) Lane line positioning method based on multi-feature fusion and polymorphism mean value
He et al. A lane detection method for lane departure warning system
CN104881661B (en) Vehicle checking method based on structural similarity
CN103440785B (en) One is traffic lane offset warning method fast
CN103313039B (en) A kind of freeway tunnel entrance security prompt device on daytime and reminding method
CN105654073A (en) Automatic speed control method based on visual detection
CN106485248A (en) A kind of slag-soil truck license plate sloped correcting method
CN103593981A (en) Vehicle model identification method based on video
CN107610506A (en) The detection method and system of parking position state
CN102902951A (en) System and method for vehicle target location and event detection on basis of high-definition video monitoring images

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant