US10118610B2 - Autonomous vehicle using path prediction - Google Patents

Autonomous vehicle using path prediction Download PDF

Info

Publication number
US10118610B2
US10118610B2 US15/252,425 US201615252425A US10118610B2 US 10118610 B2 US10118610 B2 US 10118610B2 US 201615252425 A US201615252425 A US 201615252425A US 10118610 B2 US10118610 B2 US 10118610B2
Authority
US
United States
Prior art keywords
host
targ
vehicle
controller
bound
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US15/252,425
Other versions
US20180056995A1 (en
Inventor
Kun Deng
Wangdong Luo
Nanjun LIU
Alex Maurice Miller
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ford Global Technologies LLC
Original Assignee
Ford Global Technologies LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ford Global Technologies LLC filed Critical Ford Global Technologies LLC
Assigned to FORD GLOBAL TECHNOLOGIES, LLC reassignment FORD GLOBAL TECHNOLOGIES, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LUO, WANGDONG, DENG, Kun, LIU, Nanjun, MILLER, ALEX MAURICE
Priority to US15/252,425 priority Critical patent/US10118610B2/en
Priority to GB1713487.5A priority patent/GB2556146A/en
Priority to RU2017129885A priority patent/RU2017129885A/en
Priority to CN201710758086.8A priority patent/CN107792079B/en
Priority to MX2017011127A priority patent/MX2017011127A/en
Priority to DE102017119952.7A priority patent/DE102017119952A1/en
Publication of US20180056995A1 publication Critical patent/US20180056995A1/en
Publication of US10118610B2 publication Critical patent/US10118610B2/en
Application granted granted Critical
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
    • B60W30/08Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
    • B60W30/09Taking automatic action to avoid collision, e.g. braking and steering
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/0097Predicting future conditions
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W10/00Conjoint control of vehicle sub-units of different type or different function
    • B60W10/18Conjoint control of vehicle sub-units of different type or different function including control of braking systems
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W10/00Conjoint control of vehicle sub-units of different type or different function
    • B60W10/18Conjoint control of vehicle sub-units of different type or different function including control of braking systems
    • B60W10/184Conjoint control of vehicle sub-units of different type or different function including control of braking systems with wheel brakes
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W10/00Conjoint control of vehicle sub-units of different type or different function
    • B60W10/20Conjoint control of vehicle sub-units of different type or different function including control of steering systems
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
    • B60W30/08Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
    • B60W30/095Predicting travel path or likelihood of collision
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
    • B60W30/08Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
    • B60W30/095Predicting travel path or likelihood of collision
    • B60W30/0953Predicting travel path or likelihood of collision the prediction being responsive to vehicle dynamic parameters
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units
    • B60W30/08Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
    • B60W30/095Predicting travel path or likelihood of collision
    • B60W30/0956Predicting travel path or likelihood of collision the prediction being responsive to traffic or environmental parameters
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B62LAND VEHICLES FOR TRAVELLING OTHERWISE THAN ON RAILS
    • B62DMOTOR VEHICLES; TRAILERS
    • B62D15/00Steering not otherwise provided for
    • B62D15/02Steering position indicators ; Steering position determination; Steering aids
    • B62D15/025Active steering aids, e.g. helping the driver by actively influencing the steering system after environment evaluation
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/0088Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots characterized by the autonomous decision making process, e.g. artificial intelligence, predefined behaviours
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/16Anti-collision systems
    • G08G1/166Anti-collision systems for active traffic, e.g. moving vehicles, pedestrians, bikes
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2420/00Indexing codes relating to the type of sensors based on the principle of their operation
    • B60W2420/40Photo, light or radio wave sensitive means, e.g. infrared sensors
    • B60W2420/408Radar; Laser, e.g. lidar
    • B60W2420/52
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2520/00Input parameters relating to overall vehicle dynamics
    • B60W2520/06Direction of travel
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2520/00Input parameters relating to overall vehicle dynamics
    • B60W2520/10Longitudinal speed
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2520/00Input parameters relating to overall vehicle dynamics
    • B60W2520/14Yaw
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2554/00Input parameters relating to objects
    • B60W2554/40Dynamic objects, e.g. animals, windblown objects
    • B60W2554/404Characteristics
    • B60W2554/4041Position
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2554/00Input parameters relating to objects
    • B60W2554/80Spatial relation or speed relative to objects
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2554/00Input parameters relating to objects
    • B60W2554/80Spatial relation or speed relative to objects
    • B60W2554/804Relative longitudinal speed
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2556/00Input parameters relating to data
    • B60W2600/00
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2710/00Output or target parameters relating to a particular sub-units
    • B60W2710/18Braking system
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2710/00Output or target parameters relating to a particular sub-units
    • B60W2710/20Steering systems
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2900/00Indexing codes relating to the purpose of, or problem solved of road vehicle drive control systems not otherwise provided for in groups B60W30/00

Definitions

  • Autonomous vehicles have the ability to operate without the intervention of a human operator, e.g., driver, that is, a vehicle controller makes decisions about accelerating, braking, and/or steering the vehicle.
  • a vehicle may be fully autonomous or semi-autonomous.
  • a semi-autonomous vehicle may be autonomous only in particular situations, for example, highway driving or parallel parking, or with respect to certain vehicle subsystems, for example, braking but not acceleration or steering.
  • An autonomous vehicle may include sensors for tracking an external environment surrounding the vehicle. Some types of sensors are radar sensors, scanning laser range finders, light detection and ranging (LIDAR) devices, and image processing sensors such as cameras.
  • the vehicle controller is in communication with the sensors and uses output from the sensors to analyze the external environment, for example, defining features of a surrounding landscape, detecting roads and lanes of roads on the landscape, interpreting signs and signals, and tracking and classifying objects in the environment such as vehicles, cyclists, and pedestrians. For example, a vehicle controller may classify whether a detected object is another vehicle and provide state information about the other vehicle, such as location, speed, and heading.
  • the vehicle controller uses target path prediction to predict where another vehicle will travel.
  • the vehicle controller uses the predicted path of the other vehicle to make decisions affecting operation of the vehicle.
  • inaccuracies and failures of current path prediction technologies are problematic.
  • FIG. 1 is a block diagram of a host vehicle.
  • FIG. 2 is a process flow diagram of a process for implementing target path prediction on the host vehicle.
  • FIG. 3 is a process flow diagram of a process for disabling a feature of the host vehicle based on a quality of the target path prediction.
  • FIG. 4 is a process flow diagram of a process for running a test to determine the quality of the target path prediction.
  • FIG. 5 is a process flow diagram of a process for determining the quality of the target path prediction during a test run.
  • FIG. 6 is a process flow diagram of a process for operating the host vehicle using target path prediction.
  • FIG. 7 is a diagram of an interaction between the host vehicle and a target vehicle.
  • a controller 34 in a host vehicle 30 includes a processor and a memory storing processor-executable instructions.
  • the processor is programmed to calculate an error bound for confidence intervals of a predicted position at a future time of a target vehicle 32 relative laterally to a current position and orientation of the host vehicle 30 at a current time based on a current position and velocity of the target vehicle 32 and the future time.
  • error bound makes target path prediction, and hence autonomous operation of the host vehicle 30 , more robust because the controller 34 makes decisions based on a range of the most likely predicted positions for the target vehicle 32 . Furthermore, calculation of the error bound is adaptive in that the calculation updates continuously as the host vehicle 30 travels.
  • the host vehicle 30 may be an autonomous vehicle.
  • the controller 34 may be capable of operating the vehicle independently of the intervention of a human driver, to a greater or a lesser degree, i.e., fully autonomously and/or in a semi-autonomous mode.
  • the controller 34 may be programmed to operate a steering system 40 , a propulsion 42 , brakes 44 , and/or other vehicle systems.
  • fully autonomous means that each of the steering system 40 , the propulsion 42 , and the brakes 44 of the host vehicle 30 are controlled by the controller 34 ;
  • a semi-autonomous mode is one in which one or two of the steering system 40 , the propulsion 42 , and the brakes 44 are operated by the controller 34 .
  • the controller 34 carries out various operations, including as described herein.
  • the controller 34 is a computing device that generally includes a processor and a memory, the memory including one or more forms of computer-readable media, and storing instructions executable by the processor for performing various operations, including as disclosed herein.
  • the memory of the controller 34 further generally stores remote data received via various communications mechanisms; e.g., the controller 34 is generally configured for communications on a controller area network (CAN) bus or the like, and/or for using other wired or wireless protocols, e.g., Bluetooth, etc.
  • the controller 34 may also have a connection to an onboard diagnostics connector (OBD-II).
  • OBD-II onboard diagnostics connector
  • the controller 34 may transmit messages to various devices in the vehicle and/or receive messages from the various devices, e.g., controllers, actuators, sensors, etc., e.g., controllers and sensors as discussed herein.
  • the controller 34 may receive data from sensors 36 .
  • FIG. 1 one controller 34 is shown in FIG. 1 for ease of illustration, it is to be understood that the controller 34 could include, and various operations described herein could be carried out by, one or more computing devices.
  • the controller 34 may transmit signals through a communications network 38 such as a controller area network (CAN) bus, Ethernet, Local Interconnect Network (LIN), and/or by any other wired or wireless communications network.
  • the controller 34 may be in communication with the sensors 36 , the steering system 40 , the propulsion 42 , the brakes 44 , and other vehicle subsystems and/or features.
  • the host vehicle 30 may include the sensors 36 .
  • the sensors 36 may detect internal states of the vehicle, for example, wheel speed, wheel orientation, and engine and transmission variables.
  • the sensors 36 may detect the position or orientation of the vehicle, for example, global positioning system (GPS) sensors; accelerometers such as piezo-electric or microelectromechanical systems (MEMS); gyroscopes such as rate, ring laser, or fiber-optic gyroscopes; inertial measurements units (IMU); and magnetometers.
  • the sensors 36 may detect an environment external to a body of the host vehicle 30 , for example, the sensors 36 can include one or more of radar sensors, scanning laser range finders, light detection and ranging (LIDAR) devices, and image processing sensors such as cameras.
  • the sensors 36 may include communications devices, for example, vehicle-to-infrastructure (V2I) or vehicle-to-vehicle (V2V) devices.
  • V2I vehicle-to-infrastructure
  • V2V
  • the steering system 40 is typically a known vehicle steering subsystem and controls the turning of wheels of the host vehicle 30 .
  • the steering system 40 is in communication with and receives input from a steering wheel and the controller 34 .
  • the steering system 40 may be a rack-and-pinion system with electric power-assisted steering, a steer-by-wire system, as are both known in the art, or any other suitable system.
  • the propulsion 42 of the host vehicle 30 generates energy and translates the energy into motion of the host vehicle 30 .
  • the propulsion 42 may be a known vehicle propulsion subsystem, for example, a conventional powertrain including an internal-combustion engine coupled to a transmission that transfers rotational motion to wheels; an electric powertrain including batteries, an electric motor, and a transmission that transfers rotational motion to the wheels; a hybrid powertrain including elements of the conventional powertrain and the electric powertrain; or any other type of propulsion.
  • the propulsion 42 is in communication with and receives input from the controller 34 and from a human driver.
  • the human driver may control the propulsion 42 via, e.g., an accelerator pedal and/or a gear-shift lever.
  • the brakes 44 are typically a known vehicle braking subsystem and are provided to resist the motion of the host vehicle 30 to thereby slow and/or stop the vehicle.
  • the brakes 44 may be friction brakes such as disc brakes, drum brakes, band brakes, and so on; regenerative brakes; any other suitable type of brakes; or a combination.
  • the brakes 44 are in communication with and receive input from the controller 34 and a human driver.
  • the human driver may control the brakes 44 via, e.g., a brake pedal.
  • Target path prediction means predicting, by the host vehicle 30 , a path P along which the target vehicle 32 will travel. An actual path A by the target vehicle 32 may differ from the predicted path P. Motion variables shown in FIG. 7 describing the travel of the vehicles 30 , 32 are described below with respect to a block 415 in a process 400 .
  • FIG. 2 is a process flow diagram illustrating an exemplary process 200 for implementing target path prediction on the host vehicle 30 .
  • the process 200 may be used by, for example, a system designer testing and implementing an algorithm for target path prediction on a type of vehicle.
  • the process 200 begins by performing an exemplary process 400 for running a test to determine the quality of the target path prediction, shown in FIG. 4 and described below.
  • the process 400 returns a quality measure Q of the target path prediction to the controller 34 .
  • the controller 34 determines whether the quality measure Q is above a quality threshold Q th .
  • the quality threshold is a value chosen to ensure a sufficiently robust performance of the target path prediction.
  • a system designer adjusts a path prediction algorithm for target path prediction based on the quality measure Q.
  • the system designer may adjust parameters for the target path prediction algorithm, such as the quadratic polynomial method, as is known (described below with respect to a block 420 of the process 400 and a block 515 of a process 500 ).
  • the system designer may adjust parameters for the error bounds aspect of the path prediction algorithm (described below with respect to a block 525 of the process 500 ), for example, a lateral offset bound x targ for the target vehicle 32 , a longitudinal offset bound y targ for the target vehicle 32 , a location bound b for the target vehicle 32 , a velocity bound v targ for the target vehicle 32 , a velocity bound v host for the host vehicle 30 , a yaw rate bound w host for the host vehicle 30 , and/or a default initialization values ⁇ x targ , ⁇ y targ , ⁇ v targ , ⁇ v host , and ⁇ w host .
  • the controller 34 performs an exemplary process 600 for operating the host vehicle 30 using target path prediction, as shown in FIG. 6 and described below. After the process 600 , the process 200 ends.
  • FIG. 3 is a process flow diagram illustrating an exemplary process 300 for disabling a feature of the host vehicle 30 based on a quality of the target path prediction.
  • the process 300 may be performed, for example, by the controller 34 in the host vehicle 30 .
  • the process 300 begins by performing the exemplary process 400 for running a test to determine the quality of the target path prediction, shown in FIG. 4 and described below.
  • the process 400 returns the quality measure Q of the target path prediction to the controller 34 .
  • the controller 34 determines whether the quality measure Q is above a quality threshold Q th , as described above with respect to the decision block 205 .
  • the controller 34 disables a feature relying on an error bound ⁇ (described below with respect to the block 525 of the process 500 ) based on the quality measure Q.
  • an error bound ⁇ may include autonomous or semi-autonomous operation of the host vehicle 30 , that is, operation of one or more of the steering system 40 , propulsion 42 , and brakes 44 , such as automated braking or automated steering.
  • Other features that may rely on the error bound ⁇ may include features that may anticipate movement of the host vehicle 30 , for example, automatically turning headlights.
  • the controller 34 performs the exemplary process 600 for operating the host vehicle 30 using target path prediction, as shown in FIG. 6 and described below. After the process 600 , the process 300 ends.
  • FIG. 4 is a process flow diagram illustrating the exemplary process 400 for running a test to determine the quality of the target path prediction.
  • the process 400 begins in a block 405 , in which the controller 34 initializes an index value k to 1.
  • the index value k counts a number of test runs each of a duration equal to a test-run time limit T M .
  • the controller 34 performs the kth test run.
  • Performing the test run involves driving the host vehicle 30 for a duration T M from an initial time t 0 to a final time t 0 +T M .
  • Driving the host vehicle 30 may be performed autonomously by the controller 34 , semi-autonomously by the controller 34 and a human driver, or manually by the human driver.
  • the block 410 begins an experiment cycle through blocks 410 to 430 of a number K of test runs.
  • the number K of test runs is a preset or user-specified value and may be chosen, e.g., to provide a sufficient period to gather a statistically useful set of data, as described below.
  • the controller 34 measures a set of motion variables for the host vehicle 30 and the target vehicle 32 .
  • the motion variables may be measured in a known coordinate system such as an objective over-the-ground coordinate system, a host-vehicle-centered coordinate system, or any other suitable coordinate system, and the motion variables may be converted from one coordinate system to another before further use.
  • the controller 34 may measure this set of motion variables in over-the-ground coordinates:
  • the controller 34 generates coefficients to calculate the predicted position of the target vehicle 32 at the future time using a quadratic polynomial method.
  • Another method of target path prediction besides the quadratic polynomial method such as a Kalman filter-based motion-prediction method, may be used.
  • the controller 34 performs an exemplary process 500 for determining the quality of the target path prediction during the test run, as shown in FIG. 5 and described below.
  • the process 500 returns a quality measure Q test of the target path prediction during the test run to the controller 34 .
  • the controller 34 increments the index value k by 1, so the index value k is equal to k+1.
  • the controller 34 determines whether the index value k has reached or exceeded the number K of test runs. If the index value k is still below the number K of test runs, then the process 400 proceeds back to the block 410 , and the controller 34 performs another loop of the experiment cycle of the blocks 410 to 430 with the update index value k from the block 425 .
  • the controller 34 calculates the quality measure Q equal to a minimum value during the period of the experiment of a lesser value of 1 and a ratio of the error bound ⁇ and the error metric ⁇ , as described below with respect to the process 500 . In other words, the controller 34 calculates the lowest minimum itemized quality measure Q test :
  • FIG. 5 is a process flow diagram illustrating the exemplary process 500 for determining the quality of the target path prediction during the test run.
  • the process 500 begins in a block 505 , in which the controller 34 initializes a time variable t to an initial value t 0 .
  • the controller 34 initializes a prediction time variable s to the value of the time variable t.
  • the time variable t tracks the current time
  • the prediction time variable s tracks the future time for predictions.
  • the block 510 begins a test-run cycle through blocks 510 to 550 for a duration equal to the test-run time limit T M .
  • the test-run time limit T M is a preset value and may be chosen, e.g., to provide a sufficient period to gather a statistically useful set of data, as described below.
  • the controller 34 calculates the predicted position (x targ P )(s), y targ P (s)) of the target vehicle 32 at the future time s.
  • the controller 34 may use known methods, including the quadratic polynomial method using the coefficients determined above. For example, the controller 34 may use this set of equations:
  • the controller 34 may only calculate the predicted lateral position x targ P (s) relative to the host vehicle 30 .
  • the block 515 begins a prediction cycle through blocks 515 to 540 for a duration equal to a prediction time limit T P .
  • the prediction time limit T P is a preset value and may depend on, e.g., a sensitivity of the sensors 36 or a processing speed of the controller 34 .
  • If the quadratic polynomial method is used, then the error metric ⁇ becomes: ⁇ ( t,s )
  • the controller 34 calculates the error bound ⁇ for confidence intervals of the predicted position at the future time of the target vehicle 32 relative laterally to the current position and orientation of the host vehicle 30 at the current time based on the current position and velocity of the target vehicle 32 and the future time.
  • a position of the target vehicle 32 “relative laterally to” a position of the host vehicle 30 refers to the lateral position x relative to the host vehicle 30 , that is, the host-vehicle-centered X-coordinate.
  • the error bound may be further based on the current velocity and yaw rate of the host vehicle 30 and the relative heading angle between the host and target vehicles 30 , 32 .
  • the error bound ⁇ may have a linear relationship with the future time s.
  • the control variables b(t) and c(t) are determined from these equations using the motion variables as inputs:
  • the error bound ⁇ is based only on inputs consisting of the current position and velocity of the target vehicle 32 , the future time, the current velocity and yaw rate of the host vehicle 30 , and the relative heading angle between the host and target vehicles 30 , 32 .
  • the controller 34 calculates an itemized quality measure Q item for the prediction time s.
  • the itemized quality measure Q item is equal to the lesser of 1 and a ratio of the error bound ⁇ and the error metric ⁇ ; in other words, the itemized quality measure Q item is equal to 1 if the error bound ⁇ is at least as great as the error metric ⁇ :
  • the controller 34 increments the prediction time s by an incremental prediction time step ⁇ T P .
  • the prediction time s is changed to s+ ⁇ T P .
  • the incremental prediction time step ⁇ T P is a preset value for a duration of each loop of the prediction cycle.
  • the incremental prediction time step ⁇ T P may depend on, e.g., a sensitivity of the sensors 36 or a processing speed of the controller 34 .
  • the controller 34 determines whether the prediction cycle has not yet reached the prediction time limit T P . In other words, the controller 34 determines whether the prediction time s is less than or equal to t+T P . If the prediction time s is less than or equal to t+T P , then the process 500 proceeds back to the block 515 , and the controller 34 performs another loop of the prediction cycle of the blocks 515 to 540 with the same time t and the updated prediction time s from the block 535 .
  • the controller 34 increments the time t by the incremental time step ⁇ t M .
  • the time t is changed to t+ ⁇ t M .
  • the incremental prediction time step ⁇ t M is a preset value for a duration of each loop of the test-run cycle.
  • the incremental prediction time step ⁇ t M may depend on, e.g., a sensitivity of the sensors 36 or a processing speed of the controller 34 .
  • the controller 34 determines whether the test-run cycle has not yet reached the test-run time limit T M . In other words, the controller 34 determines whether the time t is less than or equal to t 0 +T M . If the time t is less than or equal to t 0 +T M , then the process proceeds back to the block 510 , and the controller 34 performs another loop of the test-run cycle of the blocks 510 to 550 with the updated time t from the block 545 .
  • the controller 34 calculates the minimum itemized quality measure Q test , which is the lowest itemized quality measure Q item over the test run:
  • FIG. 6 is a process flow diagram illustrating the exemplary process 600 for operating the host vehicle 30 using target path prediction.
  • the process 600 begins in a block 605 .
  • the controller 34 measures a set of motion variables for the host vehicle 30 and the target vehicle 32 , as described above with respect to the block 415 .
  • the controller 34 generates coefficients to calculate the predicted position of the target vehicle 32 at the future time using a quadratic polynomial method, as described with respect to the block 420 .
  • the controller 34 initializes the prediction time s to the value of the time t.
  • the controller 34 calculates the predicted position (x targ P (s), y targ P (s)) of the target vehicle 32 at the future time s, as described above with respect to the block 515 .
  • the controller 34 calculates an error bound ⁇ for confidence intervals of the predicted position (x targ P (s), y targ P (s)) at the future time s of the target vehicle 32 relative laterally to the current position and orientation of the host vehicle 30 at the current time, as described above with respect to the block 525 .
  • the controller 34 calculates an error-bounded predicted position (x targ P,err (s), y targ P,err (s)) at the future time s. Specifically, the controller 34 may apply the error bound ⁇ to the lateral predicted position x targ P (s) relative to the host vehicle 30 .
  • the error-bounded predicted position is given by these equations:
  • the controller 34 determines an autonomous operation to perform for driving the host vehicle 30 based on the error-bounded predicted position, that is, based on the predicted position and the error bound.
  • An “autonomous operation” is an action by one or more of the steering system 40 , the propulsion 42 , and the brakes 44 upon instruction by the controller 34 , for example, activating the brakes 44 with a given force for a given period of time, changing a steering angle by a given number of degrees, etc.
  • the controller 34 may control a vehicle subsystem to evade the target vehicle 32 based on the predicted position and the error bound.
  • Vehicle subsystems include, for example, the steering system 40 , propulsion 42 , and brakes 44 . Controlling the vehicle subsystem may include the steering system 40 to turn, instructing the propulsion 42 to accelerate, and/or instructing the brakes 44 to brake.
  • the controller 34 increments the prediction time s by the incremental prediction time step ⁇ T P , as described above with respect to the block 535 .
  • the controller 34 determines whether the prediction cycle has not yet reached the prediction time limit T P . In other words, the controller 34 determines whether the prediction time s is less than or equal to t+T P . If the prediction time s is less than or equal to t+T P , then the process 600 proceeds back to the block 620 , and the controller 34 performs another loop of the blocks 620 to 645 with the same time t and the updated prediction time s from the block 640 . If the prediction time s is greater than t+T P , then the process 600 ends.
  • the computing systems and/or devices described may employ any of a number of computer operating systems, including, but by no means limited to, versions and/or varieties of the Ford Sync® application, AppLink/Smart Device Link middleware, the Microsoft Automotive® operating system, the Microsoft Windows® operating system, the Unix operating system (e.g., the Solaris® operating system distributed by Oracle Corporation of Redwood Shores, Calif.), the AIX UNIX operating system distributed by International Business Machines of Armonk, N.Y., the Linux operating system, the Mac OSX and iOS operating systems distributed by Apple Inc. of Cupertino, Calif., the BlackBerry OS distributed by Blackberry, Ltd. of Waterloo, Canada, and the Android operating system developed by Google, Inc.
  • the Microsoft Automotive® operating system e.g., the Microsoft Windows® operating system distributed by Oracle Corporation of Redwood Shores, Calif.
  • the Unix operating system e.g., the Solaris® operating system distributed by Oracle Corporation of Redwood Shores, Calif.
  • the AIX UNIX operating system distributed by International Business Machine
  • computing devices include, without limitation, an on-board vehicle computer, a computer workstation, a server, a desktop, notebook, laptop, or handheld computer, or some other computing system and/or device.
  • Computing devices generally include computer-executable instructions, where the instructions may be executable by one or more computing devices such as those listed above.
  • Computer executable instructions may be compiled or interpreted from computer programs created using a variety of programming languages and/or technologies, including, without limitation, and either alone or in combination, JavaTM, C, C++, Matlab, Simulink, Stateflow, Visual Basic, Java Script, Perl, HTML, etc. Some of these applications may be compiled and executed on a virtual machine, such as the Java Virtual Machine, the Dalvik virtual machine, or the like.
  • a processor receives instructions, e.g., from a memory, a computer readable medium, etc., and executes these instructions, thereby performing one or more processes, including one or more of the processes described herein.
  • Such instructions and other data may be stored and transmitted using a variety of computer readable media.
  • a file in a computing device is generally a collection of data stored on a computer readable medium, such as a storage medium, a random access memory, etc.
  • a computer-readable medium includes any non-transitory (e.g., tangible) medium that participates in providing data (e.g., instructions) that may be read by a computer (e.g., by a processor of a computer).
  • a medium may take many forms, including, but not limited to, non-volatile media and volatile media.
  • Non-volatile media may include, for example, optical or magnetic disks and other persistent memory.
  • Volatile media may include, for example, dynamic random access memory (DRAM), which typically constitutes a main memory.
  • Such instructions may be transmitted by one or more transmission media, including coaxial cables, copper wire and fiber optics, including the wires that comprise a system bus coupled to a processor of a ECU.
  • Computer-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, any other magnetic medium, a CD-ROM, DVD, any other optical medium, punch cards, paper tape, any other physical medium with patterns of holes, a RAM, a PROM, an EPROM, a FLASH-EEPROM, any other memory chip or cartridge, or any other medium from which a computer can read.
  • Databases, data repositories or other data stores described herein may include various kinds of mechanisms for storing, accessing, and retrieving various kinds of data, including a hierarchical database, a set of files in a file system, an application database in a proprietary format, a relational database management system (RDBMS), etc.
  • Each such data store is generally included within a computing device employing a computer operating system such as one of those mentioned above, and are accessed via a network in any one or more of a variety of manners.
  • a file system may be accessible from a computer operating system, and may include files stored in various formats.
  • An RDBMS generally employs the Structured Query Language (SQL) in addition to a language for creating, storing, editing, and executing stored procedures, such as the PL/SQL language mentioned above.
  • SQL Structured Query Language
  • system elements may be implemented as computer-readable instructions (e.g., software) on one or more computing devices (e.g., servers, personal computers, etc.), stored on computer readable media associated therewith (e.g., disks, memories, etc.).
  • a computer program product may comprise such instructions stored on computer readable media for carrying out the functions described herein.

Landscapes

  • Engineering & Computer Science (AREA)
  • Transportation (AREA)
  • Mechanical Engineering (AREA)
  • Automation & Control Theory (AREA)
  • Chemical & Material Sciences (AREA)
  • Combustion & Propulsion (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Game Theory and Decision Science (AREA)
  • Medical Informatics (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Business, Economics & Management (AREA)
  • Remote Sensing (AREA)
  • Control Of Driving Devices And Active Controlling Of Vehicle (AREA)
  • Traffic Control Systems (AREA)
  • Steering Control In Accordance With Driving Conditions (AREA)
  • Regulating Braking Force (AREA)
  • Feedback Control In General (AREA)

Abstract

A controller in a host vehicle includes a processor and a memory storing processor-executable instructions. The processor is programmed to calculate an error bound for confidence intervals of a predicted position at a future time of a target vehicle relative laterally to a current position and orientation of the host vehicle at a current time based on a current position and velocity of the target vehicle and the future time.

Description

BACKGROUND
Autonomous vehicles have the ability to operate without the intervention of a human operator, e.g., driver, that is, a vehicle controller makes decisions about accelerating, braking, and/or steering the vehicle. A vehicle may be fully autonomous or semi-autonomous. A semi-autonomous vehicle may be autonomous only in particular situations, for example, highway driving or parallel parking, or with respect to certain vehicle subsystems, for example, braking but not acceleration or steering.
An autonomous vehicle may include sensors for tracking an external environment surrounding the vehicle. Some types of sensors are radar sensors, scanning laser range finders, light detection and ranging (LIDAR) devices, and image processing sensors such as cameras. The vehicle controller is in communication with the sensors and uses output from the sensors to analyze the external environment, for example, defining features of a surrounding landscape, detecting roads and lanes of roads on the landscape, interpreting signs and signals, and tracking and classifying objects in the environment such as vehicles, cyclists, and pedestrians. For example, a vehicle controller may classify whether a detected object is another vehicle and provide state information about the other vehicle, such as location, speed, and heading.
The vehicle controller uses target path prediction to predict where another vehicle will travel. The vehicle controller uses the predicted path of the other vehicle to make decisions affecting operation of the vehicle. Thus, inaccuracies and failures of current path prediction technologies are problematic. There is opportunity to improve technology for predicting paths of objects such as target vehicles.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a block diagram of a host vehicle.
FIG. 2 is a process flow diagram of a process for implementing target path prediction on the host vehicle.
FIG. 3 is a process flow diagram of a process for disabling a feature of the host vehicle based on a quality of the target path prediction.
FIG. 4 is a process flow diagram of a process for running a test to determine the quality of the target path prediction.
FIG. 5 is a process flow diagram of a process for determining the quality of the target path prediction during a test run.
FIG. 6 is a process flow diagram of a process for operating the host vehicle using target path prediction.
FIG. 7 is a diagram of an interaction between the host vehicle and a target vehicle.
DETAILED DESCRIPTION
With reference to the Figures, wherein like numerals indicate like parts throughout the several views, a controller 34 in a host vehicle 30 includes a processor and a memory storing processor-executable instructions. The processor is programmed to calculate an error bound for confidence intervals of a predicted position at a future time of a target vehicle 32 relative laterally to a current position and orientation of the host vehicle 30 at a current time based on a current position and velocity of the target vehicle 32 and the future time.
Use of the error bound makes target path prediction, and hence autonomous operation of the host vehicle 30, more robust because the controller 34 makes decisions based on a range of the most likely predicted positions for the target vehicle 32. Furthermore, calculation of the error bound is adaptive in that the calculation updates continuously as the host vehicle 30 travels.
With reference to FIG. 1, the host vehicle 30 may be an autonomous vehicle. The controller 34, sometimes referred to as the “virtual driver,” may be capable of operating the vehicle independently of the intervention of a human driver, to a greater or a lesser degree, i.e., fully autonomously and/or in a semi-autonomous mode. The controller 34 may be programmed to operate a steering system 40, a propulsion 42, brakes 44, and/or other vehicle systems. For purposes of this disclosure, fully autonomous means that each of the steering system 40, the propulsion 42, and the brakes 44 of the host vehicle 30 are controlled by the controller 34; a semi-autonomous mode is one in which one or two of the steering system 40, the propulsion 42, and the brakes 44 are operated by the controller 34.
The controller 34 carries out various operations, including as described herein. The controller 34 is a computing device that generally includes a processor and a memory, the memory including one or more forms of computer-readable media, and storing instructions executable by the processor for performing various operations, including as disclosed herein. The memory of the controller 34 further generally stores remote data received via various communications mechanisms; e.g., the controller 34 is generally configured for communications on a controller area network (CAN) bus or the like, and/or for using other wired or wireless protocols, e.g., Bluetooth, etc. The controller 34 may also have a connection to an onboard diagnostics connector (OBD-II). Via a vehicle network using Ethernet, WiFi, the CAN bus, Local Interconnect Network (LIN), and/or other wired or wireless mechanisms, the controller 34 may transmit messages to various devices in the vehicle and/or receive messages from the various devices, e.g., controllers, actuators, sensors, etc., e.g., controllers and sensors as discussed herein. For example, the controller 34 may receive data from sensors 36. Although one controller 34 is shown in FIG. 1 for ease of illustration, it is to be understood that the controller 34 could include, and various operations described herein could be carried out by, one or more computing devices.
The controller 34 may transmit signals through a communications network 38 such as a controller area network (CAN) bus, Ethernet, Local Interconnect Network (LIN), and/or by any other wired or wireless communications network. The controller 34 may be in communication with the sensors 36, the steering system 40, the propulsion 42, the brakes 44, and other vehicle subsystems and/or features.
The host vehicle 30 may include the sensors 36. The sensors 36 may detect internal states of the vehicle, for example, wheel speed, wheel orientation, and engine and transmission variables. The sensors 36 may detect the position or orientation of the vehicle, for example, global positioning system (GPS) sensors; accelerometers such as piezo-electric or microelectromechanical systems (MEMS); gyroscopes such as rate, ring laser, or fiber-optic gyroscopes; inertial measurements units (IMU); and magnetometers. The sensors 36 may detect an environment external to a body of the host vehicle 30, for example, the sensors 36 can include one or more of radar sensors, scanning laser range finders, light detection and ranging (LIDAR) devices, and image processing sensors such as cameras. The sensors 36 may include communications devices, for example, vehicle-to-infrastructure (V2I) or vehicle-to-vehicle (V2V) devices.
The steering system 40 is typically a known vehicle steering subsystem and controls the turning of wheels of the host vehicle 30. The steering system 40 is in communication with and receives input from a steering wheel and the controller 34. The steering system 40 may be a rack-and-pinion system with electric power-assisted steering, a steer-by-wire system, as are both known in the art, or any other suitable system.
The propulsion 42 of the host vehicle 30 generates energy and translates the energy into motion of the host vehicle 30. The propulsion 42 may be a known vehicle propulsion subsystem, for example, a conventional powertrain including an internal-combustion engine coupled to a transmission that transfers rotational motion to wheels; an electric powertrain including batteries, an electric motor, and a transmission that transfers rotational motion to the wheels; a hybrid powertrain including elements of the conventional powertrain and the electric powertrain; or any other type of propulsion. The propulsion 42 is in communication with and receives input from the controller 34 and from a human driver. The human driver may control the propulsion 42 via, e.g., an accelerator pedal and/or a gear-shift lever.
The brakes 44 are typically a known vehicle braking subsystem and are provided to resist the motion of the host vehicle 30 to thereby slow and/or stop the vehicle. The brakes 44 may be friction brakes such as disc brakes, drum brakes, band brakes, and so on; regenerative brakes; any other suitable type of brakes; or a combination. The brakes 44 are in communication with and receive input from the controller 34 and a human driver. The human driver may control the brakes 44 via, e.g., a brake pedal.
With reference to FIG. 7, the exemplary processes described below pertain to target path prediction. The host vehicle 30 may interact with the target vehicle 32 while each are traveling. “Target path prediction” means predicting, by the host vehicle 30, a path P along which the target vehicle 32 will travel. An actual path A by the target vehicle 32 may differ from the predicted path P. Motion variables shown in FIG. 7 describing the travel of the vehicles 30, 32 are described below with respect to a block 415 in a process 400.
FIG. 2 is a process flow diagram illustrating an exemplary process 200 for implementing target path prediction on the host vehicle 30. The process 200 may be used by, for example, a system designer testing and implementing an algorithm for target path prediction on a type of vehicle. The process 200 begins by performing an exemplary process 400 for running a test to determine the quality of the target path prediction, shown in FIG. 4 and described below. The process 400 returns a quality measure Q of the target path prediction to the controller 34.
Next, after receiving input from the process 400, in a decision block 205, the controller 34 determines whether the quality measure Q is above a quality threshold Qth. The quality threshold is a value chosen to ensure a sufficiently robust performance of the target path prediction.
If the quality measure Q is not above the quality threshold Qth, next, in a block 210, a system designer adjusts a path prediction algorithm for target path prediction based on the quality measure Q. The system designer may adjust parameters for the target path prediction algorithm, such as the quadratic polynomial method, as is known (described below with respect to a block 420 of the process 400 and a block 515 of a process 500). Alternatively or additionally, the system designer may adjust parameters for the error bounds aspect of the path prediction algorithm (described below with respect to a block 525 of the process 500), for example, a lateral offset bound x targ for the target vehicle 32, a longitudinal offset bound y targ for the target vehicle 32, a location bound b for the target vehicle 32, a velocity bound v targ for the target vehicle 32, a velocity bound v host for the host vehicle 30, a yaw rate bound w host for the host vehicle 30, and/or a default initialization values σx targ , σy targ , σv targ , σv host , and σw host .
If, in the decision block 205, the quality measure Q is above the quality threshold Qth, the controller 34 performs an exemplary process 600 for operating the host vehicle 30 using target path prediction, as shown in FIG. 6 and described below. After the process 600, the process 200 ends.
FIG. 3 is a process flow diagram illustrating an exemplary process 300 for disabling a feature of the host vehicle 30 based on a quality of the target path prediction. The process 300 may be performed, for example, by the controller 34 in the host vehicle 30. The process 300 begins by performing the exemplary process 400 for running a test to determine the quality of the target path prediction, shown in FIG. 4 and described below. The process 400 returns the quality measure Q of the target path prediction to the controller 34.
Next, after receiving input from the process 400, in a decision block 305, the controller 34 determines whether the quality measure Q is above a quality threshold Qth, as described above with respect to the decision block 205.
If the quality measure Q is not above the quality threshold Qth, next, in a block 310, the controller 34 disables a feature relying on an error bound ε (described below with respect to the block 525 of the process 500) based on the quality measure Q. Features that may rely on the error bound ε may include autonomous or semi-autonomous operation of the host vehicle 30, that is, operation of one or more of the steering system 40, propulsion 42, and brakes 44, such as automated braking or automated steering. Other features that may rely on the error bound ε may include features that may anticipate movement of the host vehicle 30, for example, automatically turning headlights.
If, in the decision block 305, the quality measure Q is above the quality threshold Qth, the controller 34 performs the exemplary process 600 for operating the host vehicle 30 using target path prediction, as shown in FIG. 6 and described below. After the process 600, the process 300 ends.
FIG. 4 is a process flow diagram illustrating the exemplary process 400 for running a test to determine the quality of the target path prediction. The process 400 begins in a block 405, in which the controller 34 initializes an index value k to 1. The index value k counts a number of test runs each of a duration equal to a test-run time limit TM.
Next, in a block 410, the controller 34 performs the kth test run. Performing the test run involves driving the host vehicle 30 for a duration TM from an initial time t0 to a final time t0+TM. Driving the host vehicle 30 may be performed autonomously by the controller 34, semi-autonomously by the controller 34 and a human driver, or manually by the human driver. The block 410 begins an experiment cycle through blocks 410 to 430 of a number K of test runs. The number K of test runs is a preset or user-specified value and may be chosen, e.g., to provide a sufficient period to gather a statistically useful set of data, as described below.
Next, during the test run, in a block 415, the controller 34 measures a set of motion variables for the host vehicle 30 and the target vehicle 32. The motion variables may be measured in a known coordinate system such as an objective over-the-ground coordinate system, a host-vehicle-centered coordinate system, or any other suitable coordinate system, and the motion variables may be converted from one coordinate system to another before further use. For example, the controller 34 may measure this set of motion variables in over-the-ground coordinates:
{ X targ M ( t ) , Y targ M ( t ) , ψ targ M ( t ) , v targ M ( t ) , a targ M ( t ) } t 0 t t 0 + T M { X host M ( t ) , Y host M ( t ) , ψ host M ( t ) , v host M ( t ) , a host M ( t ) } t 0 t t 0 + T M
in which the subscript host denotes a variable related to the host vehicle 30, the subscript targ denotes a variable related to the target vehicle 32, the superscript M denotes a measured variable, X denotes an over-the-ground X-coordinate, Y denotes an over-the-ground Y-coordinate, ψ denotes an over-the-ground heading angle, v denotes a speed, and a denotes an acceleration. The over-the-ground motion variables may then be converted into the host-vehicle-centered coordinate system:
{ x targ M ( t ) , y targ M ( t ) , v targ M ( t ) , a targ M ( t ) , w host M ( t ) , v host M ( t ) , a host M ( t ) , a M ( t ) } t 0 t t 0 + T M
in which x denotes a lateral position relative to the host vehicle 30, that is, a host-vehicle-centered X-coordinate; y denotes a longitudinal position relative to the host vehicle 30, that is, a host-vehicle-centered Y-coordinate; w denotes a yaw rate; and a denotes a relative heading angle between the host vehicle 30 and the target vehicle 32. The motion variables in the host-vehicle-centered coordinate system are shown in FIG. 7. The motion variables continue to be recorded as the host vehicle 30 moves and as the target vehicle 32 moves along an actual path A.
Next, during the test run, in the block 420, the controller 34 generates coefficients to calculate the predicted position of the target vehicle 32 at the future time using a quadratic polynomial method. A predicted path P, which is a sequence of predicted positions of the target vehicle 32 in the host-vehicle-centered coordinate system fixed at time t, is estimated as a second-order polynomial:
x targ P =a 0(t)+a 1(t)*y targ P +a 2(t)*(y targ P)2
in which the superscript P denotes a predicted variable; and a0(t), a1(t), and a2(t) are the coefficients generated by the quadratic polynomial method, as is known. Another method of target path prediction besides the quadratic polynomial method, such as a Kalman filter-based motion-prediction method, may be used.
Next, during the test run, the controller 34 performs an exemplary process 500 for determining the quality of the target path prediction during the test run, as shown in FIG. 5 and described below. The process 500 returns a quality measure Qtest of the target path prediction during the test run to the controller 34.
Next, after receiving input from the process 500, in a block 425, the controller 34 increments the index value k by 1, so the index value k is equal to k+1.
Next, in a decision block 430, the controller 34 determines whether the index value k has reached or exceeded the number K of test runs. If the index value k is still below the number K of test runs, then the process 400 proceeds back to the block 410, and the controller 34 performs another loop of the experiment cycle of the blocks 410 to 430 with the update index value k from the block 425.
If the index value k is greater than or equal to the number K of test runs, next, in a block 435, the controller 34 calculates the quality measure Q equal to a minimum value during the period of the experiment of a lesser value of 1 and a ratio of the error bound ε and the error metric δ, as described below with respect to the process 500. In other words, the controller 34 calculates the lowest minimum itemized quality measure Qtest:
Q = min k [ 1 , K ] Q test
After the block 435, the process 400 ends.
FIG. 5 is a process flow diagram illustrating the exemplary process 500 for determining the quality of the target path prediction during the test run. The process 500 begins in a block 505, in which the controller 34 initializes a time variable t to an initial value t0.
Next, in a block 510, the controller 34 initializes a prediction time variable s to the value of the time variable t. The time variable t tracks the current time, and the prediction time variable s tracks the future time for predictions. The block 510 begins a test-run cycle through blocks 510 to 550 for a duration equal to the test-run time limit TM. The test-run time limit TM is a preset value and may be chosen, e.g., to provide a sufficient period to gather a statistically useful set of data, as described below.
Next, in a block 515, the controller 34 calculates the predicted position (xtarg P)(s), ytarg P(s)) of the target vehicle 32 at the future time s. The controller 34 may use known methods, including the quadratic polynomial method using the coefficients determined above. For example, the controller 34 may use this set of equations:
x targ P ( s ) = a 0 ( t ) + a 1 ( t ) * y targ P ( s ) + a 2 ( t ) * ( y targ P ( s ) ) 2 y targ P ( s ) = y targ M ( t ) + v targ M ( t ) * sin ( a M ( t ) + π 2 ) * ( s - t ) + 1 2 a targ M ( t ) * sin ( a M ( t ) + π 2 ) * ( s - t ) 2
Alternatively, the controller 34 may only calculate the predicted lateral position xtarg P(s) relative to the host vehicle 30. The block 515 begins a prediction cycle through blocks 515 to 540 for a duration equal to a prediction time limit TP. The prediction time limit TP is a preset value and may depend on, e.g., a sensitivity of the sensors 36 or a processing speed of the controller 34.
Next, in a block 520, the controller 34 calculates an error metric δ equal to an absolute value of a difference of the predicted position and a measured position at the future time of the target vehicle 32 relative laterally to the current position and orientation of the host vehicle 30, as described in the following equation:
δ(t,s)=|x targ M(s)−x targ P(s)|
If the quadratic polynomial method is used, then the error metric δ becomes:
δ(t,s)=|x targ M(s)−(a 0(t)+a 1(t)*y targ M(s)+a 2(t)*(y targ M(s))2)|
In both equations for the error metric δ, the measured position (xtarg M(s), ytarg M(s)) is measured at the prediction time s in the host-vehicle-centered coordinate system fixed at time t.
Next, in the block 525, the controller 34 calculates the error bound ε for confidence intervals of the predicted position at the future time of the target vehicle 32 relative laterally to the current position and orientation of the host vehicle 30 at the current time based on the current position and velocity of the target vehicle 32 and the future time. A position of the target vehicle 32 “relative laterally to” a position of the host vehicle 30 refers to the lateral position x relative to the host vehicle 30, that is, the host-vehicle-centered X-coordinate. The error bound may be further based on the current velocity and yaw rate of the host vehicle 30 and the relative heading angle between the host and target vehicles 30, 32. The error bound ε may have a linear relationship with the future time s. The error bound ε may be determined from this equation:
ε(t,s)=b(t)*(s−t)+c(t)
in which b(t) and c(t) are control variables for the error bound ε. The control variables b(t) and c(t) are determined from these equations using the motion variables as inputs:
b ( t ) = 1 1 + cos ( a M ( t ) ) * ( x targ M ( t ) + σ x targ x _ targ ) * ( y targ M ( t ) + σ y targ y _ targ ) * ( v targ M ( t ) + σ v targ v _ targ ) * ( v host M ( t ) + σ v host v _ host ) * ( w host M ( t ) + σ w host w _ host ) c ( t ) = ( x targ M ( t ) ) 2 + ( y targ M ( t ) ) 2 b _
in which ε is the error bound; t is the current time; s is the future time; (xtarg M(t), ytarg M(t)) is the current position of the target vehicle 32 in a host-vehicle-centered coordinate system; vtarg M(t) is the current velocity of the target vehicle 32; vhost M(t) is the current velocity of the host vehicle 30; whost M(t) is the current yaw rate of the host vehicle 30; αM(t) is the relative heading angle between the host and target vehicles 30, 32; x targ is a lateral offset bound for the target vehicle 32; y targ is a longitudinal offset bound for the target vehicle 32; b is a location bound for the target vehicle 32; v targ is a velocity bound for the target vehicle 32; v host is a velocity bound for the host vehicle 30; W host is a yaw rate bound for the host vehicle 30; and σx targ , σy targ , σv targ , σv host , and σw host are default initialization values. As calculated in the above equation, the error bound ε is based only on inputs consisting of the current position and velocity of the target vehicle 32, the future time, the current velocity and yaw rate of the host vehicle 30, and the relative heading angle between the host and target vehicles 30, 32.
Next, in a block 530, the controller 34 calculates an itemized quality measure Qitem for the prediction time s. The itemized quality measure Qitem is equal to the lesser of 1 and a ratio of the error bound ε and the error metric δ; in other words, the itemized quality measure Qitem is equal to 1 if the error bound ε is at least as great as the error metric δ:
Q item ( t , s ) = { 1 if δ ( t , s ) ɛ ( t , s ) ɛ ( t , s ) δ ( t , s ) otherwise
Next, in a block 535, the controller 34 increments the prediction time s by an incremental prediction time step ΔTP. In other words, the prediction time s is changed to s+ΔTP. The incremental prediction time step ΔTP is a preset value for a duration of each loop of the prediction cycle. The incremental prediction time step ΔTP may depend on, e.g., a sensitivity of the sensors 36 or a processing speed of the controller 34.
Next, in a decision block 540, the controller 34 determines whether the prediction cycle has not yet reached the prediction time limit TP. In other words, the controller 34 determines whether the prediction time s is less than or equal to t+TP. If the prediction time s is less than or equal to t+TP, then the process 500 proceeds back to the block 515, and the controller 34 performs another loop of the prediction cycle of the blocks 515 to 540 with the same time t and the updated prediction time s from the block 535.
If the prediction time s is greater than t+TP, next, in a block 545, the controller 34 increments the time t by the incremental time step ΔtM. In other words, the time t is changed to t+ΔtM. The incremental prediction time step ΔtM is a preset value for a duration of each loop of the test-run cycle. The incremental prediction time step ΔtM may depend on, e.g., a sensitivity of the sensors 36 or a processing speed of the controller 34.
Next, in a decision block 550, the controller 34 determines whether the test-run cycle has not yet reached the test-run time limit TM. In other words, the controller 34 determines whether the time t is less than or equal to t0+TM. If the time t is less than or equal to t0+TM, then the process proceeds back to the block 510, and the controller 34 performs another loop of the test-run cycle of the blocks 510 to 550 with the updated time t from the block 545.
If the time t is greater than t0+TM, next, in a block 555, the controller 34 calculates the minimum itemized quality measure Qtest, which is the lowest itemized quality measure Qitem over the test run:
Q test = min t [ 0 , t + T M ] ( min s [ t , t + T P ] Q item ( t , s ) )
After the block 555, the process 500 ends.
FIG. 6 is a process flow diagram illustrating the exemplary process 600 for operating the host vehicle 30 using target path prediction. The process 600 begins in a block 605. The controller 34 measures a set of motion variables for the host vehicle 30 and the target vehicle 32, as described above with respect to the block 415.
Next, in a block 610, the controller 34 generates coefficients to calculate the predicted position of the target vehicle 32 at the future time using a quadratic polynomial method, as described with respect to the block 420.
Next, in a block 615, the controller 34 initializes the prediction time s to the value of the time t.
Next, in a block 620, the controller 34 calculates the predicted position (xtarg P(s), ytarg P(s)) of the target vehicle 32 at the future time s, as described above with respect to the block 515.
Next, in a block 625, the controller 34 calculates an error bound ε for confidence intervals of the predicted position (xtarg P(s), ytarg P(s)) at the future time s of the target vehicle 32 relative laterally to the current position and orientation of the host vehicle 30 at the current time, as described above with respect to the block 525.
Next, in a block 630, the controller 34 calculates an error-bounded predicted position (xtarg P,err(s), ytarg P,err(s)) at the future time s. Specifically, the controller 34 may apply the error bound ε to the lateral predicted position xtarg P(s) relative to the host vehicle 30. The error-bounded predicted position is given by these equations:
{ x targ P , err ( s ) = x targ P ( s ) ± ɛ ( t , s ) y targ P , err ( s ) = y targ P ( s )
Next, in a block 635, the controller 34 determines an autonomous operation to perform for driving the host vehicle 30 based on the error-bounded predicted position, that is, based on the predicted position and the error bound. An “autonomous operation” is an action by one or more of the steering system 40, the propulsion 42, and the brakes 44 upon instruction by the controller 34, for example, activating the brakes 44 with a given force for a given period of time, changing a steering angle by a given number of degrees, etc. Specifically, if the host vehicle 30 may cross a region specified by the error-bounded predicted position, then the controller 34 may control a vehicle subsystem to evade the target vehicle 32 based on the predicted position and the error bound. Vehicle subsystems include, for example, the steering system 40, propulsion 42, and brakes 44. Controlling the vehicle subsystem may include the steering system 40 to turn, instructing the propulsion 42 to accelerate, and/or instructing the brakes 44 to brake.
Next, in a block 640, the controller 34 increments the prediction time s by the incremental prediction time step ΔTP, as described above with respect to the block 535.
Next, in a decision block 645, the controller 34 determines whether the prediction cycle has not yet reached the prediction time limit TP. In other words, the controller 34 determines whether the prediction time s is less than or equal to t+TP. If the prediction time s is less than or equal to t+TP, then the process 600 proceeds back to the block 620, and the controller 34 performs another loop of the blocks 620 to 645 with the same time t and the updated prediction time s from the block 640. If the prediction time s is greater than t+TP, then the process 600 ends.
In general, the computing systems and/or devices described may employ any of a number of computer operating systems, including, but by no means limited to, versions and/or varieties of the Ford Sync® application, AppLink/Smart Device Link middleware, the Microsoft Automotive® operating system, the Microsoft Windows® operating system, the Unix operating system (e.g., the Solaris® operating system distributed by Oracle Corporation of Redwood Shores, Calif.), the AIX UNIX operating system distributed by International Business Machines of Armonk, N.Y., the Linux operating system, the Mac OSX and iOS operating systems distributed by Apple Inc. of Cupertino, Calif., the BlackBerry OS distributed by Blackberry, Ltd. of Waterloo, Canada, and the Android operating system developed by Google, Inc. and the Open Handset Alliance, or the QNX® CAR Platform for Infotainment offered by QNX Software Systems. Examples of computing devices include, without limitation, an on-board vehicle computer, a computer workstation, a server, a desktop, notebook, laptop, or handheld computer, or some other computing system and/or device.
Computing devices generally include computer-executable instructions, where the instructions may be executable by one or more computing devices such as those listed above. Computer executable instructions may be compiled or interpreted from computer programs created using a variety of programming languages and/or technologies, including, without limitation, and either alone or in combination, Java™, C, C++, Matlab, Simulink, Stateflow, Visual Basic, Java Script, Perl, HTML, etc. Some of these applications may be compiled and executed on a virtual machine, such as the Java Virtual Machine, the Dalvik virtual machine, or the like. In general, a processor (e.g., a microprocessor) receives instructions, e.g., from a memory, a computer readable medium, etc., and executes these instructions, thereby performing one or more processes, including one or more of the processes described herein. Such instructions and other data may be stored and transmitted using a variety of computer readable media. A file in a computing device is generally a collection of data stored on a computer readable medium, such as a storage medium, a random access memory, etc.
A computer-readable medium (also referred to as a processor-readable medium) includes any non-transitory (e.g., tangible) medium that participates in providing data (e.g., instructions) that may be read by a computer (e.g., by a processor of a computer). Such a medium may take many forms, including, but not limited to, non-volatile media and volatile media. Non-volatile media may include, for example, optical or magnetic disks and other persistent memory. Volatile media may include, for example, dynamic random access memory (DRAM), which typically constitutes a main memory. Such instructions may be transmitted by one or more transmission media, including coaxial cables, copper wire and fiber optics, including the wires that comprise a system bus coupled to a processor of a ECU. Common forms of computer-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, any other magnetic medium, a CD-ROM, DVD, any other optical medium, punch cards, paper tape, any other physical medium with patterns of holes, a RAM, a PROM, an EPROM, a FLASH-EEPROM, any other memory chip or cartridge, or any other medium from which a computer can read.
Databases, data repositories or other data stores described herein may include various kinds of mechanisms for storing, accessing, and retrieving various kinds of data, including a hierarchical database, a set of files in a file system, an application database in a proprietary format, a relational database management system (RDBMS), etc. Each such data store is generally included within a computing device employing a computer operating system such as one of those mentioned above, and are accessed via a network in any one or more of a variety of manners. A file system may be accessible from a computer operating system, and may include files stored in various formats. An RDBMS generally employs the Structured Query Language (SQL) in addition to a language for creating, storing, editing, and executing stored procedures, such as the PL/SQL language mentioned above.
In some examples, system elements may be implemented as computer-readable instructions (e.g., software) on one or more computing devices (e.g., servers, personal computers, etc.), stored on computer readable media associated therewith (e.g., disks, memories, etc.). A computer program product may comprise such instructions stored on computer readable media for carrying out the functions described herein.
In the drawings, the same reference numbers indicate the same elements. Further, some or all of these elements could be changed. With regard to the media, processes, systems, methods, heuristics, etc. described herein, it should be understood that, although the steps of such processes, etc. have been described as occurring according to a certain ordered sequence, such processes could be practiced with the described steps performed in an order other than the order described herein. It further should be understood that certain steps could be performed simultaneously, that other steps could be added, or that certain steps described herein could be omitted. In other words, the descriptions of processes herein are provided for the purpose of illustrating certain embodiments, and should in no way be construed so as to limit the claims.
Accordingly, it is to be understood that the above description is intended to be illustrative and not restrictive. Many embodiments and applications other than the examples provided would be apparent to those of skill in the art upon reading the above description. The scope of the invention should be determined, not with reference to the above description, but should instead be determined with reference to the appended claims, along with the full scope of equivalents to which such claims are entitled. It is anticipated and intended that future developments will occur in the arts discussed herein, and that the disclosed systems and methods will be incorporated into such future embodiments. In sum, it should be understood that the invention is capable of modification and variation and is limited only by the following claims.
All terms used in the claims are intended to be given their plain and ordinary meanings as understood by those skilled in the art unless an explicit indication to the contrary in made herein. In particular, use of the singular articles such as “a,” “the,” “said,” etc. should be read to recite one or more of the indicated elements unless a claim recites an explicit limitation to the contrary.
The disclosure has been described in an illustrative manner, and it is to be understood that the terminology which has been used is intended to be in the nature of words of description rather than of limitation. Many modifications and variations of the present disclosure are possible in light of the above teachings, and the disclosure may be practiced otherwise than as specifically described.

Claims (20)

What is claimed is:
1. A controller comprising a processor and a memory storing processor-executable instructions, wherein the processor is programmed to:
calculate an error bound for confidence intervals of a predicted position of a target vehicle; and
instruct a steering system of a host vehicle to turn based on the error bound;
wherein the predicted position is at a future time;
the predicted position is relative laterally to a current position and orientation of the host vehicle at a current time;
calculating the error bound is based on a current position and velocity of the target vehicle and the future time; and
the error bound is a product of the future time and a value that is constant with respect to the future time.
2. The controller of claim 1, wherein the error bound is further based on a current velocity and yaw rate of the host vehicle.
3. The controller of claim 2, wherein the error bound is further based on a relative heading angle between the host and target vehicles.
4. The controller of claim 3, wherein the error bound has a linear relationship with the future time.
5. The controller of claim 4, wherein the error bound is determined from an equation:

ε(t,s)=b(t)*(s−t)+c(t), in which
b ( t ) = 1 1 + cos ( a M ( t ) ) * ( x targ M ( t ) + σ x targ x _ targ ) * ( y targ M ( t ) + σ y targ y _ targ ) * ( v targ M ( t ) + σ v targ v _ targ ) * ( v host M ( t ) + σ v host v _ host ) * ( w host M ( t ) + σ w host w _ host ) , and c ( t ) = ( x targ M ( t ) ) 2 + ( y targ M ( t ) ) 2 b _ ;
in which
ε is the error bound, t is the current time, s is the future time, (xtarg M(t), ytarg M(t)) is the current position of the target vehicle in a host-vehicle-centered coordinate system, vtarg M(t) is the current velocity of the target vehicle, vhost M(t) is the current velocity of the host vehicle, whost M(t) is the current yaw rate of the host vehicle, αM (t) is the relative heading angle between the host and target vehicles, x targ is a lateral offset bound for the target vehicle, y targ is a longitudinal offset bound for the target vehicle, b is a location bound for the target vehicle, v targ is a velocity bound for the target vehicle, v host is a velocity bound for the host vehicle, w host is a yaw rate bound for the host vehicle, and σx targ , σy targ , σv targ , σv host and σw host are default initialization values.
6. The controller of claim 3, wherein the error bound is based on inputs consisting of the current position and velocity of the target vehicle, the future time, the current velocity and yaw rate of the host vehicle, and the relative heading angle between the host and target vehicles.
7. The controller of claim 1, wherein the processor is further programmed to calculate the predicted position at the future time.
8. The controller of claim 7, wherein the processor is programmed to generate coefficients to calculate the predicted position at the future time using a quadratic polynomial method.
9. The controller of claim 7, wherein the processor is further programmed to control a vehicle subsystem to evade the target vehicle based on the predicted position and the error bound.
10. The controller of claim 9, wherein controlling the vehicle subsystem includes instructing brakes to brake.
11. The controller of claim 9, wherein controlling the vehicle subsystem includes instructing the steering system to turn.
12. The controller of claim 1, wherein the processor is further programmed to calculate an error metric equal to an absolute value of a difference of the predicted position and a measured position of the target vehicle;
wherein the measured position is at the future time; and
the measured position is relative laterally to the current position and orientation of the host vehicle.
13. The controller of claim 12, wherein the processor is further programmed to calculate a quality measure equal to a minimum value of a lesser value of 1 and a ratio of the error bound to the error metric, wherein the minimum value is calculated over a period.
14. The controller of claim 13, wherein the processor is further programmed to disable a feature relying on the error bound based on the quality measure.
15. The controller of claim 14, wherein the feature is one of automated braking, automated steering, and autonomous operation.
16. A method comprising:
calculating an error bound for confidence intervals of a predicted position of a target vehicle; and
instructing a steering system of a host vehicle to turn based on the error bound;
wherein the predicted position is at a future time;
the predicted position is relative laterally to a current position and orientation of the host vehicle at a current time;
calculating the error bound is based on a current position and velocity of the target vehicle and the future time; and
the error bound is a product of the future time and a value that is constant with respect to the future time.
17. The method of claim 16, wherein the error bound is further based on a current velocity and yaw rate of the host vehicle and a relative heading angle between the host and target vehicles.
18. The method of claim 17, further comprising calculating an error metric equal to an absolute value of a difference of the predicted position and a measured position of the target vehicle;
wherein the measured position is at the future time; and
the measured position is relative laterally to the current position and orientation of the host vehicle.
19. The method of claim 18, further comprising calculating a quality measure equal to a minimum value of a lesser value of 1 and a ratio of the error bound to the error metric, wherein the minimum value is calculated over a period.
20. The method of claim 19, further comprising adjusting a path prediction algorithm based on the quality measure.
US15/252,425 2016-08-31 2016-08-31 Autonomous vehicle using path prediction Active 2037-02-03 US10118610B2 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
US15/252,425 US10118610B2 (en) 2016-08-31 2016-08-31 Autonomous vehicle using path prediction
GB1713487.5A GB2556146A (en) 2016-08-31 2017-08-22 Autonomous vehicle using path prediction
RU2017129885A RU2017129885A (en) 2016-08-31 2017-08-24 CONTROLLER AND METHOD
CN201710758086.8A CN107792079B (en) 2016-08-31 2017-08-29 Autonomous vehicle with path prediction
MX2017011127A MX2017011127A (en) 2016-08-31 2017-08-30 Autonomous vehicle using path prediction.
DE102017119952.7A DE102017119952A1 (en) 2016-08-31 2017-08-30 AUTONOMIC VEHICLE WITH COURAGE RECORDING

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US15/252,425 US10118610B2 (en) 2016-08-31 2016-08-31 Autonomous vehicle using path prediction

Publications (2)

Publication Number Publication Date
US20180056995A1 US20180056995A1 (en) 2018-03-01
US10118610B2 true US10118610B2 (en) 2018-11-06

Family

ID=59996755

Family Applications (1)

Application Number Title Priority Date Filing Date
US15/252,425 Active 2037-02-03 US10118610B2 (en) 2016-08-31 2016-08-31 Autonomous vehicle using path prediction

Country Status (6)

Country Link
US (1) US10118610B2 (en)
CN (1) CN107792079B (en)
DE (1) DE102017119952A1 (en)
GB (1) GB2556146A (en)
MX (1) MX2017011127A (en)
RU (1) RU2017129885A (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10732261B1 (en) * 2019-12-31 2020-08-04 Aurora Innovation, Inc. Generating data using radar observation model based on machine learning
US10933890B2 (en) * 2019-01-30 2021-03-02 Toyota Jtdosha Kabushiki Kaisha Driving assistance device
US20210116916A1 (en) * 2019-10-22 2021-04-22 Baidu Usa Llc End dynamics and constraints relaxation algorithm on optimizing an open space trajectory
US11021148B2 (en) * 2019-03-25 2021-06-01 Zoox, Inc. Pedestrian prediction based on attributes
US11351991B2 (en) * 2019-03-25 2022-06-07 Zoox, Inc. Prediction based on attributes
US11351999B2 (en) * 2020-09-16 2022-06-07 Xuan Binh Luu Traffic collision warning device
US11409284B2 (en) * 2019-05-15 2022-08-09 Baidu Usa Llc Relaxation optimization model to plan an open space trajectory for autonomous vehicles
US11467591B2 (en) * 2019-05-15 2022-10-11 Baidu Usa Llc Online agent using reinforcement learning to plan an open space trajectory for autonomous vehicles
US11493926B2 (en) 2019-05-15 2022-11-08 Baidu Usa Llc Offline agent using reinforcement learning to speedup trajectory planning for autonomous vehicles
US11608063B2 (en) * 2018-11-15 2023-03-21 Volvo Car Corporation Vehicle safe stop
US11926299B2 (en) 2020-11-23 2024-03-12 Aptiv Technologies Limited System and method for predicting road collisions with a host vehicle

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA3036819C (en) * 2016-09-29 2023-10-31 Teleste Oyj A method for detecting link status
JP6493364B2 (en) * 2016-11-18 2019-04-03 トヨタ自動車株式会社 Driving assistance device
WO2018216058A1 (en) * 2017-05-22 2018-11-29 三菱電機株式会社 Position estimation device, position estimation method, and position estimation program
US10409286B2 (en) * 2017-07-21 2019-09-10 Ford Global Technologies, Llc Highway detection systems and methods
US11400927B2 (en) * 2018-01-29 2022-08-02 Ford Global Technologies, Llc Collision avoidance and mitigation
CN110962856B (en) * 2018-09-30 2021-07-27 毫末智行科技有限公司 Method and device for determining area of vehicle where environmental target is located
CN109991976B (en) * 2019-03-01 2022-05-13 江苏理工学院 Method for avoiding dynamic vehicle by unmanned vehicle based on standard particle swarm optimization
CN109948289B (en) * 2019-04-01 2020-11-10 清华大学 Vehicle autonomous parking function evaluation method based on maximum global encroachment rate search
JP7227072B2 (en) * 2019-05-22 2023-02-21 日立Astemo株式会社 vehicle controller
CN112346479B (en) * 2020-11-18 2023-08-22 大连海事大学 Unmanned aircraft state estimation method based on centralized Kalman filtering
CN113788031B (en) * 2021-10-19 2023-08-25 广州小鹏自动驾驶科技有限公司 Vehicle speed prediction method and device, electronic equipment and storage medium

Citations (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11175896A (en) 1997-12-05 1999-07-02 Fujitsu Ltd Method and system for preventing collision at intersection, storage medium in which collision at intersection preventing program is stored and intersection device
US20030016143A1 (en) 2001-07-23 2003-01-23 Ohanes Ghazarian Intersection vehicle collision avoidance system
US6675094B2 (en) 2000-09-08 2004-01-06 Raytheon Company Path prediction system and method
JP2005173703A (en) 2003-12-08 2005-06-30 Nissan Motor Co Ltd Intersection collision preventing apparatus
CN1766232A (en) 2005-11-21 2006-05-03 钱红波 Novel traffic layout and control method for city road net and crossing
US20070005609A1 (en) 1997-10-22 2007-01-04 Intelligent Technologies International, Inc. Vehicular Communication Arrangement and Method
US20070030131A1 (en) * 2005-08-02 2007-02-08 Nissan Motor Co., Ltd. Vehicle obstacle verification system
US20070276600A1 (en) 2006-03-06 2007-11-29 King Timothy I Intersection collision warning system
US20080114530A1 (en) 2006-10-27 2008-05-15 Petrisor Gregory C Thin client intelligent transportation system and method for use therein
JP2009087372A (en) 2008-12-25 2009-04-23 Mitsubishi Electric Corp Road-to-vehicle communication system, and road-to-vehicle communication device
JP2009146219A (en) 2007-12-14 2009-07-02 Toshiba Denpa Products Kk Intersection collision accident preventing device
WO2010110109A1 (en) 2009-03-23 2010-09-30 本田技研工業株式会社 Information providing device for vehicle
CN202057015U (en) 2011-05-04 2011-11-30 宁波奥克斯空调有限公司 Air outlet structure of indoor unit of air conditioner
US20110298603A1 (en) 2006-03-06 2011-12-08 King Timothy I Intersection Collision Warning System
WO2011156621A2 (en) 2010-06-09 2011-12-15 Toyota Motor Engineering & Manufacturing North America, Inc. Computationally efficient intersection collision avoidance system
US20120095646A1 (en) 2009-09-15 2012-04-19 Ghazarian Ohanes D Intersection vehicle collision avoidance system
DE102011006176A1 (en) 2011-03-25 2012-09-27 Judo Wasseraufbereitung Gmbh Method for operating a metering pump and associated metering device
US20130026916A1 (en) 2011-07-28 2013-01-31 Denso Corporation Headlamp light distribution control device
CN202868581U (en) 2012-10-24 2013-04-10 苏坚池 Cutting device used for stage light imaging system
US8457827B1 (en) * 2012-03-15 2013-06-04 Google Inc. Modifying behavior of autonomous vehicle based on predicted behavior of other vehicles
GB2498438A (en) 2012-01-10 2013-07-17 Ford Global Tech Llc Collision avoidance which accounts for vehicle length
US20130253815A1 (en) 2012-03-23 2013-09-26 Institut Francais Des Sciences Et Technologies Des Transports, De L'amenagement System of determining information about a path or a road vehicle
DE102013005362A1 (en) 2013-03-28 2013-10-10 Daimler Ag Method for analyzing traffic conditions between vehicle and road user at e.g. road crossings, involves dynamically determining danger areas based on points of intersection of predicted movement trajectories
US8577550B2 (en) 2009-10-05 2013-11-05 Ford Global Technologies, Llc System for vehicle control to mitigate intersection collisions and method of using the same
JP2014006609A (en) 2012-06-22 2014-01-16 Mitsubishi Motors Corp Driving support device
CN103871272A (en) 2012-12-18 2014-06-18 江南大学 ZigBee-based vehicle-mounted intersection collision avoidance system
CN103898822A (en) 2014-03-27 2014-07-02 华南理工大学 Simple flyover
US20140303882A1 (en) 2013-04-05 2014-10-09 Electronics And Telecommunications Research Institute Apparatus and method for providing intersection collision-related information
CN104134370A (en) 2013-05-03 2014-11-05 株式会社万都 Method of preventing collision in vehicle
US20140347207A1 (en) 2013-05-23 2014-11-27 GM Global Technology Operations LLC Probabilistic target selection and threat assessment method and application to intersection collision alert system
US20150066348A1 (en) 2013-08-29 2015-03-05 Denso Corporation Collision probability determination apparatus and program
US20150183431A1 (en) 2012-08-08 2015-07-02 Toyota Jidosha Kabushiki Kaisha Collision prediction apparatus
US20150344030A1 (en) 2014-05-30 2015-12-03 Honda Research Institute Europe Gmbh Method and vehicle with an advanced driver assistance system for risk-based traffic scene analysis
US20160325743A1 (en) * 2015-05-04 2016-11-10 Honda Research Institute Europe Gmbh Method for improving performance of a method for computationally predicting a future state of a target object, driver assistance system, vehicle including such driver assistance system and respective program storage medium and program

Family Cites Families (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4073574B2 (en) * 1999-05-10 2008-04-09 本田技研工業株式会社 Steering control device for auto following vehicle
US6819991B2 (en) * 2001-11-29 2004-11-16 Ford Global Technologies, Llc Vehicle sensing based pre-crash threat assessment system
US6826478B2 (en) * 2002-04-12 2004-11-30 Ensco, Inc. Inertial navigation system for mobile objects with constraints
US6753804B2 (en) * 2002-05-21 2004-06-22 Visteon Global Technologies, Inc. Target vehicle identification based on the theoretical relationship between the azimuth angle and relative velocity
KR100651912B1 (en) * 2005-08-25 2006-12-01 엘지전자 주식회사 A location based service system and a position information updating method therefor
JP5016889B2 (en) * 2006-10-11 2012-09-05 日立オートモティブシステムズ株式会社 Preventive safety device
JP4254844B2 (en) * 2006-11-01 2009-04-15 トヨタ自動車株式会社 Travel control plan evaluation device
US8560218B1 (en) * 2008-12-31 2013-10-15 Dp Technologies, Inc. Method and apparatus to correct for erroneous global positioning system data
US20130238181A1 (en) * 2012-03-12 2013-09-12 Toyota Motor Eng. & Man. North America (Tema) On-board vehicle path prediction using processed sensor information
JP6003349B2 (en) * 2012-07-27 2016-10-05 トヨタ自動車株式会社 Vehicle behavior prediction device
US9286793B2 (en) * 2012-10-23 2016-03-15 University Of Southern California Traffic prediction using real-world transportation data
US9665694B2 (en) * 2013-01-31 2017-05-30 Codexis, Inc. Methods, systems, and software for identifying bio-molecules with interacting components
US20140257686A1 (en) * 2013-03-05 2014-09-11 GM Global Technology Operations LLC Vehicle lane determination
US8924066B2 (en) * 2013-05-22 2014-12-30 General Electric Company Systems and methods for determining route location
MX344376B (en) * 2013-10-17 2016-12-13 Ford Global Tech Llc Road characteristic prediction.
US9616898B2 (en) * 2013-12-12 2017-04-11 Ford Global Technologies, Llc System and method for determining effective road grade characteristic
US9346400B2 (en) * 2013-12-20 2016-05-24 Ford Global Technologies, Llc Affective user interface in an autonomous vehicle
DE102014205014A1 (en) * 2014-03-18 2015-09-24 Ford Global Technologies, Llc Method and device for detecting moving objects in the surroundings of a vehicle
CN105035090B (en) * 2015-06-06 2017-10-13 吉林大学 Autonomous land vehicle trajectory predictions control method based on traffic lights

Patent Citations (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070005609A1 (en) 1997-10-22 2007-01-04 Intelligent Technologies International, Inc. Vehicular Communication Arrangement and Method
JPH11175896A (en) 1997-12-05 1999-07-02 Fujitsu Ltd Method and system for preventing collision at intersection, storage medium in which collision at intersection preventing program is stored and intersection device
US6675094B2 (en) 2000-09-08 2004-01-06 Raytheon Company Path prediction system and method
US20030016143A1 (en) 2001-07-23 2003-01-23 Ohanes Ghazarian Intersection vehicle collision avoidance system
JP2005173703A (en) 2003-12-08 2005-06-30 Nissan Motor Co Ltd Intersection collision preventing apparatus
US20070030131A1 (en) * 2005-08-02 2007-02-08 Nissan Motor Co., Ltd. Vehicle obstacle verification system
CN1766232A (en) 2005-11-21 2006-05-03 钱红波 Novel traffic layout and control method for city road net and crossing
US20110298603A1 (en) 2006-03-06 2011-12-08 King Timothy I Intersection Collision Warning System
US20070276600A1 (en) 2006-03-06 2007-11-29 King Timothy I Intersection collision warning system
US20080114530A1 (en) 2006-10-27 2008-05-15 Petrisor Gregory C Thin client intelligent transportation system and method for use therein
JP2009146219A (en) 2007-12-14 2009-07-02 Toshiba Denpa Products Kk Intersection collision accident preventing device
JP2009087372A (en) 2008-12-25 2009-04-23 Mitsubishi Electric Corp Road-to-vehicle communication system, and road-to-vehicle communication device
WO2010110109A1 (en) 2009-03-23 2010-09-30 本田技研工業株式会社 Information providing device for vehicle
US20120095646A1 (en) 2009-09-15 2012-04-19 Ghazarian Ohanes D Intersection vehicle collision avoidance system
US8577550B2 (en) 2009-10-05 2013-11-05 Ford Global Technologies, Llc System for vehicle control to mitigate intersection collisions and method of using the same
WO2011156621A2 (en) 2010-06-09 2011-12-15 Toyota Motor Engineering & Manufacturing North America, Inc. Computationally efficient intersection collision avoidance system
DE102011006176A1 (en) 2011-03-25 2012-09-27 Judo Wasseraufbereitung Gmbh Method for operating a metering pump and associated metering device
CN202057015U (en) 2011-05-04 2011-11-30 宁波奥克斯空调有限公司 Air outlet structure of indoor unit of air conditioner
US20130026916A1 (en) 2011-07-28 2013-01-31 Denso Corporation Headlamp light distribution control device
GB2498438A (en) 2012-01-10 2013-07-17 Ford Global Tech Llc Collision avoidance which accounts for vehicle length
US8457827B1 (en) * 2012-03-15 2013-06-04 Google Inc. Modifying behavior of autonomous vehicle based on predicted behavior of other vehicles
US20130253815A1 (en) 2012-03-23 2013-09-26 Institut Francais Des Sciences Et Technologies Des Transports, De L'amenagement System of determining information about a path or a road vehicle
JP2014006609A (en) 2012-06-22 2014-01-16 Mitsubishi Motors Corp Driving support device
US20150183431A1 (en) 2012-08-08 2015-07-02 Toyota Jidosha Kabushiki Kaisha Collision prediction apparatus
CN202868581U (en) 2012-10-24 2013-04-10 苏坚池 Cutting device used for stage light imaging system
CN103871272A (en) 2012-12-18 2014-06-18 江南大学 ZigBee-based vehicle-mounted intersection collision avoidance system
DE102013005362A1 (en) 2013-03-28 2013-10-10 Daimler Ag Method for analyzing traffic conditions between vehicle and road user at e.g. road crossings, involves dynamically determining danger areas based on points of intersection of predicted movement trajectories
US20140303882A1 (en) 2013-04-05 2014-10-09 Electronics And Telecommunications Research Institute Apparatus and method for providing intersection collision-related information
CN104134370A (en) 2013-05-03 2014-11-05 株式会社万都 Method of preventing collision in vehicle
US20140347207A1 (en) 2013-05-23 2014-11-27 GM Global Technology Operations LLC Probabilistic target selection and threat assessment method and application to intersection collision alert system
US20150066348A1 (en) 2013-08-29 2015-03-05 Denso Corporation Collision probability determination apparatus and program
CN103898822A (en) 2014-03-27 2014-07-02 华南理工大学 Simple flyover
US20150344030A1 (en) 2014-05-30 2015-12-03 Honda Research Institute Europe Gmbh Method and vehicle with an advanced driver assistance system for risk-based traffic scene analysis
US20160325743A1 (en) * 2015-05-04 2016-11-10 Honda Research Institute Europe Gmbh Method for improving performance of a method for computationally predicting a future state of a target object, driver assistance system, vehicle including such driver assistance system and respective program storage medium and program

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
UKIPO Search Report for Application No. GB1713487.5 dated Feb. 8, 2018 (4 pages).

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11608063B2 (en) * 2018-11-15 2023-03-21 Volvo Car Corporation Vehicle safe stop
US10933890B2 (en) * 2019-01-30 2021-03-02 Toyota Jtdosha Kabushiki Kaisha Driving assistance device
US11021148B2 (en) * 2019-03-25 2021-06-01 Zoox, Inc. Pedestrian prediction based on attributes
US11351991B2 (en) * 2019-03-25 2022-06-07 Zoox, Inc. Prediction based on attributes
US11409284B2 (en) * 2019-05-15 2022-08-09 Baidu Usa Llc Relaxation optimization model to plan an open space trajectory for autonomous vehicles
US11493926B2 (en) 2019-05-15 2022-11-08 Baidu Usa Llc Offline agent using reinforcement learning to speedup trajectory planning for autonomous vehicles
US11467591B2 (en) * 2019-05-15 2022-10-11 Baidu Usa Llc Online agent using reinforcement learning to plan an open space trajectory for autonomous vehicles
US20210116916A1 (en) * 2019-10-22 2021-04-22 Baidu Usa Llc End dynamics and constraints relaxation algorithm on optimizing an open space trajectory
US11815891B2 (en) * 2019-10-22 2023-11-14 Baidu Usa Llc End dynamics and constraints relaxation algorithm on optimizing an open space trajectory
US10732261B1 (en) * 2019-12-31 2020-08-04 Aurora Innovation, Inc. Generating data using radar observation model based on machine learning
US10976410B1 (en) 2019-12-31 2021-04-13 Aurora Innovation, Inc. Generating data using radar observation model based on machine learning
US11351999B2 (en) * 2020-09-16 2022-06-07 Xuan Binh Luu Traffic collision warning device
US11926299B2 (en) 2020-11-23 2024-03-12 Aptiv Technologies Limited System and method for predicting road collisions with a host vehicle

Also Published As

Publication number Publication date
GB2556146A (en) 2018-05-23
DE102017119952A1 (en) 2018-03-01
CN107792079A (en) 2018-03-13
RU2017129885A (en) 2019-02-26
MX2017011127A (en) 2018-02-27
GB201713487D0 (en) 2017-10-04
US20180056995A1 (en) 2018-03-01
CN107792079B (en) 2022-05-31

Similar Documents

Publication Publication Date Title
US10118610B2 (en) Autonomous vehicle using path prediction
CN109421742B (en) Method and apparatus for monitoring autonomous vehicles
US11003182B2 (en) Vehicle monitoring and control infrastructure
US20200073405A1 (en) Vehicle navigation and control
US20200114921A1 (en) Sensor-limited lane changing
US11657635B2 (en) Measuring confidence in deep neural networks
US20220111859A1 (en) Adaptive perception by vehicle sensors
CN108394458B (en) Steering wheel feedback mechanism
US11574463B2 (en) Neural network for localization and object detection
CN108394457B (en) Steering wheel control mechanism for autonomous vehicle
US11555919B2 (en) Radar calibration system
CN112622922A (en) Non-autonomous vehicle speed prediction with autonomous vehicle reference
US10025319B2 (en) Collision-warning system
US20220172062A1 (en) Measuring confidence in deep neural networks
US20220207348A1 (en) Real-time neural network retraining
US11262201B2 (en) Location-based vehicle operation
US10246101B2 (en) Driver training system
US12097859B2 (en) Vehicle lane-change operations
US20240326815A1 (en) Vehicle maneuvering
US12128889B2 (en) Adaptive messaging within a cloud and edge computing environment for V2X applications
US20230219572A1 (en) Traction control system using feedforward control
US20230339517A1 (en) Autonomous driving evaluation system

Legal Events

Date Code Title Description
AS Assignment

Owner name: FORD GLOBAL TECHNOLOGIES, LLC, MICHIGAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:DENG, KUN;LUO, WANGDONG;LIU, NANJUN;AND OTHERS;SIGNING DATES FROM 20160826 TO 20160830;REEL/FRAME:039599/0486

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4