EP4020009A1 - System und verfahren zur klassifizierung von objekten in einem fahrzeug unter verwendung von merkmalsvektoren - Google Patents
System und verfahren zur klassifizierung von objekten in einem fahrzeug unter verwendung von merkmalsvektoren Download PDFInfo
- Publication number
- EP4020009A1 EP4020009A1 EP21161064.7A EP21161064A EP4020009A1 EP 4020009 A1 EP4020009 A1 EP 4020009A1 EP 21161064 A EP21161064 A EP 21161064A EP 4020009 A1 EP4020009 A1 EP 4020009A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- features
- signals
- vehicle
- velocity
- values
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
- 238000000034 method Methods 0.000 title claims abstract description 27
- 239000013598 vector Substances 0.000 title description 5
- 230000004044 response Effects 0.000 claims abstract description 15
- 230000008569 process Effects 0.000 claims abstract description 12
- 239000000284 extract Substances 0.000 claims abstract description 9
- 238000001514 detection method Methods 0.000 claims description 94
- 238000012545 processing Methods 0.000 claims description 12
- 230000015654 memory Effects 0.000 claims description 7
- 238000005516 engineering process Methods 0.000 description 7
- 230000003068 static effect Effects 0.000 description 5
- 238000013459 approach Methods 0.000 description 4
- 230000033001 locomotion Effects 0.000 description 4
- 238000010418 babysitting Methods 0.000 description 3
- 239000004744 fabric Substances 0.000 description 3
- 239000000463 material Substances 0.000 description 3
- 230000029058 respiratory gaseous exchange Effects 0.000 description 3
- 230000001629 suppression Effects 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 239000011159 matrix material Substances 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 238000010606 normalization Methods 0.000 description 2
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000008933 bodily movement Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000003066 decision tree Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000014759 maintenance of location Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000006403 short-term memory Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012706 support-vector machine Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/02—Systems using reflection of radio waves, e.g. primary radar systems; Analogous systems
- G01S13/04—Systems determining presence of a target
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/59—Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/02—Systems using reflection of radio waves, e.g. primary radar systems; Analogous systems
- G01S13/06—Systems determining position data of a target
- G01S13/08—Systems for measuring distance only
- G01S13/32—Systems for measuring distance only using transmission of continuous waves, whether amplitude-, frequency-, or phase-modulated, or unmodulated
- G01S13/34—Systems for measuring distance only using transmission of continuous waves, whether amplitude-, frequency-, or phase-modulated, or unmodulated using transmission of continuous, frequency-modulated waves while heterodyning the received signal, or a signal derived therefrom, with a locally-generated signal related to the contemporaneously transmitted signal
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/02—Systems using reflection of radio waves, e.g. primary radar systems; Analogous systems
- G01S13/06—Systems determining position data of a target
- G01S13/42—Simultaneous measurement of distance and other co-ordinates
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/02—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00
- G01S7/41—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional objects
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R21/00—Arrangements or fittings on vehicles for protecting or preventing injuries to occupants or pedestrians in case of accidents or other traffic risks
- B60R21/01—Electrical circuits for triggering passive safety arrangements, e.g. airbags, safety belt tighteners, in case of vehicle accidents or impending vehicle accidents
- B60R21/015—Electrical circuits for triggering passive safety arrangements, e.g. airbags, safety belt tighteners, in case of vehicle accidents or impending vehicle accidents including means for detecting the presence or position of passengers, passenger seats or child seats, and the related safety parameters therefor, e.g. speed or timing of airbag inflation in relation to occupant position or seat belt use
- B60R21/01512—Passenger detection systems
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/02—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00
- G01S7/41—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section
- G01S7/415—Identification of targets based on measurements of movement associated with the target
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/02—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00
- G01S7/41—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section
- G01S7/417—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section involving the use of neural networks
Definitions
- the present disclosure relates, in general, to a sensing system, and more specifically, relates to frequency modulated continuous wave (FMCW) radar sensing for classification of objects in a vehicle using feature vectors.
- FMCW frequency modulated continuous wave
- Radar FMCW technology is nowadays widely used to detect objects of interest within a seat occupancy zone in a vehicle.
- Occupancy sensors based on radar technology offer some advantages in comparison to other occupancy detection methods. It is a contactless and invisible measurement system, which can easily be integrated.
- Few exemplary existing technologies in the field of radar systems may use vital signs for occupancy detection by measuring the smallest motions like heartbeat or breathing of the occupant within the range of micrometres. Few other existing technologies may use synthetic wave analysis to determine the presence of the human body, where these technologies determine the presence only when the internal temperature rises to the dangerous level.
- exemplary existing technologies suffer from the limitations of inaccurate detection to determine the presence of an occupant and to distinguish the occupant, when other objects move in the vehicle, where the other objects may include engine vibration, shaking car in a bump, moving water bottle jacket and the like. Further, these existing technologies are also prone to delayed response due to the time taken to correctly establish the breathing rate and/or heart rate of the occupant and may require external accelerometer to remove car vibration.
- An object of the present disclosure relates, in general, to a sensing system, and more specifically, relates to frequency modulated continuous wave (FMCW) radar sensing for classification of objects in a vehicle using feature vectors.
- FMCW frequency modulated continuous wave
- Another object of the present disclosure is to provide a system that can use at least one sensor to cover more than one seat/location, with a minimum of one sensor per seat to a maximum of one sensor per whole car covering two rows, five seats, footwell and truck region
- Another object of the present disclosure is to provide a system that can be extended to larger vehicles like 6/7/8 seaters by increasing the field of view (FoV) of the sensor and/or by adding additional sensors of same type.
- FoV field of view
- Another object of the present disclosure that ensures faster response time of less than a second, when compared to other existing radar-based approaches that use vital signs for occupancy detection
- Another object of the present disclosure is to provide a system that can cater various in-cabin features
- Another object of the present disclosure is to provide a system that can be capable to operate under contactless and low ambient light condition.
- Another object of the present disclosure is to provide a system that can be capable to operate even when living objects are covered by some materials such as blanket, jacket, sun cover, cloth and the like.
- Yet another object of the present disclosure is to provide a system that can be capable to distinguish person leaning on the car door from a baby sitting inside the car.
- the present disclosure relates, in general, to a sensing system, and more specifically, relates to frequency modulated continuous wave (FMCW) radar sensing for classification of objects in a vehicle using feature vectors.
- FMCW frequency modulated continuous wave
- the present disclosure provides a system for differentiating objects present in a vehicle, the system including one or more sensors adapted to be placed within the vehicle to generate a set of signals in response to an object being present in one or more zones within the vehicle, an analogue-to-digital converter (ADC) configured to receive the set of signals and convert the received set of signals to a digital form, and a processor operatively coupled to the one or more sensors, the processor operatively coupled to a memory, the memory storing instructions executable by the processor to receive, from the ADC, the digital set of signals, process the received digital set of signals, to generate point cloud dataset of the received digital set of signals, the point cloud dataset pertaining to any or a combination of range, angle, velocity and reflected power of the object being present in the one or more zones within the vehicle, extract, from the point cloud dataset, a first set of features pertaining to a single frame that may include any or a combination of mean of values, and distribution of values, extract, from the point cloud dataset, a second set of features
- the one or more sensors may include a transmitter unit that emits signals onto the object, and a receiver unit that receives the emitted signals reflected from the object.
- the one or more sensors may include a mixer that combines the signals received from the transmitter unit and the receiver unit.
- the processor may perform fast fourier transform (FFT) on the received digital set of signals to calculate any or a combination of range and velocity of the received digital set of signals, wherein, the processor may detect prominent reflection points of the received set of signals based on predefined threshold value.
- FFT fast fourier transform
- processor may perform bearing angle estimation of the detected prominent reflection points, and a grouping unit may group the detected prominent reflection points.
- the processor operatively coupled to a classifier that receives the extracted first set of features and the second set of features to differentiate types of objects, object attributes and any combination thereof.
- the extracted set of features may include any or a combination of expectation of position, range/angle normalized average power, expectation of power, expectation of velocity, number of detections, detection density, gating factor, position spread, spread in velocity, spread in power, velocity distribution, distribution of power, and power distribution across velocity.
- the number of detections comprise total number of point cloud detections in respective one or more zones
- the detection density may include volume of detections in the one or more zones and the gating factor may include the density of detections from the entire one or more zones, wherein the one or more zones may include any or a combination of specific seat positions, specific footwell regions, entire row, entire footwell region, truck region, and entire vehicle.
- the velocity distribution and distribution of power may include maximum number of detections across successive set of bins, and power distribution across velocity may include a ratio of total number of detections on negative side of histogram peak by total number of detections on positive side of histogram peak.
- the present disclosure provides a method for differentiating objects present in a vehicle, the method including receiving, at a computing device, a digital set of signals, from an analogue-to-digital converter (ADC), the ADC configured to receive the set of signals from one or more sensors and convert the received set of signals to a digital form, the one or more sensors adapted to be placed within a vehicle to generate the set of signals in response to an object being present in one or more zones within the vehicle; processing, at the computing device, the received digital set of signals, to generate point cloud dataset of the received digital set of signals, the point cloud dataset pertaining to any or a combination of range, angle, velocity and reflected power of the object being present in the one or more zones within the vehicle, extracting, at the computing device, from the point cloud dataset, a first set of features pertaining to a single frame that may include any or a combination of mean of values, and distribution of values, extracting, at the computing device, from the point cloud dataset, a second set of features pertaining to a multi-
- the present disclosure relates, in general, to a sensing system, and more specifically, relates to frequency modulated continuous wave (FMCW) radar sensing for classification of objects in a vehicle using feature vectors.
- FMCW frequency modulated continuous wave
- the present disclosure provides a method of robust vehicle occupancy system by using a strong set of features extracted from point cloud information of FMCW radar system, which is capable of reliably detecting a life present in the whole vehicle including footwell region, detecting occupant present in specific zones, and distinguishing the present occupant as a child or adult.
- the present disclosure provides a robust in-cabin sensing with FMCW radar system that are applicable within the vehicle, which includes various features such as life presence detection, seat occupancy detection, adult vs child classification, child/pet left behind detection, seat belt reminder, out of position and the like
- the present disclosure provides the radar sensor system and a method of operating the radar sensor system to detect occupancy of the interior of the vehicle by extracting the set of features.
- the system may perform signal processing steps of Fast Fourier Transform (FFT) calculation, removal of static object reflections, and applying a threshold to extract the point cloud information.
- FFT Fast Fourier Transform
- the set of features are extracted and provided to a classifier for all classification applications of the vehicle interior.
- the set of features that enable to provide all the above-listed features of in-cabin sensing, these features are extracted from point cloud information of the FMCW radar sensor, where these features may be used as an input to the classifier to distinguish the living object from other static objects.
- the system can use one sensor for more than one seat/zone in the vehicle.
- the present disclosure can cover a minimum of one seat per sensor up to a maximum of the full interior of the car with a single sensor spanning multiple rows, seats, footwell and trunk region.
- the system may include a radar sensor that may include a transmitter unit to transmit the radar signals, a receiver unit to receive the radar signals, a signal processing unit that performs fast FFT and thresholding techniques to extract the point cloud information.
- the present disclosure further includes the classifier to extract the feature set from the point cloud information to perform various classifier applications to determine any or a combination of life presence detection, seat occupancy detection, adult vs child classification and child left behind detection that carter to the in-cabin requirements of any automotive vehicle.
- the present disclosure can be described in enabling detail in the following examples, which may represent more than one embodiment of the present disclosure.
- FIG. 1 illustrates an exemplary representation of a FMCW radar system for differentiating objects in a vehicle, in accordance with an embodiment of the present disclosure.
- FMCW radar system 100 may be configured to differentiate objects 104 also interchangeably referred to as targets 104 in an interior of a vehicle, where the objects 104 may be an adult, child, infant and the like.
- the system 100 may include one or more sensors 102, for example, a radar sensor that may be mounted within the vehicle with its radio frequency (RF) emitting direction pointing towards the interior of the vehicle.
- the one or more sensors 102 may include transmitter unit 110, a receiver unit 112, a mixer 114, a low pass filter (LPF) 116, an analogue to digital converter (ADC) 118, a processor 106, and a classifier 120.
- the system 100 can classify the objects 104 into various classes to determine the life presence detection, seat occupancy detection, adult vs infant/child detection, child left behind detection, airbag deployment, out of position detection, airbag suppression, automated child lock and the like.
- the vehicle as presented in the example may be a four-wheeler vehicle, e.g., a car.
- the present disclosure may not be limited to this configuration but may be extended to other configurations such as bus, trucks and the like.
- the at least one sensor may be preferably mounted within the vehicle to generate a set of signals in response to the object 104 being present/positioned in one or more zones within the vehicle.
- different vehicles may require one or more sensor mounting configurations, where the sensor arrangement can be divided into one or more zones.
- the at least one sensor may cover more than one seat/zone/location in the car covering two rows, five seats, footwell and truck region, whereas one or more sensors may be used to increase the field of view (FoV) of the sensor in larger vehicles with 6/7/8 seaters.
- the one or more zones also interchangeably referred to as seat group may refer to the group of detection points that are from a predefined area within the vehicle, where the area can refer to the reflections from any objects 104 present in the single-seat location, single row location, trunk, multiple seats grouped, or the entire car.
- the area dimensions may be known with respect to the car and the sensor mounting in the car.
- the transmitter unit 110 may include at least one antenna to emit high-frequency signals (radar signals) illuminating the interior of the vehicle
- the receiver unit 112 may include at least one antenna that may receive the emitted signals after getting reflected on the objects 104.
- the system 100 may use the transmitter unit 110 and receiver unit 112 to transmit and receive the signal frequency in GHz or any required suitable range.
- the frequency range may include 60-64GHz band, 77-81GHz band and 24 GHz band and any combination thereof.
- the system 100 may emit the selected radar signals and receive the same signal back after reflecting from the object 104, where the reflected signal may include the information specifically about the reflected object.
- a mixer 114 may be operatively coupled to the transmitter unit 110 and the receiver unit 112 to combine the signals received from the transmitter unit 110 and the receiver unit 112, the mixer 114 may be operatively coupled to the LPF 116 to obtain the intermediate frequency signal also interchangeably referred to as the set of signals, which may be considered as an input data, where the intermediate frequency signal may include range, velocity and bearing angle information about the reflected object.
- the received intermediate signal has information from multiple reflections from all objects 104 in the FoV of the one or more sensors 102.
- the ADC 118 may convert the received set of signals to process in a digital domain in the processor 106.
- the set of signals also interchangeably referred to as input data may be collected using FMCW radar with any or a combination of single waveform pattern and multiple waveform patterns.
- the waveform pattern may be up-chirp waveform pattern, with a constant slope.
- the input data may include one or more samples within the chirp, for more than one chirp and more than one receiver antenna, the input data may be arranged in a cube pattern as illustrated in FIG. 2 and explained in detail below.
- the processor 106 may be a signal processing unit.
- the processor 106 may include a memory 108 for storing the information, where the memory 108 can be part of the processor 106 or can be a separate unit associated with the processor 106 depending upon the application.
- the processor 106 may receive the digital set of signals from the ADC 118 to extract prominent reflected signals.
- the processor 106 may process the received digital set of signals, to generate point cloud dataset also interchangeable referred to as point cloud information/list of the received digital set of signals using two-dimensional Fast Fourier Transform (2D FFT), thresholding technique, and Direction of Arrival (DoA) algorithm.
- 2D FFT two-dimensional Fast Fourier Transform
- DoA Direction of Arrival
- the 2D FFT may be processed across the samples in the chirp to process any or a combination of range information and velocity/doppler information, where a predefined threshold value (thresholding technique) may be used to detect the prominent reflection points in the received digital set of signals.
- the DoA algorithm may be used to estimate the bearing angle of the detected prominent reflection points.
- a grouping mechanism also interchangeably referred to as grouping unit may group the detected prominent reflection points based on the position with the region of interest within the vehicle and with respect to the mounted sensor.
- the resultant of all these signal processing may generate the point cloud dataset/list having details of the prominent reflection points, where the point cloud dataset may include the information about the range, angle (azimuth and/or elevation), velocity and/or reflected power of the targets, where the targets may include any or a combination of adult, children, baby, empty seats and other objects 104 inside the car that are within the FoV of the one or more sensors 102.
- the processor 106 may extract, from the point cloud dataset, first set of features pertaining to a single frame that may include any or a combination of the mean of values, and distribution of values.
- the processor 106 may extract, from the point cloud dataset, second set of features pertaining to a multi-frame that may include any or a combination of time-averaged values and variation over time value of the first set of features.
- the first set of features also interchangeably referred to as single frame features that may include the mean and/or expectation of values i.e., expectation of position, range/angle normalized average power, expectation of power, expectation of velocity, number of detections, detection density, gating factor.
- the single frame features may include distribution or spread of values i.e., position spread, spread in velocity, spread in power, velocity distribution, distribution of power, power distribution across velocity.
- the second set of features also interchangeably referred to as the multi-frame features that may include time averaged values and variation and/or distribution over time of the following values i.e., expectation of position, range/angle normalized average power, expectation of power, expectation of velocity, number of detections, detection density, gating factor, position spread, spread in velocity, spread in power, velocity distribution, distribution of power, power distribution across velocity.
- the one or more zones may be defined in any or a combination of two dimension and three dimensions, where the zones can be defined in cartesian or polar coordinates.
- the zones can be a single cuboid and/or rectangle or a group of multiple cuboids and/or rectangle, the zones specifically show the areas of interest in the point cloud dataset. For every frame the point clouds are grouped with respect to the different zones, the features are extracted based on the detection points that fall within each respective zone.
- the processor 106 may be operatively coupled to the classifier 120.
- the classifier 120 may be any or a combination of a simple decision tree, discriminant analysis, support vector machines, neural network, and an ensemble of multiple classifiers.
- the classifier 120 may classify the extracted first set of features and the second set of features to extract a set of values i.e., output data from the classified set of features as explained in detail below.
- the processor 106 may be configured to differentiate the objects 104 present in the one or more zones within the vehicle, where the classifier 120 differentiate types of object 104 e.g., living from static objects, and object attributes, e.g., adult from child, human from pet and any combination thereof.
- the living objects e.g., adult, child and pet may be differentiated from the static objects, e.g., empty child restraint system, water bottle, shaking the car, engine vibrations, corner reflectors and other noise signals.
- deployment force for the airbag may be controlled based on the classification, where if the classification indicates that the adult is the seat occupant then the airbag is deployed in a normal manner. If the classification indicates that a child or infant is the seat occupant then the airbag may not be deployed or may be deployed at a significantly lower deployment force.
- the system 100 can ensure a faster response time of less than a second, when compared to other existing radar-based approaches that use vital signs for occupancy detection.
- the system 100 can provide various in-cabin features that may include any or a combination of life presence detection, seat occupancy detection or occupancy detection, adult vs infant/child detection, a child left behind detection, passenger classification system, seat belt reminder, airbag deployment system, airbag suppression and the like.
- the system 100 can be capable to operate in any or a combination of contactless and low ambient light conditions and can be capable to operate even when the living objects are covered by some materials such as a blanket, jacket, sun cover, and cloth. Further, the system can be capable to distinguish a person leaning on the car door from baby sitting inside the car.
- FIG. 2 illustrates an exemplary view of three dimensional (3D) Fast Fourier Transform radar cube representation, in accordance with an embodiment of the present disclosure.
- the input data may be collected using FMCW radar with any or a combination of single waveform pattern and multiple waveform patterns.
- the waveform pattern may be up-chirp waveform pattern, with the constant slope.
- the input data may include multiple samples within the chirp, for more than one chirp and for more than one receiver antenna, the input data can be arranged in cube pattern.
- the intermediate frequency signal may include range, velocity and bearing angle information about the reflected object 104, where the received intermediate signal may include information from multiple reflections from the objects 104 in the FoV of the one or more sensors 102.
- the distance of the object 104 from the radar and the relative velocity of the object 104 may be determined by the peaks in the 2D FFT of the input data, where the first dimension refers to the direction along the ADC samples within the chirp (also referred to as range direction) and the second dimension refers to the direction along the chirps (also referred to as velocity direction).
- the present disclosure can be used for the FMCW radar with a minimum of one transmitter antenna and one receiver antenna.
- the FMCW radar with one or more transmitter and/or receiver antenna may have object bearing angle related phase information across the antenna, where the third dimension refers to the object bearing angle related phase information across the antenna.
- one or more transmitter and receiver antennas can be arranged only in one direction (either azimuth or elevation) or/and in both azimuth and elevation direction.
- the 2D FFT may determine the bearing angle in azimuth and elevation.
- Other DoA estimation algorithms like barlett, Capon/( Minimum Variance Distortionless Response beamformer) MvDR, MUSIC, estimation of signal parameters via rotational invariant techniques (ESPRIT) or Matrix Pencil can be used for better accuracy and resolution with higher computation cost.
- the present disclosure is independent of the selected DoA algorithm.
- the present disclosure describes the signal processing pipeline, including, but not limited to, range estimation, velocity estimation, threshold process and the DoA estimation.
- the present disclosure may not be limited to this signal processing pipeline but may be extended to other signal processing pipeline having range and velocity estimation followed by angle estimation (azimuth and/or elevation) and then thresholding to detect prominent reflection points.
- Another possible signal processing pipeline may include range and bearing angle estimation followed by thresholding and then velocity estimation.
- the resultant of all these signal processing pipeline may be the point cloud list having details of the prominent reflection points and their information may include position (range, angle), velocity and power.
- the present disclosure may include any signal processing pipeline that can provide the point cloud list or object list, which may include the information on the prominent peaks or reflections from the objects 104.
- the point cloud list is a set of detection points, which represents the reflected signals, where the reflected signals are the peaks that may be seen in the fourier transform plots and these may be detected by the threshold techniques.
- the point cloud list has information about the range, angle (azimuth and/or elevation), velocity and/or reflected power of the targets, where the targets include any or a combination of adult, children, baby, empty seats and other objects 104 inside the car that are within the FoV of the one or more sensors 102.
- the features described in the present disclosure are with reference to the set of detection points within the one or more zones/seat group, where the zone/seat group can refer to the group of detection points that are from the predefined area within the vehicle.
- the area can refer to the reflections from any objects 104 present in the single seat location, single row location, trunk, multiple seats grouped, or the entire car.
- the area dimensions may be already known with respect to the car and the sensor mounting in the car, where the area may include, but not limited to, two-dimensions.
- the zones may be defined in any or a combination of two dimension or three dimensions, where the zones can be defined in cartesian or polar coordinates.
- the zones can be the single cuboid and/or rectangle or the group of multiple cuboids and/or rectangle, the zones specifically show the areas of interest in the point cloud.
- the point clouds are grouped with respect to the different zones. The features are extracted based on the detection points that fall within each respective zone. In a certain frame there might be no points within the zone or only one point or a plural of detection points.
- the present disclosure describes the different features that can be extracted from the point cloud list, where the set of features i.e., first set of features and the second set of features are extracted for the group of points that gets reflected from a signal object 104, each object can have singular or plural of detection points assigned.
- the first set of features may include any or a combination of mean or expectation of values in a signal time instant (single frame) and distribution or spread of values in a single time instant (single frame).
- the second set of features may include any or a combination of time averaged values across time (multi-frames) and variation or distribution over time (multi-frames).
- the features of signal time instant refer to the features that are derived from the current single frame, these features contribute to faster response time, wherein if faster response time is not a required feature of the system, then these features need not be included.
- the features derived over time includes the information over multi-frames, these features can represent the values that are present in the previous frames but may not be present in the current frame. These features provide a short-term memory feature to the classification process. These features may be responsible for consistent detection output irrespective of any changes in detection capability of the target.
- the features defined by the mean, expectation or averaged values represent the object 104 information without any outliers. These set of features remove the noisy information and keep important centre value of the spread.
- the features defined by distribution, spread or variation represent the variation of values in the single frame or across multi-frames. In the single frame, this value represents the spread of the corresponding value across multiple points from the object 104. In the multi-frame, this value represents the variation over time.
- Mean or expectation of values in the signal time instant may include expectation of position, range/angle normalized average power, expectation of power, expectation of velocity, number of detections, detection density, gating factor.
- Distribution or spread of values in a single time instant may include position spread, spread in velocity, spread in power, velocity distribution, distribution of power, power distribution across velocity.
- Time averaged values may include TMA of expectation of position, TMA of range/angle normalized average power, TMA of expectation of power, TMA of expectation of velocity, TMA of number of detections, TMA of detection density, TMA of gating factor, TMA of position spread, TMA of spread in velocity, TMA of spread in power, TMA of velocity distribution, TMA of distribution of power, TMA of power distribution across velocity
- Variation or distribution over time may include DoT of expectation of position, DoT of range/angle normalized average power, DoT of expectation of power, DoT of expectation of velocity, DoT of number of detections, DoT of detection density, DoT of gating factor, DoT of position spread, DoT of spread in velocity, DoT of spread in power, DoT of velocity distribution, DoT of distribution of power, DoT of power distribution across velocity.
- Expectation of position refers to the mean/averaged value of the different position information of the point cloud from the same zone.
- the position information can be in cartesian or polar coordinates.
- the position information can be represented in a three-dimensional form or in two-dimensional form.
- Ex Az angle Sum az_angle i / N
- Ex[El angle] Sum (az_anglei)/N, where Range i refers to the range value of i th point, az_anglei refers to the azimuth angle value of i th point, El_angle i refers to the elevation angle value of i th point and N refers to the total number of points in the group.
- the expectation of position can be represented in XYZ coordinates. The expectation of the position information need not be always in three dimensional.
- the final set of features can include only the information from any one of the coordinates, or combination of two of the coordinates or all the three coordinates as described.
- Range/angle normalized average power refers to the mean/averaged value of normalized reflected power from all the points that belong to a specific zone and/or object
- Ex[Norm_Power] Sum (Norm_Power i )/N, where Norm_Power i refers to the normalized reflected power value of i th point and N refers to the total number of points in the group.
- the normalization of the reflected power value obtained from the point cloud can be achieved by the inverse antenna beam pattern. It is known that the reflected power reduces with increasing range and as the object's bearing angle is away from 0 degree. This reduction in power can be derived from the antenna beam pattern and/or by actual measurements.
- the reflected power value is compensated for this attenuation based on its position (range and/or angle).
- the normalization can be performed for range, azimuth angle and elevation angle together or a combination of either of two values or as individual three values. Hence, this feature range/angle normalized average power represents any one or few or all the seven combinations as described.
- Number of detections is the total number of point cloud detections in the respective group or zone/seat.
- Detection density refers to the density of detections from the object. This has information about the activity of the object/living person like stable or slight movements or breathing/heart rate movements or huge bodily movements.
- Detect_density [detection_volume]/N, where N refers to the total number of detections in the group and detection_volume refers to the volume of detections in the group and is determined by the [range_spread x Az_angle_spread x El_angle_spread].
- the volume can also be determined by including only one angle information of either azimuth or elevation as [range_spread x angle_spread].
- the zone might refer to specific seat positions and/or specific footwell regions and/or entire row and or entire footwell region and/or and/or truck region and/or entire car and/or other combinations of regions within the car.
- Positional Spread refers to the difference in maximum and minimum values of the different position information of the point cloud from the same zone.
- the position information can be in cartesian or polar coordinates.
- the position information can be represented in a three-dimensional form or in two-dimensional form.
- the positional spread can be represented in XYZ coordinates (cartesian).
- the positional spread feature need not be always in three dimensional.
- the final set of features can include only the information from any one of the coordinates, or combination of two of the coordinates or all the three coordinates as described.
- Velocity_spread (Velocity max - Velocity min ), where Velocity max refers to the maximum range value and Velocity min refers to the minimum range value among all the points from the group.
- Power_spread refers to the difference in the max and min values of the reflected power from all the points that belong to a specific zone or object.
- Power spread (Power max - Power min ), where Power max refers to the maximum reflected power value and Power min refers to the minimum reflected power value among all the points from the group.
- Velocity distribution refers to the parameters representing the distribution of velocity in a position frame.
- the highest velocity detection point is selected as a peak in range-azimuth-elevation plane.
- the number of detections in the successive bins in this 3D plane is counted. If this number of detections is not zero, then the next successive bins in this 3D space is considered and the number of detections in this space is seen. This process is repeated till number of detections is zero or the extremes of the zone is reached.
- the velocity distribution is defined as the maximum number of detections across successive set of bins.
- Distrb_velocity max (no_dectection i ), where no_dectection i refers to the number of detections in the i th set of bins around the peak value.
- the bins can be selected in any one of the three-position axis in cartesian or polar coordinates, or in combinations of two axis or along all the three axis also.
- Distribution of power refers to the parameters representing the distribution of reflected power in a position frame.
- the highest reflected power detection point is selected as a peak in range-azimuth-elevation plane.
- the number of detections in the successive bins in this 3D plane is counted. If this number of detections is not zero, then the next successive bins in this 3D space is considered and the number of detections in this space is seen. This process is repeated till number of detections is zero or the extremes of the zone is reached.
- the bins can be selected in any one of the three-position axis in cartesian or polar coordinates, or in combinations of two axis or along all the three axis also.
- Power_distrb_vel refers to the distribution of detections points across velocity.
- the negative side and positive side of the histogram pot is defined with respect to the peak of the histogram and the positive side refers to the increasing velocity value whereas the negative side refers to the decreasing velocity value.
- the set of features grouped under time averaged (TMA) values represent a moving averaged value of the single frame features.
- the number of frames considered can be as low as zero or as high as total number of frames.
- the moving average is applied on the above defined thirteen single frame features belonging to both mean group and distribution group.
- TMA j (feature i ) ALPHA x TMA (j-1) (feature i ) + feature i,j , where feature i refers to the i th feature from the previous list, feature i,j refers to the same i th feature on the j th frame or time instant.
- TMA j refers to the time average operation on the j th frame or time instant.
- ALPHA refers to the retention factor and takes a value between 0 to 1. A value of 0 for ALPHA means that there is no previous information and a value of 1 means that all previous values are included.
- TMA j (feature i ) Sum (feature i-N ........... feature i ) / N, where the feature values of the previous N frames is averaged. This approach also provides a time averaged feature value. If the above representation is modified as below, providing the same result, then it is less computationally intensive and is preferred for embedded implementation.
- TMA j feature i TMA j ⁇ 1 feature i ⁇ N ⁇ feature i ⁇ N + feature i / N
- the set of features grouped under distribution over time (DoT) features represent the variation or spread of all the single frame features over time or multi-frames.
- the number of frames considered can be as low as zero or as large as the current total number of frames.
- the distribution over time value for a feature is the difference between the maximum and the minimum value of the feature across a predefined set of frames (N) and is presented as the formula below:
- DoT j feature i Max feature i ⁇ N .
- feature i Min feature i ⁇ N .
- Both time-averaged implementation and distribution over time implementation may be applied for all the thirteen single frame features that are listed and explained above.
- the extracted thirty-nine features are evaluated for minimum redundancy and maximum relevance to the classification.
- Each feature mentioned here has a major role in the classification of at least one of the use cases of in-cabin sensing.
- These features extracted from the point cloud information can be provided to the classifier 120 to distinguish living vs static object, distinguish infant vs child vs adult, and distinguish human vs pet.
- the described features of the present disclosure can be used in any combination for the desired results. The selection of features from the set of twenty-six features obtained based on memory availability, the complexity of classifier 120, the linearity of the classifier 120, use case, feature and required region of interest.
- the selection of features coupled with the classifier 120 can be used to perform all the radar-based in-vehicle applications such as life presence detection, seat occupancy detection, adult vs child classification, child and/or pet left behind detection, passenger classification system, out of position, automatic child lock, intrusion detection, seat belt reminder, airbag deployment, airbag suppression, airbag low-risk deployment and the like.
- FIG. 3 illustrates an exemplary top view of the antenna averaged data at zero degrees angle, in accordance with an embodiment of the present disclosure.
- the FMCW radar with one or more transmitter and/or receiver antenna may have an object bearing angle related phase information across the antenna.
- the one or more transmitter and receiver antennas can be arranged only in one direction (either azimuth or elevation) or/and in both azimuth and elevation direction, to determine the bearing angle of the object 104 in both azimuth and/or elevation, as subtended from the objects 104 on the normal line, the 2D FFT may determine the bearing angle in azimuth and elevation.
- Other DoA estimation algorithms like barlett, Capon/ MvDR, MUSIC, ESPRIT or Matrix Pencil can be used for better accuracy and resolution with higher computation cost.
- the present disclosure is independent of the selected DoA algorithm.
- FIG.4 illustrates an exemplary view of the radar mounting position within the vehicle, in accordance with an embodiment of the present disclosure.
- the one or more sensors 102 mounted inside the vehicle with its RF emitting direction pointing towards the interior of the vehicle, for example, the one or more sensors 102 mounted in front, top and rear portion within the vehicle.
- the system 100 of the present disclosure tuned to operate the one or more sensors 102 placed at any position as long as the FoV of the sensor covers the required region of interest.
- the present disclosure may be extensively used in all automotive such as passenger cars, trucks, buses and the like.
- FIG. 5 illustrates an exemplary flow diagram of a method for differentiating objects in a vehicle, in accordance with an embodiment of the present disclosure.
- method 500 can be implemented using a computing device, which can include the processors 106.
- the computing device may receive a digital set of signals from an analogue-to-digital converter (ADC), the ADC configured to receive the set of signals from one or more sensors and convert the received set of signals to a digital form, the one or more sensors adapted to be placed within a vehicle to generate the set of signals in response to an object being present in one or more zones within the vehicle.
- ADC an analogue-to-digital converter
- the computing device may process the received digital set of signals, to generate point cloud dataset of the received digital set of signals, the point cloud dataset pertaining to any or a combination of range, angle, velocity and reflected power of the object being present in the one or more zones within the vehicle.
- the computing device may extract, from the point cloud dataset, the first set of features pertaining to a single frame that may include any or a combination of mean of values, and distribution of values.
- the computing device may extract, from the point cloud dataset, the second set of features pertaining to a multi-frame that may include any or a combination of time averaged values and variation over time values of the first set of features.
- the computing device may classify the extracted first set of features and the second set of features to extract a set of values for the classified set of features.
- the computing device may be configured to differentiate the object positioned in one or more zones within the vehicle, based on the determination of the extracted set of values from the classified set of features.
- the present disclosure provides a system that can use at least one sensor to cover more than one seat/location, with a minimum of one sensor per seat to a maximum of one sensor per whole car covering two rows, five seats, footwell and truck region
- the present disclosure provides a system that can be extended to larger vehicles like 6/7/8 seaters by increasing the field of view of the sensor and/or by adding additional sensors of same type.
- the present disclosure ensures faster response time of less than a second, when compared to other existing radar-based approaches that use vital signs for occupancy detection
- the present disclosure provides a system that can cater various in-cabin features
- the present disclosure provides a system that can be capable to operate under contactless and low ambient light condition.
- the present disclosure provides a system that can be capable to operate even when the living objects are covered by some materials such as blanket, jacket, sun cover, cloth and the like.
- the present disclosure provides a system that can be capable to distinguish person leaning on the car door from a baby sitting inside the car.
Landscapes
- Engineering & Computer Science (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Bioinformatics & Cheminformatics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Radar Systems Or Details Thereof (AREA)
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
IN202041055923 | 2020-12-22 |
Publications (1)
Publication Number | Publication Date |
---|---|
EP4020009A1 true EP4020009A1 (de) | 2022-06-29 |
Family
ID=74859363
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP21161064.7A Withdrawn EP4020009A1 (de) | 2020-12-22 | 2021-03-05 | System und verfahren zur klassifizierung von objekten in einem fahrzeug unter verwendung von merkmalsvektoren |
Country Status (3)
Country | Link |
---|---|
US (1) | US11527080B2 (de) |
EP (1) | EP4020009A1 (de) |
WO (1) | WO2022136930A1 (de) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP4328095A1 (de) * | 2022-08-24 | 2024-02-28 | Robert Bosch GmbH | Klassifizierungssysteme und -verfahren zur fahrzeuginternen erfassung mit radar |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2022136930A1 (en) * | 2020-12-22 | 2022-06-30 | PathPartner Technology Private Limited | System and method for classification of objects in vehicle using feature vectors |
US12017657B2 (en) * | 2022-01-07 | 2024-06-25 | Ford Global Technologies, Llc | Vehicle occupant classification using radar point cloud |
US20230406158A1 (en) * | 2022-05-30 | 2023-12-21 | Toyota Connected North America, Inc. | In-cabin detection framework |
CN115883812B (zh) * | 2022-11-30 | 2024-07-05 | 重庆大学 | 基于粒子群算法的多传感器协同布站方法 |
DE102023103210A1 (de) * | 2023-02-09 | 2024-08-14 | Audi Aktiengesellschaft | Vorrichtung zur Aktivierung/Deaktivierung einer Airbagfunktion in einem Fahrzeug |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160200276A1 (en) * | 2013-08-14 | 2016-07-14 | Iee International Electronics & Engineering S.A. | Radar sensing of vehicle occupancy |
WO2019238575A1 (en) * | 2018-06-11 | 2019-12-19 | Iee International Electronics & Engineering S.A. | Method for robust vehicle occupancy detection with vital sign monitoring |
US20200341114A1 (en) * | 2017-03-28 | 2020-10-29 | Sri International | Identification system for subject or activity identification using range and velocity data |
Family Cites Families (39)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7783403B2 (en) * | 1994-05-23 | 2010-08-24 | Automotive Technologies International, Inc. | System and method for preventing vehicular accidents |
US8731736B2 (en) * | 2011-02-22 | 2014-05-20 | Honda Motor Co., Ltd. | System and method for reducing driving skill atrophy |
US8872447B2 (en) * | 2012-08-08 | 2014-10-28 | Honda Motor Co., Ltd. | Dual function power door |
US9193359B2 (en) * | 2013-08-12 | 2015-11-24 | GM Global Technology Operations LLC | Vehicle systems and methods for identifying a driver |
US9436410B2 (en) * | 2013-12-13 | 2016-09-06 | Netapp, Inc. | Replication of volumes on demands using absent allocation |
EP2891589B1 (de) * | 2014-01-06 | 2024-09-25 | Harman International Industries, Incorporated | Automatische Fahreridentifikation |
US20150242605A1 (en) * | 2014-02-23 | 2015-08-27 | Qualcomm Incorporated | Continuous authentication with a mobile device |
US9428054B2 (en) * | 2014-04-04 | 2016-08-30 | Here Global B.V. | Method and apparatus for identifying a driver based on sensor information |
US20160300050A1 (en) * | 2014-05-28 | 2016-10-13 | GM Global Technology Operations LLC | Verifying a user with biometric data |
DE102014212758A1 (de) * | 2014-07-02 | 2016-01-07 | Robert Bosch Gmbh | Verfahren und Vorrichtung zur Erkennung eines Fahrers eines Fahrzeugs |
US9418491B2 (en) * | 2014-09-22 | 2016-08-16 | Brian K. Phillips | Method and system for automatically identifying a driver by creating a unique driver profile for a vehicle from driving habits |
US9779565B2 (en) * | 2015-03-24 | 2017-10-03 | Panasonic Automotive Systems Company Of America, Division Of Panasonic Corporation Of North America | Method and apparatus for managing operation of a driver's mobile telephone |
US20170186054A1 (en) * | 2015-12-28 | 2017-06-29 | Bosch Automotive Service Solutions Inc. | System To Identify A Driver |
US10137848B2 (en) * | 2016-01-13 | 2018-11-27 | Ford Global Technologies, Llc | System identifying a driver before they approach the vehicle using wireless communication protocols |
US10315665B2 (en) * | 2016-01-29 | 2019-06-11 | Faraday & Future Inc. | System and method for driver pattern recognition, identification, and prediction |
US9707911B1 (en) * | 2016-03-21 | 2017-07-18 | Ford Global Technologies, Llc | Identifying a driver of a vehicle |
US12044789B2 (en) * | 2016-04-22 | 2024-07-23 | Azar Zandifar | Systems and methods for occupancy detection using WiFi sensing technologies |
US10053110B2 (en) * | 2016-05-06 | 2018-08-21 | Toyota Motor Engineering & Manufacturing North America, Inc. | Systems and methodologies for controlling an autonomous vehicle |
US10154029B1 (en) * | 2016-05-31 | 2018-12-11 | Wells Fargo Bank, N.A. | Biometric knowledge extraction for mutual and multi-factor authentication and key exchange |
US9994232B2 (en) * | 2016-06-24 | 2018-06-12 | GM Global Technology Operations LLC | Dynamic assignment of driver identifiers and related adjustment of vehicle settings based on detection of driver identifiers |
WO2018009567A1 (en) * | 2016-07-05 | 2018-01-11 | Nauto Global Limited | System and method for automatic driver identification |
US20180018179A1 (en) * | 2016-07-12 | 2018-01-18 | Ford Global Technologies, Llc | Intelligent pre-boot and setup of vehicle systems |
US9953244B2 (en) * | 2016-08-16 | 2018-04-24 | RFNAV, Inc. | Apparatus and method for single look main lobe and sidelobe discrimination in spectral domain images |
US10464530B2 (en) * | 2017-01-17 | 2019-11-05 | Nio Usa, Inc. | Voice biometric pre-purchase enrollment for autonomous vehicles |
EP3373037B1 (de) * | 2017-03-10 | 2023-06-28 | The Hi-Tech Robotic Systemz Ltd | Fortschrittliches fahrerassistenzsystem mit einzelnem gehäuse |
US11847651B2 (en) * | 2017-05-23 | 2023-12-19 | Kenneth A Kopf | Systems and methods for facilitating biometric tokenless authentication for services |
US10710588B2 (en) * | 2017-05-23 | 2020-07-14 | Toyota Motor Engineering & Manufacturing North America, Inc. | Merging and lane change acceleration prediction energy management |
US10600270B2 (en) * | 2017-08-28 | 2020-03-24 | Ford Global Technologies, Llc | Biometric authentication for a vehicle without prior registration |
US10621112B2 (en) * | 2018-01-25 | 2020-04-14 | Dell Products L.P. | System and method of identifying a device driver |
CA3087506A1 (en) * | 2018-01-31 | 2019-08-08 | Xirgo Technologies, Llc | Enhanced vehicle sharing system |
US20200160263A1 (en) * | 2018-11-06 | 2020-05-21 | Marcus Kuettner | System, Method, And Packaging For Secure Food Delivery |
US11511199B2 (en) * | 2019-02-28 | 2022-11-29 | Vsn Vision Inc. | Systems and methods for creating and sharing virtual and augmented experiences |
US10908677B2 (en) * | 2019-03-25 | 2021-02-02 | Denso International America, Inc. | Vehicle system for providing driver feedback in response to an occupant's emotion |
US11586991B2 (en) * | 2019-04-19 | 2023-02-21 | Whitney Skaling | Secure on-demand transportation service |
US10915779B2 (en) * | 2019-04-26 | 2021-02-09 | Unikie Oy | Method for extracting uniform features from point cloud and system therefor |
AU2020326442A1 (en) * | 2019-08-05 | 2022-03-10 | Tellus You Care, Inc. | Non-contact identification of multi-person presence for elderly care |
US11662449B2 (en) * | 2020-06-22 | 2023-05-30 | Honeywell International Inc. | Methods and systems for improving target detection performance of an indoor radar sensor |
US11172339B1 (en) * | 2020-07-11 | 2021-11-09 | Gregory J. Hummer | Method and devices for detecting chemical compositions and biological pathogens |
WO2022136930A1 (en) * | 2020-12-22 | 2022-06-30 | PathPartner Technology Private Limited | System and method for classification of objects in vehicle using feature vectors |
-
2021
- 2021-02-03 WO PCT/IB2021/050852 patent/WO2022136930A1/en active Application Filing
- 2021-02-23 US US17/182,585 patent/US11527080B2/en active Active
- 2021-03-05 EP EP21161064.7A patent/EP4020009A1/de not_active Withdrawn
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160200276A1 (en) * | 2013-08-14 | 2016-07-14 | Iee International Electronics & Engineering S.A. | Radar sensing of vehicle occupancy |
US20200341114A1 (en) * | 2017-03-28 | 2020-10-29 | Sri International | Identification system for subject or activity identification using range and velocity data |
WO2019238575A1 (en) * | 2018-06-11 | 2019-12-19 | Iee International Electronics & Engineering S.A. | Method for robust vehicle occupancy detection with vital sign monitoring |
Non-Patent Citations (3)
Title |
---|
ALIZADEH MOSTAFA ET AL: "Low-cost low-power in-vehicle occupant detection with mm-wave FMCW radar", 2019 IEEE SENSORS, IEEE, 27 October 2019 (2019-10-27), pages 1 - 4, XP033685782, DOI: 10.1109/SENSORS43011.2019.8956880 * |
TEXAS INSTRUMENTS: "Vehicle Occupant Detection Reference Design", 30 April 2018 (2018-04-30), pages 1 - 21, XP055663148, Retrieved from the Internet <URL:https://www.ti.com/lit/ug/tidue95/tidue95.pdf> [retrieved on 20200129] * |
WOHLER CHRISTIAN ET AL: "Comparison of random forest and long short-term memory network performances in classification tasks using radar", 2017 SENSOR DATA FUSION: TRENDS, SOLUTIONS, APPLICATIONS (SDF), IEEE, 10 October 2017 (2017-10-10), pages 1 - 6, XP033274064, DOI: 10.1109/SDF.2017.8126350 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP4328095A1 (de) * | 2022-08-24 | 2024-02-28 | Robert Bosch GmbH | Klassifizierungssysteme und -verfahren zur fahrzeuginternen erfassung mit radar |
Also Published As
Publication number | Publication date |
---|---|
US11527080B2 (en) | 2022-12-13 |
US20220198205A1 (en) | 2022-06-23 |
WO2022136930A1 (en) | 2022-06-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11527080B2 (en) | System and method for classification of objects in vehicle using feature vectors | |
US9671492B2 (en) | Radar sensing of vehicle occupancy | |
EP3546979B1 (de) | System und verfahren zur steuerung des zugangs zu einem kofferraum eines fahrzeugs mittels eines radarsensors | |
CN110023780B (zh) | 用于检测和分类车辆内部车辆乘员和其他对象的极化雷达系统和方法 | |
CN109606357A (zh) | 用于使用毫米波雷达传感器控制车辆的操作的系统和方法 | |
JP6887066B1 (ja) | 電子機器、電子機器の制御方法、及びプログラム | |
JP2003507716A (ja) | ニューラルネットワークレーダプロセッサ | |
Abedi et al. | AI-powered in-vehicle passenger monitoring using low-cost mm-wave radar | |
Li et al. | In-vehicle occupant detection system using mm-wave radar | |
EP4077059B1 (de) | Verfahren zur sitzbesetzungsdetektion | |
CN113740855B (zh) | 占位识别方法、装置、毫米波雷达及存储介质 | |
US20240027597A1 (en) | Electronic device, method for controlling electronic device, and program | |
US20220317246A1 (en) | System and method for automotive in-vehicle applications using cascaded classifiers | |
EP3978947A1 (de) | Elektronische vorrichtung, verfahren zum steuern der elektronischen vorrichtung und programm | |
EP3978946A1 (de) | Elektronische vorrichtung, verfahren zur steuerung einer elektronischen vorrichtung und programm | |
JP7520256B2 (ja) | 乗員状態検知装置 | |
US20230227045A1 (en) | Physique estimation device, physique estimation method, seatbelt reminder system, and airbag control system | |
WO2023002871A1 (ja) | 電子機器、電子機器の制御方法、及びプログラム | |
WO2024185715A1 (ja) | 電子機器、電子機器の制御方法、及びプログラム | |
WO2024142893A1 (ja) | 電子機器、電子機器の制御方法、及びプログラム | |
임소희 | RADAR SENSOR BASED SIGNAL PROCESSING TECHNIQUES FOR INDOOR AND OUTDOOR AUTONOMOUS DRIVING | |
Kamann et al. | Radar-based Environment Perception for Pre-Crash Safety Systems | |
Abedi et al. | On the Use of Low-Cost Radars and Machine Learning for In-Vehicle Passenger Detection |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION HAS BEEN PUBLISHED |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20221228 |
|
RBV | Designated contracting states (corrected) |
Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION HAS BEEN WITHDRAWN |
|
17Q | First examination report despatched |
Effective date: 20240321 |
|
18W | Application withdrawn |
Effective date: 20240411 |