CN108320051B - Mobile robot dynamic collision avoidance planning method based on GRU network model - Google Patents

Mobile robot dynamic collision avoidance planning method based on GRU network model Download PDF

Info

Publication number
CN108320051B
CN108320051B CN201810044018.XA CN201810044018A CN108320051B CN 108320051 B CN108320051 B CN 108320051B CN 201810044018 A CN201810044018 A CN 201810044018A CN 108320051 B CN108320051 B CN 108320051B
Authority
CN
China
Prior art keywords
data
output
mobile robot
layer
input
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810044018.XA
Other languages
Chinese (zh)
Other versions
CN108320051A (en
Inventor
王宏健
刘大伟
秦佳宇
王莹
周佳加
袁建亚
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Engineering University
Original Assignee
Harbin Engineering University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Engineering University filed Critical Harbin Engineering University
Priority to CN201810044018.XA priority Critical patent/CN108320051B/en
Publication of CN108320051A publication Critical patent/CN108320051A/en
Application granted granted Critical
Publication of CN108320051B publication Critical patent/CN108320051B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • G06Q10/047Optimisation of routes or paths, e.g. travelling salesman problem
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/004Artificial life, i.e. computing arrangements simulating life
    • G06N3/006Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/061Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using biological neurons, e.g. biological neurons connected to an integrated circuit
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Business, Economics & Management (AREA)
  • Molecular Biology (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Data Mining & Analysis (AREA)
  • Human Resources & Organizations (AREA)
  • Strategic Management (AREA)
  • Economics (AREA)
  • Neurology (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Development Economics (AREA)
  • Microelectronics & Electronic Packaging (AREA)
  • Game Theory and Decision Science (AREA)
  • Tourism & Hospitality (AREA)
  • Marketing (AREA)
  • Operations Research (AREA)
  • General Business, Economics & Management (AREA)
  • Quality & Reliability (AREA)
  • Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
  • Feedback Control In General (AREA)

Abstract

The invention discloses a dynamic collision avoidance planning method for a mobile robot based on a GRU network model, belonging to the field of mobile robot navigation; the invention relates to a collision avoidance algorithm based on a deep learning network, which is characterized in that data of a sensor are subjected to early-stage normalization processing and then input into a GRU network model, the data are transmitted to a hidden layer through an input layer, the data are processed through a GRU module unit of the hidden layer, and the processed data are output to an output layer to obtain the direction theta and the speed v of a mobile robot in a global coordinate system at the next moment; the robot can have a high intelligent dynamic planning level by using simple sensing equipment under the action of the algorithm, and the reaction speed of the mobile robot is superior to that of the traditional collision avoidance algorithm on the premise of ensuring safety.

Description

Mobile robot dynamic collision avoidance planning method based on GRU network model
Technical Field
The invention relates to the field of mobile robot navigation, in particular to a dynamic collision avoidance planning method for a mobile robot based on a GRU network model.
Background
When the mobile robot works in an unknown environment, the dynamic collision avoidance planning of a local path is carried out by sensing surrounding environment information through the laser range finder, the safety of the path is guaranteed when the planning is carried out, and meanwhile the real-time performance is also met. At present, a plurality of documents about autonomous collision avoidance of mobile robots exist, most collision avoidance methods utilize traditional algorithms or group intelligent methods, and the algorithms need to establish an environment model and iterate the algorithms, so that the calculation time of the algorithms is prolonged, and the real-time performance is influenced. In the modern day with highly developed science and technology, deep learning gradually enters the lives of people, and the rise of unmanned automobiles is the specific application of deep learning networks in the aspects of collision avoidance and path planning. Compared with the traditional algorithm, the deep learning algorithm does not need to construct a model, and after a large amount of training, the strong nonlinear fitting capacity of the deep learning network can express various input and output. For a trained network, the iteration of the algorithm is not needed, and the corresponding output result can be obtained only through forward propagation, so that the real-time performance is good.
For the problem of dynamic collision avoidance of the mobile robot, the result of the planning is not only related to the environmental information at the current moment, but also comprehensively considers the environmental information and the planning result at the previous moments, the input information is related to the time sequence, and in order to process the input data related to the time sequence, a proper network needs to be selected.
Aiming at the problem of robot collision avoidance in an unknown environment, the invention designs a collision avoidance algorithm based on a deep learning network, and under the action of the algorithm, the robot can have a high intelligent dynamic planning level by using simple sensing equipment, so that the reaction speed of the mobile robot is superior to that of the traditional collision avoidance algorithm on the premise of ensuring safety.
Disclosure of Invention
The invention aims to design a collision avoidance algorithm based on a deep learning network aiming at the robot collision avoidance problem in an unknown environment, the robot can have a high intelligent dynamic programming level by using simple sensing equipment under the action of the algorithm, and the reaction speed of the mobile robot is superior to that of the traditional collision avoidance algorithm on the premise of ensuring safety.
A dynamic collision avoidance planning method for a mobile robot based on a GRU network model is characterized by comprising an input layer, a hidden layer and an output layer; the input original data is changed into 61-dimensional data after normalization and combination processing, and the 61-dimensional data sequentially enters an input layer, a hidden layer and an output layer.
The input layer receives input data with 61 dimensions, wherein the first 60-dimensional data is distance information detected by a laser range finder, and the last one-dimensional data is an angle of a connection line between the mobile robot and a target point in a global coordinate system
Figure BDA0001550311750000011
The data is stored in the form of row vectors, and the input layer and the hidden layer are in a full connection mode.
The hidden layer is two layers in total, the hidden layer 1 is composed of 40 GRU module units, the hidden layer 2 comprises 30 neurons, and the connection mode between the hidden layers is full connection.
The output layer comprises 2 neurons which respectively correspond to the direction theta and the speed v of the mobile robot in the global coordinate system at the next planned moment; the hidden layer data output then enters the output layer in a fully connected manner through the activation function.
A dynamic collision avoidance planning method for a mobile robot based on a GRU network model is characterized by comprising the following implementation methods:
the method comprises the following steps: establishing a coordinate system model, wherein the origin of the coordinate system is selected at the center of the laser range finder, the X axis points to the right front of the laser range finder, and the Y axis points to the left of the laser range finder; the horizontal opening angle of the laser range finder is 180 degrees, the maximum detection radius is set to be 8m, 181 beams are totally formed, and the beam angle is 1 degree;
step two: collecting and preprocessing sample data, performing dynamic collision avoidance and path planning by a mobile robot under the control of a teacher system in a simulation training field, collecting the sample data at the same time, wherein each group of sample data has 184 dimensions, then performing normalization processing, merging the collected data from 1 to 181 dimensions, keeping the 182 th dimension unchanged, and inputting the processed data with 61 dimensions;
step three: inputting the 61-dimensional data into an input layer;
step four: the GRU module unit in the hidden layer receives input data from the input layer in time sequence, firstly xt-9Producing an output h after forward propagationt-9,ht-9And xt-8Together asThe next unit inputs, and so on, the data is transmitted in this way, the output of the module at the last moment is the output of the unit structure, and the calculation formula of the GRU output value is:
zt=σ(xtWz+ht-1Uz)
rt=σ(xtWr+ht-1Ur)
Figure BDA0001550311750000021
Figure BDA0001550311750000022
wherein: z is a radical oft、rtRespectively representing the output of an update gate and a reset gate at the time t; x is the number oftAnd htInput and output vectors at time t; wz、WrAnd W is the module input and update gate, reset gate and
Figure BDA0001550311750000023
an inter-weight matrix; u shapez、UrAnd U is the output of the module at the time t-1, the refresh gate, the reset gate and
Figure BDA0001550311750000024
an inter-weight matrix;
step five: transmitting the calculation result of the hidden layer to an output layer to be used as the direction theta and the speed v of the mobile robot in the global coordinate system at the next moment;
step six: training a GRU network model; initializing the weight of the GRU network model by using standard positive-space distribution;
step seven: training times k is 0, i is 0;
step eight: randomly extracting data [ x ] of 10 continuous time instants from a training sett-9,…,xt]Input network for obtaining corresponding output y through forward propagation processt
Step nine: computing network output ytAnd a corresponding label ltAn error of (2);
step ten: updating the weight by using a gradient descent algorithm;
step eleven: if i is more than or equal to 500, making i equal to 0, randomly selecting 50 sequences in the test set, testing by using the current network, calculating the mean square error, and stopping training if the mean square error in the test set is not reduced for 10 times continuously;
step twelve: k +1, i + 1;
step thirteen: if k is larger than the maximum training times, jumping to the step fourteen, otherwise, jumping to the step eight;
fourteen steps: finishing the training; and obtaining the trained GRU network model.
The invention has the beneficial effects that:
after the network training is finished, the mobile robot uses the training result of the GRU network model to perform dynamic collision avoidance planning in an unknown environment and compares the dynamic collision avoidance planning with the ant colony algorithm, the two algorithms can enable the mobile robot to avoid all obstacles, the route is smooth, and the algorithm can guarantee the safety of the robot. However, the reaction time of the mobile robot is short under the GRU network model, the real-time performance is higher, and the high efficiency of the algorithm is proved.
Drawings
FIG. 1 is a diagram of a coordinate system of a mobile robot according to the present invention;
FIG. 2 shows the angle θ of the present invention
Figure BDA0001550311750000031
A schematic diagram;
FIG. 3 is a diagram of a GRU-based dynamic path planning model according to the present invention;
FIG. 4 is a time-series expanded view of a single GRU modular unit of the present invention;
fig. 5 is a diagram of the effect of GRU network and ant colony algorithm planning in the present invention.
Detailed Description
The invention is further described below with reference to the accompanying drawings.
And (3) constructing a coordinate system model, as shown in fig. 1, wherein a global coordinate system adopts a northeast coordinate system, the lower left corner of a map is an origin, the east-righting direction is an X-axis, the north-righting direction is a Y-axis, the origin of a local coordinate system is selected at the center of the laser range finder, the X-axis points to the front of the laser range finder, and the Y-axis points to the left of the laser range finder. The horizontal opening angle of the laser range finder is 180 degrees, the maximum detection radius is 8m, 181 wave beams are totally formed, and the wave beam angle is 1 degree.
Collecting and processing sample data, wherein before training, sample data collection and processing are firstly carried out, a mobile robot carries out dynamic collision avoidance and path planning under the control of a teacher system in a simulation training field, and simultaneously sample data are collected, wherein each group of sample data has 184 dimensions, the data of the first 181 dimensions are obtained by collecting surrounding environment information by a laser range finder, the sensor returns 181 points representing distances, the data of the second 182 dimensions are angle information of a connecting line of the mobile robot and a target point in a global coordinate system, as shown in figure 2,
Figure BDA0001550311750000041
namely, 182 th dimension data. The last two-dimensional data are the robot's heading angle θ and velocity v. Where the first 182-dimensional data are input values to the network and the last two-dimensional data are training labels. In order to make the training result more accurate, the data must be correspondingly processed before entering the input layer, firstly, normalization processing is carried out, then, the data of 1 to 181 dimensions are merged, finally, the data of 182 th dimension is kept unchanged, and the input data has 61 dimensions after processing.
The mobile robot dynamically plans the structural design of the GRU network model, and the structural design of the GRU network model designed by the invention is shown in figure 3. According to the forward propagation direction, input original data are changed into 61-dimension data after normalization and combination processing, and the input data sequentially enter an input layer, a hidden layer and an output layer.
An input layer: the input data of the layer is 61-dimensional, wherein the first 60-dimensional data is distance information detected by a laser range finder, and the last one-dimensional data is an angle of a connection line of the mobile robot and a target point in a global coordinate system
Figure BDA0001550311750000042
The data is stored in the form of row vectors, and the input layer and the hidden layer are in a full connection mode.
Hiding the layer: the hidden layer is two layers in total, hidden layer 1 comprises 40 GRU modular unit, and hidden layer 2 contains 30 neurons, and the connection mode between the hidden layer is full connection.
An output layer: the number of neurons in the output layer is 2, and the neurons respectively correspond to the direction theta and the speed v of the mobile robot in the global coordinate system at the next planned moment. The hidden layer data output then enters the output layer in a fully connected manner through the activation function.
The most important structure in the above-mentioned GRU network model is the GRU module unit, each unit includes two gate control units, namely an update gate and a reset gate, and the inputs of the two gates are the current time input xtAnd the output h of the previous momentt-1The long-term information can be retained by the cooperation of the two doors. The calculation method of the unit output is shown as the following formula:
zt=σ(xtWz+ht-1Uz)
rt=σ(xtWr+ht-1Ur)
Figure BDA0001550311750000043
Figure BDA0001550311750000044
wherein: z is a radical oft、rtRespectively representing the output of an update gate and a reset gate at the time t; x is the number oftAnd htInput and output vectors at time t; wz、WrAnd W is the module input and update gate, reset gate and
Figure BDA0001550311750000045
an inter-weight matrix; u shapez、UrAnd U is the output and update gate of the module at the time t-1,Reset gate and
Figure BDA0001550311750000046
the weight matrix of (2).
The GRU module units are expanded according to time sequence as shown in FIG. 4, the GRU module units receive input data according to time sequence, firstly, xt-9Producing an output h after forward propagationt-9,ht-9And xt-8The data are transmitted in this way, and the output of the module at the last moment is the output of the unit structure.
GRU network model training process:
the method comprises the following steps: initializing the weight of the GRU network model by using standard positive-space distribution;
step two: training times k is 0, i is 0;
step three: randomly extracting data [ x ] of 10 continuous time instants from a training sett-9,…,xt]Input network for obtaining corresponding output y through forward propagation processt
Step four: computing network output ytAnd a corresponding label ltAn error of (2);
step five: updating the weight by using a gradient descent algorithm;
step six: if i is more than or equal to 500, making i equal to 0, randomly selecting 50 sequences in the test set, testing by using the current network, calculating the mean square error, and stopping training if the mean square error in the test set is not reduced for 10 times continuously;
step seven: k +1, i + 1;
step eight: if k is larger than the maximum training times, jumping to the step nine, otherwise, jumping to the step three;
step nine: and finishing the training.
The invention has the following effects:
after the network training is finished, the mobile robot uses the training result of the GRU network model to perform dynamic collision avoidance planning in an unknown environment and compares the dynamic collision avoidance planning with the ant colony algorithm, the result is shown in fig. 5, and as can be seen from the figure, the two algorithms can enable the mobile robot to avoid all obstacles, and the route is smooth, which indicates that the algorithm can ensure the safety of the robot. The average reaction time required by the robot for collision avoidance under the two algorithms is 236ms and 391ms respectively, so that the reaction time of the mobile robot under the GRU network model is obviously short, the real-time performance is higher, and the high efficiency of the algorithm is proved.

Claims (4)

1. A dynamic collision avoidance planning method for a mobile robot based on a GRU network model is characterized by comprising an input layer, a hidden layer and an output layer; input original data are changed into 61-dimensional data after normalization and combination processing, and the input data sequentially enter an input layer, a hidden layer and an output layer; the realization method comprises the following steps:
the method comprises the following steps: establishing a coordinate system model, wherein the origin of the coordinate system is selected at the center of the laser range finder, the X axis points to the right front of the laser range finder, and the Y axis points to the left of the laser range finder; the horizontal opening angle of the laser range finder is 180 degrees, the maximum detection radius is set to be 8m, 181 beams are totally formed, and the beam angle is 1 degree;
step two: collecting and preprocessing sample data, performing dynamic collision avoidance and path planning by a mobile robot under the control of a teacher system in a simulation training field, collecting the sample data at the same time, wherein each group of sample data has 184 dimensions, then performing normalization processing, merging the collected data from 1 to 181 dimensions, keeping the 182 th dimension unchanged, and inputting the processed data with 61 dimensions;
step three: inputting the 61-dimensional data into an input layer;
step four: the GRU module unit in the hidden layer receives input data from the input layer in time sequence, firstly xt-9Producing an output h after forward propagationt-9,ht-9And xt-8The data are transmitted in this way, the output of the module at the last moment is the output of the unit structure, and the calculation formula of the output value of the GRU is as follows:
zt=σ(xtWz+ht-1Uz)
rt=σ(xtWr+ht-1Ur)
Figure FDA0003267658980000011
Figure FDA0003267658980000012
wherein: z is a radical oft、rtRespectively representing the output of an update gate and a reset gate at the time t; x is the number oftAnd htInput and output vectors at time t; wz、WrAnd W is the module input and update gate, reset gate and
Figure FDA0003267658980000013
an inter-weight matrix; u shapez、UrAnd U is the output of the module at the time t-1, the refresh gate, the reset gate and
Figure FDA0003267658980000014
an inter-weight matrix;
step five: transmitting the calculation result of the hidden layer to an output layer to be used as the direction theta and the speed v of the mobile robot in the global coordinate system at the next moment;
step six: training a GRU network model; initializing the weight of the GRU network model by using standard positive-space distribution;
step seven: training times k is 0, i is 0;
step eight: randomly extracting data [ x ] of 10 continuous time instants from a training sett-9,…,xt]Input network for obtaining corresponding output y through forward propagation processt
Step nine: computing network output ytAnd a corresponding label ltAn error of (2);
step ten: updating the weight by using a gradient descent algorithm;
step eleven: if i is more than or equal to 500, making i equal to 0, randomly selecting 50 sequences in the test set, testing by using the current network, calculating the mean square error, and stopping training if the mean square error in the test set is not reduced for 10 times continuously;
step twelve: k +1, i + 1;
step thirteen: if k is larger than the maximum training times, jumping to the step fourteen, otherwise, jumping to the step eight;
fourteen steps: finishing the training; and obtaining the trained GRU network model.
2. The dynamic collision avoidance planning method for the mobile robot based on the GRU network model as claimed in claim 1, wherein the input layer receives input data with 61 dimensions, the first 60 dimensions being distance information detected by the laser range finder, and the last one dimensions being angles of a connection line between the mobile robot and a target point in a global coordinate system
Figure FDA0003267658980000021
The data is stored in the form of row vectors, and the input layer and the hidden layer are in a full connection mode.
3. The dynamic collision avoidance planning method for the mobile robot based on the GRU network model as claimed in claim 1, wherein the hidden layer has two layers, the hidden layer 1 is composed of 40 GRU module units, the hidden layer 2 contains 30 neurons, and the connection mode between the hidden layers is full connection.
4. The dynamic GRU network model-based collision avoidance planning method for the mobile robot according to claim 1, wherein the output layer comprises 2 neurons respectively corresponding to the planned direction θ and speed v of the mobile robot at the next moment in the global coordinate system; the hidden layer data output then enters the output layer in a fully connected manner through the activation function.
CN201810044018.XA 2018-01-17 2018-01-17 Mobile robot dynamic collision avoidance planning method based on GRU network model Active CN108320051B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810044018.XA CN108320051B (en) 2018-01-17 2018-01-17 Mobile robot dynamic collision avoidance planning method based on GRU network model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810044018.XA CN108320051B (en) 2018-01-17 2018-01-17 Mobile robot dynamic collision avoidance planning method based on GRU network model

Publications (2)

Publication Number Publication Date
CN108320051A CN108320051A (en) 2018-07-24
CN108320051B true CN108320051B (en) 2021-11-23

Family

ID=62895127

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810044018.XA Active CN108320051B (en) 2018-01-17 2018-01-17 Mobile robot dynamic collision avoidance planning method based on GRU network model

Country Status (1)

Country Link
CN (1) CN108320051B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111027313A (en) * 2018-10-08 2020-04-17 中国科学院沈阳计算技术研究所有限公司 BiGRU judgment result tendency analysis method based on attention mechanism
CN109556609B (en) * 2018-11-15 2020-11-17 武汉南华工业设备工程股份有限公司 Artificial intelligence-based collision avoidance method and device
CN109739218A (en) * 2018-12-24 2019-05-10 江苏大学 It is a kind of that outstanding driver's lane-change method for establishing model is imitated based on GRU network
CN109901589B (en) * 2019-03-29 2022-06-07 北京易达图灵科技有限公司 Mobile robot control method and device
CN117031365A (en) * 2023-08-02 2023-11-10 迈胜医疗设备有限公司 Magnetic field measuring device, control method thereof, electronic equipment and related device

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106875511A (en) * 2017-03-03 2017-06-20 深圳市唯特视科技有限公司 A kind of method for learning driving style based on own coding regularization network
CN106950969A (en) * 2017-04-28 2017-07-14 深圳市唯特视科技有限公司 It is a kind of based on the mobile robot continuous control method without map movement planner
CN107065881A (en) * 2017-05-17 2017-08-18 清华大学 A kind of robot global path planning method learnt based on deeply
CN107122736A (en) * 2017-04-26 2017-09-01 北京邮电大学 A kind of human body based on deep learning is towards Forecasting Methodology and device
CN107229967A (en) * 2016-08-22 2017-10-03 北京深鉴智能科技有限公司 A kind of hardware accelerator and method that rarefaction GRU neutral nets are realized based on FPGA
CN108334677A (en) * 2018-01-17 2018-07-27 哈尔滨工程大学 A kind of UUV Realtime collision free planing methods based on GRU networks

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103926927A (en) * 2014-05-05 2014-07-16 重庆大学 Binocular vision positioning and three-dimensional mapping method for indoor mobile robot
CN105045260A (en) * 2015-05-25 2015-11-11 湖南大学 Mobile robot path planning method in unknown dynamic environment
CN106774942A (en) * 2017-01-18 2017-05-31 华南理工大学 A kind of real-time 3D remote human-machines interactive system

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107229967A (en) * 2016-08-22 2017-10-03 北京深鉴智能科技有限公司 A kind of hardware accelerator and method that rarefaction GRU neutral nets are realized based on FPGA
CN106875511A (en) * 2017-03-03 2017-06-20 深圳市唯特视科技有限公司 A kind of method for learning driving style based on own coding regularization network
CN107122736A (en) * 2017-04-26 2017-09-01 北京邮电大学 A kind of human body based on deep learning is towards Forecasting Methodology and device
CN106950969A (en) * 2017-04-28 2017-07-14 深圳市唯特视科技有限公司 It is a kind of based on the mobile robot continuous control method without map movement planner
CN107065881A (en) * 2017-05-17 2017-08-18 清华大学 A kind of robot global path planning method learnt based on deeply
CN108334677A (en) * 2018-01-17 2018-07-27 哈尔滨工程大学 A kind of UUV Realtime collision free planing methods based on GRU networks

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
A Novel GRU-RNN Network Model for Dynamic;JIANYA YUAN等;《IEEEAccess》;20190124;第7卷;第15140-15151 *

Also Published As

Publication number Publication date
CN108320051A (en) 2018-07-24

Similar Documents

Publication Publication Date Title
CN114384920B (en) Dynamic obstacle avoidance method based on real-time construction of local grid map
CN108320051B (en) Mobile robot dynamic collision avoidance planning method based on GRU network model
Xie et al. Motion trajectory prediction based on a CNN-LSTM sequential model
CN108319293B (en) UUV real-time collision avoidance planning method based on LSTM network
Michelmore et al. Uncertainty quantification with statistical guarantees in end-to-end autonomous driving control
CN108334677B (en) UUV real-time collision avoidance planning method based on GRU network
Cao et al. Target search control of AUV in underwater environment with deep reinforcement learning
CN110782481A (en) Unmanned ship intelligent decision method and system
CN109634309B (en) Autonomous obstacle avoidance system and method for aircraft and aircraft
CN113435644B (en) Emergency prediction method based on deep bidirectional long-short term memory neural network
CN111695299B (en) Mesoscale vortex trajectory prediction method
CN113848974B (en) Aircraft trajectory planning method and system based on deep reinforcement learning
CN111445498A (en) Target tracking method adopting Bi-L STM neural network
CN111507161B (en) Method and device for heterogeneous sensor fusion by utilizing merging network
CN111027627A (en) Vibration information terrain classification and identification method based on multilayer perceptron
CN112857370A (en) Robot map-free navigation method based on time sequence information modeling
Zhang et al. A convolutional neural network method for self-driving cars
Masmoudi et al. Autonomous car-following approach based on real-time video frames processing
Xu et al. Avoidance of manual labeling in robotic autonomous navigation through multi-sensory semi-supervised learning
CN108459614B (en) UUV real-time collision avoidance planning method based on CW-RNN network
CN114495036A (en) Vehicle track prediction method based on three-stage attention mechanism
CN114089751A (en) Mobile robot path planning method based on improved DDPG algorithm
CN108639177A (en) A kind of autonomous full traversal climbing robot
CN112560571A (en) Intelligent autonomous visual navigation method based on convolutional neural network
Li et al. Vision-based obstacle avoidance algorithm for mobile robot

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant