CN114610070A - Unmanned aerial vehicle-cooperated wind power plant intelligent inspection method - Google Patents
Unmanned aerial vehicle-cooperated wind power plant intelligent inspection method Download PDFInfo
- Publication number
- CN114610070A CN114610070A CN202210274635.5A CN202210274635A CN114610070A CN 114610070 A CN114610070 A CN 114610070A CN 202210274635 A CN202210274635 A CN 202210274635A CN 114610070 A CN114610070 A CN 114610070A
- Authority
- CN
- China
- Prior art keywords
- unmanned aerial
- aerial vehicle
- wind
- wind power
- power plant
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000007689 inspection Methods 0.000 title claims abstract description 42
- 238000000034 method Methods 0.000 title claims abstract description 35
- 238000005265 energy consumption Methods 0.000 claims abstract description 38
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 31
- 238000013439 planning Methods 0.000 claims abstract description 28
- 230000002787 reinforcement Effects 0.000 claims abstract description 21
- 238000002922 simulated annealing Methods 0.000 claims abstract description 15
- 238000012545 processing Methods 0.000 claims abstract description 6
- 238000013528 artificial neural network Methods 0.000 claims description 14
- 238000010586 diagram Methods 0.000 claims description 12
- 238000012549 training Methods 0.000 claims description 9
- XOFYZVNMUHMLCC-ZPOLXVRWSA-N prednisone Chemical compound O=C1C=C[C@]2(C)[C@H]3C(=O)C[C@](C)([C@@](CC4)(O)C(=O)CO)[C@@H]4[C@@H]3CCC2=C1 XOFYZVNMUHMLCC-ZPOLXVRWSA-N 0.000 claims description 8
- 238000007781 pre-processing Methods 0.000 claims description 7
- 238000004364 calculation method Methods 0.000 claims description 6
- 238000011478 gradient descent method Methods 0.000 claims description 6
- 238000012423 maintenance Methods 0.000 claims description 5
- 238000000137 annealing Methods 0.000 claims description 4
- 238000000354 decomposition reaction Methods 0.000 claims description 3
- 230000005484 gravity Effects 0.000 claims description 3
- 238000009499 grossing Methods 0.000 claims description 3
- 238000010606 normalization Methods 0.000 claims description 3
- 230000007704 transition Effects 0.000 claims description 3
- 238000013024 troubleshooting Methods 0.000 claims description 3
- 238000005457 optimization Methods 0.000 claims 1
- 230000007613 environmental effect Effects 0.000 abstract description 4
- 238000012544 monitoring process Methods 0.000 abstract description 2
- 238000005516 engineering process Methods 0.000 description 5
- 230000005611 electricity Effects 0.000 description 4
- 238000013459 approach Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000007726 management method Methods 0.000 description 2
- 238000004519 manufacturing process Methods 0.000 description 2
- 238000010248 power generation Methods 0.000 description 2
- 206010063385 Intellectualisation Diseases 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000013523 data management Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 230000003203 everyday effect Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 239000002699 waste material Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/10—Simultaneous control of position or course in three dimensions
- G05D1/101—Simultaneous control of position or course in three dimensions specially adapted for aircraft
-
- H—ELECTRICITY
- H02—GENERATION; CONVERSION OR DISTRIBUTION OF ELECTRIC POWER
- H02J—CIRCUIT ARRANGEMENTS OR SYSTEMS FOR SUPPLYING OR DISTRIBUTING ELECTRIC POWER; SYSTEMS FOR STORING ELECTRIC ENERGY
- H02J13/00—Circuit arrangements for providing remote indication of network conditions, e.g. an instantaneous record of the open or closed condition of each circuitbreaker in the network; Circuit arrangements for providing remote control of switching means in a power distribution network, e.g. switching in and out of current consumers by using a pulse code signal carried by the network
- H02J13/00002—Circuit arrangements for providing remote indication of network conditions, e.g. an instantaneous record of the open or closed condition of each circuitbreaker in the network; Circuit arrangements for providing remote control of switching means in a power distribution network, e.g. switching in and out of current consumers by using a pulse code signal carried by the network characterised by monitoring
Landscapes
- Engineering & Computer Science (AREA)
- Power Engineering (AREA)
- Aviation & Aerospace Engineering (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Automation & Control Theory (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
- Supply And Distribution Of Alternating Current (AREA)
Abstract
The invention provides an unmanned aerial vehicle cooperative wind power plant intelligent inspection method, and belongs to the technical field of data processing. According to the invention, the unmanned aerial vehicle carries a task load to complete the routing inspection work of the wind turbine generator, the flight path of the unmanned aerial vehicle is planned through a deep reinforcement learning-simulated annealing algorithm model, and the routing inspection route planning of the intelligent wind power plant with the lowest energy consumption is realized. The invention fully considers the physical and environmental characteristics of the wind power plant, realizes the routing planning of the intelligent wind power plant inspection route with low energy consumption, has strong adaptability, and can be applied to the wind power plants in different geographic positions and terrains. The invention not only considers the characteristic of convenient real-time charging in the wind power plant, but also fully considers the climate characteristic of the wind power plant, and innovatively takes the wind speed and the wind direction into consideration of the unmanned aerial vehicle track planning; the wind power station monitoring system can adapt to wind power stations in different terrains and monsoon areas, can timely cope with sudden meteorological changes, dynamically adjusts a routing inspection route, and is very suitable for meteorological characteristics of the wind power stations.
Description
Technical Field
The invention belongs to the technical field of data processing, and relates to an intelligent inspection method for an unmanned aerial vehicle-cooperated wind power plant. According to the invention, the unmanned aerial vehicle carries a task load to complete the routing inspection work of the wind turbine generator, the flight path of the unmanned aerial vehicle is planned through a deep reinforcement learning-simulated annealing algorithm model, and the routing inspection route planning of the intelligent wind power plant with the lowest energy consumption is realized.
Background
With the rapid development of wind power industry in China, the number of wind power generation sets is increased sharply. However, wind farms are often in a complex natural environment, possibly in remote mountainous areas or offshore, and are exposed to long periods of inclement weather. The wind motors are separated by hundreds of meters, the turbine cabin is as high as tens of meters, and the manual inspection cost is high, the efficiency is low, the error rate is high, and the danger is high. How to reduce the operation and maintenance management cost of huge wind turbine generator system, impel wind power system intellectuality and informationization to promote the electricity generation income of wind-powered electricity generation field, become the important problem that the wind-powered electricity generation industry is waited to solve urgently.
With the introduction and the proposal of an intelligent wind power plant concept, the wind power system is comprehensively optimized and upgraded in the aspects of part manufacturing, data management, operation and maintenance and the like. The intelligent wind power plant mainly realizes intellectualization in aspects of fan control management, equipment state perception, routing inspection maintenance and the like based on a measurement and control technology, a communication technology, a sensing technology, a big data processing technology and various intelligent algorithms. The working state data of each device is obtained through various sensors in the wind motor, the sensor data are integrated and processed through an edge server in the wind motor, whether the wind motor is in a normal working state or not can be obtained, and possible reasons of faults can be found out when the wind motor breaks down.
However, wind power systems have not achieved true intelligence. The operation data of the fan equipment still needs to be dumped and maintained by manpower generally; meanwhile, cracks on the wind turbine such as blades are difficult to confirm through a sensor, and manual work is usually required to be matched with a hanging basket or a high-power telescope for routing inspection. This leads to wind power system's system of patrolling and examining still to waste time and energy and the rate of accuracy is low, is difficult to reach the operation requirement of wisdom wind-powered electricity generation field.
Due to the flexible and mobile characteristics, unmanned aerial vehicles are playing an increasingly important role in smart wind farms. Carry the task load through unmanned aerial vehicle and accomplish the work of patrolling and examining the fan, can increase substantially the efficiency and the degree of accuracy of the work of patrolling and examining, reduce the danger that the manpower was patrolled and examined to improve power production's economic benefits. Has great research significance and practical value.
Disclosure of Invention
The method solves the problem of how to plan the routing inspection route of the unmanned aerial vehicle by using meteorological data of the environment of the wind power plant and using a depth reinforcement learning algorithm, so that the energy consumed by the planned route while traversing all wind power generation sets is minimum. In the process, the wind turbine generator is photographed by adopting an image recognition method for fault diagnosis of the wind turbine generator, and meanwhile, the related data of the wind turbine generator is uploaded; the time cost and the labor cost in the process of inspecting personnel are saved, and the safety of the wind turbine generator is improved. The invention realizes the intelligent inspection method of the wind power plant with the cooperation of the unmanned aerial vehicle, combines the technologies of deep reinforcement learning and the like, and finally provides theoretical basis and practical experience for the unmanned aerial vehicle in the field of intelligent inspection of the wind power plant with low power consumption.
The technical scheme of the invention is as follows:
an unmanned aerial vehicle cooperative wind power plant intelligent inspection method comprises an unmanned aerial vehicle route planning system based on deep reinforcement learning and simulated annealing, and a wind power plant fault detection and wind power data uploading system based on an unmanned aerial vehicle. The method comprises the following specific steps:
the method comprises the following steps: acquiring meteorological prediction data of a wind power plant in the future 4 hours, and preprocessing the data.
Step two: for a plurality of wind turbine generators X ═ X under the same wind power plant1,x2,...,xnAnd fourthly, carrying out depth reinforcement learning algorithm on any two wind turbine generators x within the maximum cruising radius of the unmanned aerial vehiclei、xj(i is not equal to j) planning the lowest power consumption flight path and calculating the corresponding power consumption Eij。
Step three: and determining the starting position of the unmanned aerial vehicle according to the predicted weather main wind direction, and performing flight path planning based on a simulated annealing algorithm on the unmanned aerial vehicle according to the current weather data and the learning experience of the step two.
Step four: and D, according to the planned flight path in the step three, each unit of the wind power plant is inspected so as to facilitate troubleshooting and data uploading.
The data preprocessing in the first step comprises the following specific steps:
step 1.1: and checking the meteorological data, and if the meteorological data have missing parts, smoothing the missing data.
Step 1.2: for any time t, according to wind direction data thetatWill wind speedOrthogonal decomposition is carried out to decompose the wind speed into the wind speeds in three mutually perpendicular directions in the three-dimensional space
Step 1.3: and carrying out data normalization processing on the wind direction and wind speed data.
In the second step, the deep reinforcement learning algorithm is constructed by the following steps:
step 2.1: firstly, establishing a Markov decision process model related to the planning of tracks of two wind turbine generators by an unmanned aerial vehicle, and determining a quintuple in the process<S,A,P,R,γ>Wherein S represents the current environmental state quantity of the unmanned aerial vehicle, A is the action quantity executed by the unmanned aerial vehicle, P is the transition probability among different states, R is the reward quantity obtained when the unmanned aerial vehicle executes the action A in the state S, and gamma is the reinforcement learning attenuation rate. The state quantity S can completely represent the current state of the unmanned aerial vehicle, in the invention, a three-dimensional coordinate system is established according to the current position of the unmanned aerial vehicle, and the state quantity S comprises the position coordinate Pos of the current unmanned aerial vehicleU(x, y, z), and the wind velocity vector of the location where the drone is located at the current time
Simplifying the unmanned aerial vehicle motion model, wherein the action quantity A which can be executed by the unmanned aerial vehicle is equal to<a>Indicating the unmanned plane is at speed(Vector)Moving a distance in a fixed direction within a time slice τ. In order to enable the unmanned aerial vehicle to reach the target unit with the least power consumption, the following rewarding modes are designed:
wherein, | ds′L is the linear distance from the target point after the action is executed, | dsI is the linear distance from the target point before the action is performed, Ess′Energy consumption for this execution of action, EmaxThe maximum energy consumption of the unmanned aerial vehicle. When the unmanned aerial vehicle approaches the target with less energy consumption, more reward values can be obtained, and when the unmanned aerial vehicle reaches the target, a great reward value can be obtained, so that the target can absorb the unmanned aerial vehicle.
Energy consumption E required by unmanned aerial vehicle to execute action each timess′The calculation formula is Ess′=Pu·τ。PuFor unmanned aerial vehicle power, including horizontal flight powerVertical flight powerAnd resistance power Wherein W ═ mg is drone gravity; ρ is the air density;is the total area of the rotor of the unmanned aerial vehicle; cD0Is the drag coefficient related to rotor geometry;andthe relative speed of the unmanned aerial vehicle to the wind speed in the horizontal direction and the vertical direction respectively;the speed of horizontal flight of the unmanned aerial vehicle;hovering power for the unmanned aerial vehicle.
Step 2.2: and initializing a playback experience pool D for storing data generated by the unmanned aerial vehicle in a trial and error process. Initializing an Actor real network mu and a Critic real network Q randomly, wherein the parameters corresponding to the two neural networks are theta respectivelyμAnd thetaQ(ii) a Randomly initializing an Actor target network mu 'and a criticic target network Q', wherein the parameters corresponding to the two neural networks are theta respectivelyμ′And thetaQ′And make thetaμ′=θμ,θQ′=θQ。
Step 2.3: recording the initial state quantity s1A random noise N conforming to a gaussian distribution is generated.
Step 2.4: the state quantity x at the current moment is calculatedtInput to at thetaμAdding random noise N at the current moment into the actual Actor network of the parametert. Outputting the action quantity a by the Actor real networkt=μ(xt,θμ)+NtExecuting the action and calculating the reward r obtained by the action through a reward functiontWhile updating the state quantities to obtain xt+1。
Step 2.5: creating quadruplets<xt,at,rt,xt+1>It is stored in the playback experience pool D.
Step 2.6: randomly selecting a group of data from a playback experience pool D<xj,aj,rj,xj+1>X is to bej,ajInput to Critic real networkIn the formula, Q ═ Q (x) is obtainedj,aj,θQ) X is to bej+1Inputting the calculated action amount a into the Actor target networkj+1=μ′(xj+1,θμ′) And x isj+1And aj+1Input into Critic target network together to obtain Q (x)j+1,aj+1,θQ) Then the target value Q' is xj+γQ′(xj+1,aj+1,θQ′) Training the Critic reality network to make the calculated Q value infinitely close to the target value Q 'by using Q' as a label, and updating theta by using a gradient descent methodQ。
Step 2.7: and updating the actual network of the Actor to ensure that the Q value of the output action quantity is maximum in the computed Q value of the Critic actual network. Updating theta by gradient descent methodμThe strategy gradient calculation method is
Step 2.8: updating theta mu' ← alpha theta for target network parameterμ′+(1-α)θμ,θQ′←αθQ′+(1-α)θQ。
Step 2.9: and (5) repeating the step 2.3 to the step 2.8 until the loss values of the Actor target network and the Critic target network are converged and the network parameters are unchanged. After the network converges, for any two wind generating sets xi、xj(i ≠ j), the deep reinforcement learning model can give the flight track with the minimum power consumption of the unmanned aerial vehicle and the power consumption Eij。
In the third step, the simulated annealing algorithm comprises the following specific steps:
step 3.1: and determining the starting position of the unmanned aerial vehicle according to the current day rose diagram. If the current wind power plant has a dominant wind direction, setting the starting position of the unmanned aerial vehicle as a corner fan opposite to the dominant wind direction; if the current dominant wind direction is not obvious, then set unmanned aerial vehicle initial position as the fan of central point position.
Step 3.2: two neural networks on the same day are trained. According to the playback experience pool and the next weather in the step twoThe predicted wind direction and wind speed are used for determining any two wind generating sets x in the maximum cruising radius of the unmanned aerial vehiclei、xjFlight path and flight energy consumption.
Step 3.3: starting from the initial position, sequentially selecting the lowest energy consumption track within the cruising radius of the unmanned aerial vehicle until all fans are traversed, and taking the lowest energy consumption track as an initialization path c. Simultaneously initializing a start temperature T and an end temperature T0And an annealing speed alpha.
Step 3.4: by random thermal perturbation, another path c' is generated in the neighborhood of c. Unlike conventional simulated annealing algorithms, this disturbance can only occur between fans located in the same cruise area.
Step 3.5: and calculating the difference value delta E of the energy consumption of the unmanned aerial vehicle between the two paths c and c'. If Δ E is less than or equal to 0, updating the path, and making c ═ c', T ← α T; otherwise, a random number rand between 0-1 is generated, ifThen the path is updated, let c ═ c', T ← α T.
Step 3.6: judging T > T0If yes, continuing to execute the step 3.4; otherwise, a near-optimal solution of the routing inspection route planning with the lowest energy consumption is obtained, and the unmanned aerial vehicle carries out maintenance and data acquisition on the fan in sequence according to the routing inspection route.
Step 3.7: and judging whether the routing inspection route needs to be updated or not based on the real-time meteorological data. When the real-time wind direction detected by a wind sensor in the wind power plant and the wind direction predicted by the meteorology are not in the same direction or the real-time wind speed and the wind speed level predicted by the meteorology are different by more than two levels, counting the wind turbine set X' ═ X which is not traversed yet1,x2,...,xn′}. And 3.2, re-executing the step, and re-planning the lowest energy consumption flight path of the unmanned aerial vehicles of the rest wind turbine generators. And ending the algorithm until the unmanned aerial vehicle traverses the wind turbine generator.
The invention has the beneficial effects that: compared with other unmanned aerial vehicle track planning methods, the method provided by the invention fully considers the physical and environmental characteristics of the wind power plant, realizes the routing of the intelligent wind power plant inspection route with low energy consumption, has strong adaptability, and can be applied to wind power plants in different geographic positions and terrains. According to the invention, the routing of the unmanned aerial vehicle in the wind power plant is planned by developing an energy-saving routing inspection method in the intelligent wind power plant, so that the characteristic of convenience for real-time charging in the wind power plant is considered, the climatic characteristics of the wind power plant are also fully considered, and the wind speed and the wind direction are innovatively taken into consideration of the unmanned aerial vehicle track planning; the wind power station monitoring system can adapt to wind power stations in different terrains and monsoon areas, can timely cope with sudden meteorological changes, dynamically adjusts a routing inspection route, and is very suitable for meteorological characteristics of the wind power stations.
Drawings
FIG. 1 is a diagram of a wind farm intelligent patrol scenario of the present invention.
FIG. 2 is a timing diagram of the intelligent inspection of a wind farm of the present invention.
FIG. 3 is a flow chart of the data preprocessing of the present invention.
Fig. 4 is a schematic diagram of the unmanned aerial vehicle track planning algorithm based on reinforcement learning.
Fig. 5 is a detailed design diagram of the unmanned aerial vehicle track planning algorithm based on reinforcement learning.
Fig. 6 is a flow chart of the unmanned aerial vehicle flight path planning algorithm based on simulated annealing.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, embodiments of the present invention will be described in detail below with reference to the accompanying drawings.
FIG. 1 is a diagram of a wind farm intelligent patrol scenario of the present invention.
FIG. 2 is a timing diagram of intelligent inspection of a wind farm according to the present invention, which provides an intelligent inspection method of a wind farm with cooperation of unmanned aerial vehicles, and specifically includes the following steps:
the method comprises the following steps: and acquiring meteorological data of the wind power plant in the future 4 hours, and preprocessing the data.
Step two: for a plurality of wind turbine generators X ═ X under the same wind power plant1,x2,...,xnAnd fourthly, carrying out depth reinforcement learning algorithm on any two winds within the maximum cruising radius of the unmanned aerial vehicleMotor unit xi、xj(i ≠ j) carries out the lowest power consumption track planning and calculates the corresponding power consumption Eij。
Step three: and determining the starting position of the unmanned aerial vehicle according to the predicted weather main wind direction, and performing flight path planning based on a simulated annealing algorithm on the unmanned aerial vehicle according to the current weather data and the learning experience of the step two.
Step four: and D, according to the planned flight path in the step three, each unit of the wind power plant is patrolled so as to facilitate troubleshooting and data uploading.
Fig. 3 is a flow chart of data preprocessing in the first step of the present invention, which specifically includes the following steps:
step 1.1: and checking the meteorological data, and if the meteorological data have missing parts, smoothing the missing data.
Step 1.2: for any time t, according to wind direction data thetatWill wind speedOrthogonal decomposition is carried out to decompose the wind speed into the wind speeds in three mutually perpendicular directions in the three-dimensional space
Step 1.3: and carrying out normalization processing on the wind direction and the wind speed data.
Before the deep learning-simulated annealing algorithm of the present invention is explained in detail, the following description is given to the problem:
firstly, according to physical parameters and geographical environment of a wind power plant, relevant parameters of an unmanned aerial vehicle and meteorological records of nearly three years, environment construction of a simulation experiment is carried out, then the data are sent to a neural network for training and prediction, the recognition capability of the neural network on different wind speeds and wind directions is analyzed, the structure and the parameters of the neural network are continuously adjusted according to an experiment result, and further the feasibility of the method is improved. The invention adopts an Actor-critical frame-based deep reinforcement learning algorithm as a main structure, and optimizes parameters to improve the adaptability of the algorithm. Before unmanned aerial vehicle inspection starts every day, training the neural network according to weather prediction data of the day. And calculating the near-optimal unmanned aerial vehicle energy-saving routing inspection route based on a simulated annealing algorithm according to the training result. In the unmanned aerial vehicle inspection process, if the wind direction detected in real time and the wind speed are different from the weather prediction data, the path planning is carried out on the remaining fans again according to the new weather prediction data so as to ensure that the final unmanned aerial vehicle inspection route achieves the aim of low energy consumption.
Fig. 4 is a schematic diagram of the unmanned aerial vehicle path planning algorithm based on reinforcement learning.
Fig. 5 is a detailed design diagram of the unmanned aerial vehicle path planning algorithm based on reinforcement learning, and the flow specifically includes the following steps:
step 2.1: firstly, establishing a Markov decision process model related to the planning of tracks of two wind turbine generators by an unmanned aerial vehicle, and determining a quintuple in the process<S,A,P,R,γ>Wherein S represents the current environmental state quantity of the unmanned aerial vehicle, A is the action quantity executed by the unmanned aerial vehicle, P is the transition probability among different states, R is the reward quantity obtained when the unmanned aerial vehicle executes the action A in the state S, and gamma is the reinforcement learning attenuation rate. The state quantity S can completely represent the current state of the unmanned aerial vehicle, in the invention, a three-dimensional coordinate system is established according to the current position of the unmanned aerial vehicle, and the state quantity S comprises the position coordinate Pos of the current unmanned aerial vehicleU(x, y, z), and the wind velocity vector of the location where the drone is located at the current time
Simplifying the unmanned aerial vehicle motion model, wherein the action quantity A which can be executed by the unmanned aerial vehicle is equal to<a>Representing the unmanned plane velocity vectorMoving a distance in a fixed direction within a time slice τ. In order to enable the unmanned aerial vehicle to reach the target unit with the least power consumption, the following rewarding modes are designed:
wherein, | ds′L is the linear distance from the target point after the action is executed, | dsI is the linear distance from the target point before the action is performed, Ess′Energy consumption for this execution of action, EmaxThe maximum energy consumption of the unmanned aerial vehicle. When the unmanned aerial vehicle approaches the target with less energy consumption, more reward values can be obtained, and when the unmanned aerial vehicle reaches the target, a great reward value can be obtained, so that the target can absorb the unmanned aerial vehicle.
Energy consumption E required by unmanned aerial vehicle to execute action each timess′The calculation formula is Ess′=Pu·τ。PuFor unmanned aerial vehicle power, including horizontal flight powerVertical flight powerAnd resistance power Wherein W ═ mg is drone gravity; ρ is the air density;is the total area of the rotor of the unmanned aerial vehicle; cD0Is the drag coefficient related to rotor geometry;andthe relative speed of the unmanned aerial vehicle to the wind speed in the horizontal direction and the vertical direction respectively;the speed of horizontal flight of the unmanned aerial vehicle;hovering power for the drone.
Step 2.2: and initializing a playback experience pool D for storing data generated by the unmanned aerial vehicle in the trial and error process. Initializing an Actor real network mu and a Critic real network Q randomly, wherein the parameters corresponding to the two neural networks are theta respectivelyμAnd thetaQ(ii) a Randomly initializing an Actor target network mu 'and a criticic target network Q', wherein the parameters corresponding to the two neural networks are theta respectivelyμ′And thetaQ′And make thetaμ′=θμ,θQ′=θQ。
Step 2.3: recording the initial state quantity s1A random noise N conforming to a gaussian distribution is generated.
Step 2.4: the state quantity x at the current moment is calculatedtInput to at thetaμAdding random noise N at the current moment into the actual Actor network of the parametert. Outputting the action amount a by an Actor real networkt=μ(xt,θμ)+NtExecuting the action and calculating the reward r obtained by the action through a reward functiontWhile updating the state quantities to obtain xt+1。
Step 2.5: creating quadruplets<xt,at,rt,xt+1>It is stored in the playback experience pool D.
Step 2.6: randomly selecting a group of data from a playback experience pool D<xj,aj,rj,xj+1>X is to bej,ajInputting to Critic real network to obtain Q ═ Q (x)j,aj,θQ) X is to bej+1Inputting the calculated action amount a into the Actor target networkj+1=μ′(xj+1,θμ′) And x isj+1And aj+1Input into Critic target network together to obtain Q (x)j+1,aj+1,θQ) Then the target value Q' is rj+γQ′(xj+1,aj+1,θQ′) Training a Critic reality network with Q 'as a label to enable the calculated Q value to be infinitely close to a target value Q', and updating theta by using a gradient descent methodQ。
Step 2.7: and updating the actual network of the Actor to ensure that the Q value of the output action quantity is maximum in the computed Q value of the Critic actual network. Updating theta by gradient descent methodμThe strategy gradient calculation method is
Step 2.8: updating a target network parameter θμ′←αθμ′+(1-α)θμ,θQ′←αθQ′+(1-α)θQ。
Step 2.9: and (5) repeating the step 2.3 to the step 2.8 until the loss values of the Actor target network and the Critic target network are converged and the network parameters are unchanged. After the network converges, for any two wind generating sets xi、xj(i ≠ j), the deep reinforcement learning model can give the flight track with the minimum power consumption of the unmanned aerial vehicle and the power consumption Eij。
Fig. 6 is a flow chart of the unmanned aerial vehicle track planning algorithm based on simulated annealing, which specifically comprises the following steps:
step 3.1: and determining the start position cap of the unmanned aerial vehicle according to the current day rose diagram. If the current wind power plant has a dominant wind direction, setting the starting position of the unmanned aerial vehicle as a corner fan opposite to the dominant wind direction; if the current dominant wind direction is not obvious, then set unmanned aerial vehicle initial position as the fan of central point position.
Step 3.2: and training the neural network on the same day. Determining any two wind generating sets x in the maximum cruising radius of the unmanned aerial vehicle according to the playback experience pool in the step two and the wind direction and the wind speed predicted by the next-day weatheri、xjFlight path and flight energy consumption.
Step 3.3: starting from the initial position, sequentially selecting the lowest energy consumption track within the cruising radius of the unmanned aerial vehicle until all the fans are traversed, and taking the lowest energy consumption track as an initialization path c. Initializing the start temperature T and the end temperature simultaneouslyT0And an annealing speed alpha.
Step 3.4: by random thermal perturbation, another path c' is generated in the neighborhood of c. Unlike conventional simulated annealing algorithms, this disturbance can only occur between fans located in the same cruise area.
Step 3.5: and calculating the difference value delta E of the energy consumption of the unmanned aerial vehicle between the two paths c and c'. If Δ E is less than or equal to 0, updating the path, and making c ═ c', T ← α T; otherwise, a random number rand between 0-1 is generated, ifThen the path is updated, let c ═ c', T ← α T.
Step 3.6: judging T > T0If yes, continuing to execute the step 3.4; otherwise, obtaining a near-optimal solution of the routing inspection route planning with the lowest energy consumption, and the unmanned aerial vehicle sequentially overhauls and acquires data of the fan according to the routing inspection route.
Step 3.7: and judging whether the routing inspection route needs to be updated or not based on the real-time meteorological data. When the real-time wind direction detected by a wind sensor in the wind power plant and the wind direction predicted by the meteorology are not in the same direction or the real-time wind speed and the wind speed level predicted by the meteorology are different by more than two levels, counting the wind turbine set X' ═ X which is not traversed yet1,x2,...,xn′}. And 3.2, re-executing the step, and re-planning the lowest energy consumption flight path of the unmanned aerial vehicles of the rest wind turbine generators. And ending the algorithm until the unmanned aerial vehicle traverses the wind turbine generator.
Claims (5)
1. An unmanned aerial vehicle-cooperated wind power plant intelligent inspection method is characterized by comprising the following steps:
the method comprises the following steps: acquiring meteorological prediction data of a wind power plant in the future 4 hours, and preprocessing the data;
step two: for a plurality of wind turbine generators X ═ X under the same wind power plant1,x2,...,xnAnd (5) carrying out optimization on any two wind turbine generators x within the maximum cruising radius of the unmanned aerial vehicle by utilizing a depth reinforcement learning algorithmi、xjPlanning the lowest power consumption track and calculating the corresponding power consumption EijWherein i ≠ j;
step three: determining the starting position of the unmanned aerial vehicle according to the predicted weather main wind direction, and performing flight path planning based on a simulated annealing algorithm on the unmanned aerial vehicle according to the current weather data and the learning experience in the step two;
step four: and D, according to the planned flight path in the step three, each unit of the wind power plant is inspected so as to facilitate troubleshooting and data uploading.
2. The unmanned aerial vehicle-coordinated wind farm intelligent inspection method according to claim 1, wherein in the first step, the specific steps of data preprocessing are as follows:
step 1.1: checking meteorological data, and if a missing part exists, smoothing the missing data;
step 1.2: for any time t, according to wind direction data thetatWill wind speedOrthogonal decomposition is carried out to decompose the wind speed into the wind speeds in three mutually perpendicular directions in the three-dimensional space
Step 1.3: and carrying out normalization processing on the wind direction and the wind speed data.
3. The intelligent inspection method for the wind farm with cooperation of the unmanned aerial vehicle according to claim 1 or 2, wherein in the second step, the deep reinforcement learning algorithm is constructed by the following steps:
step 2.1: firstly, establishing a Markov decision process model related to the planning of tracks of two wind turbine generators by an unmanned aerial vehicle, and determining a quintuple < S, A, P, R, gamma > in the process, wherein S represents the state quantity of the environment where the unmanned aerial vehicle is located currently, A is the action quantity executed by the unmanned aerial vehicle, P is the transition probability among different states, R is the reward quantity obtained when the unmanned aerial vehicle executes the action A in the state S, and gamma is the reinforcement learning attenuation rate;
establishing a three-dimensional coordinate system by using the current position of the unmanned aerial vehicle, wherein the state quantity S comprises the position coordinate Pos of the current unmanned aerial vehicleU(x, y, z) and the wind velocity vector of the current position of the drone
Simplifying the unmanned aerial vehicle motion model, wherein the action quantity A which can be executed by the unmanned aerial vehicle is equal to<a>Representing the unmanned plane velocity vectorMoving a distance in a fixed direction within a time slice τ; in order to enable the unmanned aerial vehicle to reach the target unit with the least power consumption, the following rewarding modes are designed:
wherein, | ds′I is the linear distance between the executed action and the target point; | dsI is the linear distance between the target point and the target point before executing the action; emaxThe maximum energy consumption of the unmanned aerial vehicle is achieved; ess′For the energy consumption of the execution action, the calculation formula is Ess′=Pu·τ,PuFor unmanned aerial vehicle power, including horizontal flight powerVertical flight powerAnd resistance power Wherein W ═ mg is drone gravity; ρ is the air density;is the total area of the rotor of the unmanned aerial vehicle; cD0Is the drag coefficient related to rotor geometry;andthe relative speed of the unmanned aerial vehicle to the wind speed in the horizontal direction and the vertical direction respectively;the speed of horizontal flight of the unmanned aerial vehicle;hovering power for the unmanned aerial vehicle;
step 2.2: initializing a playback experience pool D, and storing data generated by the unmanned aerial vehicle in a trial and error process; initializing an Actor real network mu and a Critic real network Q randomly, wherein the parameters corresponding to the two neural networks are theta respectivelyμAnd thetaQ(ii) a Randomly initializing an Actor target network mu 'and a criticic target network Q', wherein the parameters corresponding to the two neural networks are theta respectivelyμ′And thetaQ′And make thetaμ′=θμ,θQ′=θQ;
Step 2.3: recording the initial state quantity s1Generating a random noise N which accords with Gaussian distribution;
step 2.4: the state quantity x at the current moment is calculatedtInput to at thetaμAdding random noise N at the current moment into the actual Actor network of the parametert(ii) a Outputting the action quantity a by the Actor real networkt=μ(xt,θμ)+NtExecute the action, go throughThe reward r obtained by the action is calculated by the reward functiontWhile updating the state quantities to obtain xt+1;
Step 2.5: creating quadruplets<xt,at,rt,xt+1>Storing the data into a playback experience pool D;
step 2.6: randomly selecting a group of data from a playback experience pool D<xj,aj,rj,xj+1>X is to bej,ajInputting into Critic real network to obtain Q = Q (x)j,aj,θQ) X is to bej+1Inputting the calculated action amount a into the Actor target networkj+1=μ′(xj+1,θμ′) And x isj+1And aj+1Input into Critic target network together to obtain Q (x)j+1,aj+1,θQ) Then the target value Q' is rj+γQ′(xj+1,aj+1,θQ′) Training the Critic reality network to make the calculated Q value infinitely close to the target value Q 'by using Q' as a label, and updating theta by using a gradient descent methodQ;
Step 2.7: updating the actual network of the Actor to enable the Q value of the output action quantity to be maximum in the computed Q value of the actual network of the Critic; updating theta by gradient descent methodμThe strategy gradient calculation method is
Step 2.8: updating a target network parameter θμ′←αθμ′+(1-α)θμ,θQ′←αθQ′+(1-α)θQ;
Step 2.9: repeating the step 2.3 to the step 2.8 until the loss values of the Actor target network and the Critic target network are converged and the network parameters are unchanged; after the network converges, for any two wind generating sets xi、xjThe deep reinforcement learning model can provide the flight track with minimum power consumption of the unmanned aerial vehicle and the power consumption Eij。
4. The unmanned aerial vehicle-coordinated wind farm intelligent inspection method according to claim 1 or 2, wherein in the third step, the specific steps of the simulated annealing algorithm are as follows:
step 3.1: determining the starting position of the unmanned aerial vehicle according to the current day rose diagram; if the current wind power plant has a dominant wind direction, setting the starting position of the unmanned aerial vehicle as a corner fan opposite to the dominant wind direction; if the current dominant wind direction is not obvious, setting the starting position of the unmanned aerial vehicle as a fan at the central position;
step 3.2: training two neural networks on the same day; determining any two wind generating sets x in the maximum cruising radius of the unmanned aerial vehicle according to the playback experience pool in the step two and the wind direction and the wind speed predicted by the next-day weatheri、xjFlight path and flight energy consumption between;
step 3.3: starting from the initial position, sequentially selecting the lowest energy consumption track within the cruising radius of the unmanned aerial vehicle until all fans are traversed, and taking the lowest energy consumption track as an initialization path c; simultaneously initializing a start temperature T and an end temperature T0And an annealing speed α;
step 3.4: generating another path c' in the neighborhood of c by random thermal perturbation;
step 3.5: calculating the difference value delta E of the energy consumption of the unmanned aerial vehicle between the two paths c and c'; if Δ E is less than or equal to 0, updating the path, and making c ═ c', T ← α T; otherwise, a random number rand between 0-1 is generated, ifUpdating the path, and making c ═ c', T ← α T;
step 3.6: judging T > T0If yes, continuing to execute the step 3.4; otherwise, obtaining a near-optimal solution of the routing inspection route with the lowest energy consumption, and sequentially overhauling and acquiring data of the fan by the unmanned aerial vehicle according to the routing inspection route;
step 3.7: judging whether the routing inspection route needs to be updated or not based on the real-time meteorological data; when the real-time wind direction detected by the wind sensor in the wind power plant is not in the same direction as the wind direction predicted by the weather, or the real-time wind speed and the weather are predictedWhen the wind speed level of the wind turbine is different by more than two levels, counting the wind turbine generator X' which is not traversed yet1,x2,...,xn′Executing the step 3.2 again, and replanning the lowest energy consumption flight path of the unmanned aerial vehicles of the rest wind turbine generators; and ending the algorithm until the unmanned aerial vehicle traverses the wind turbine generator.
5. The unmanned aerial vehicle-coordinated wind farm intelligent inspection method according to claim 3, wherein in the third step, the specific steps of the simulated annealing algorithm are as follows:
step 3.1: determining the starting position of the unmanned aerial vehicle according to the current day rose diagram; if the current wind power plant has a dominant wind direction, setting the starting position of the unmanned aerial vehicle as a corner fan opposite to the dominant wind direction; if the current dominant wind direction is not obvious, setting the starting position of the unmanned aerial vehicle as a fan at the central position;
step 3.2: training two neural networks on the same day; determining any two wind generating sets x in the maximum cruising radius of the unmanned aerial vehicle according to the playback experience pool in the step two and the wind direction and the wind speed predicted by the next-day weatheri、xjFlight path and flight energy consumption between;
step 3.3: starting from the initial position, sequentially selecting the lowest energy consumption track within the cruise radius of the unmanned aerial vehicle until all fans are traversed, and taking the track as an initialization path c; simultaneously initializing a start temperature T and an end temperature T0And an annealing speed α;
step 3.4: generating another path c' in the neighborhood of c by random thermal perturbation;
step 3.5: calculating the difference value delta E of the energy consumption of the unmanned aerial vehicle between the two paths c and c'; if Δ E is less than or equal to 0, updating the path, and making c ═ c', T ← α T; otherwise, a random number rand between 0-1 is generated, ifUpdating the path, and making c ═ c', T ← α T;
step 3.6: judging T > T0If yes, continuing to execute the step 3.4; otherwise, the patrol with the lowest energy consumption is obtainedThe method comprises the following steps that (1) a near-optimal solution of routing planning is checked, and the unmanned aerial vehicle carries out maintenance and data acquisition on a fan in sequence according to the routing;
step 3.7: judging whether the routing inspection route needs to be updated or not based on the real-time meteorological data; when the real-time wind direction detected by a wind sensor in the wind power plant is not in the same direction with the meteorological predicted wind direction or the real-time wind speed is different from the meteorological predicted wind speed by more than two levels, counting the wind turbine set X' ═ X not traversed yet1,x2,...,xn′Executing the step 3.2 again, and replanning the lowest energy consumption flight path of the unmanned aerial vehicles of the rest wind turbine generators; and ending the algorithm until the unmanned aerial vehicle traverses the wind turbine generator.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210274635.5A CN114610070B (en) | 2022-03-21 | 2022-03-21 | Unmanned aerial vehicle-coordinated intelligent inspection method for wind farm |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210274635.5A CN114610070B (en) | 2022-03-21 | 2022-03-21 | Unmanned aerial vehicle-coordinated intelligent inspection method for wind farm |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114610070A true CN114610070A (en) | 2022-06-10 |
CN114610070B CN114610070B (en) | 2024-06-21 |
Family
ID=81864395
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210274635.5A Active CN114610070B (en) | 2022-03-21 | 2022-03-21 | Unmanned aerial vehicle-coordinated intelligent inspection method for wind farm |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114610070B (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115933787A (en) * | 2023-03-14 | 2023-04-07 | 西安英图克环境科技有限公司 | Indoor multi-terminal intelligent control system based on indoor environment monitoring |
CN116878518A (en) * | 2023-09-06 | 2023-10-13 | 滨州市华亿电器设备有限公司 | Unmanned aerial vehicle inspection path planning method for urban power transmission line maintenance |
CN118819175B (en) * | 2024-09-20 | 2024-11-19 | 云南视限科技有限公司 | Unmanned aerial vehicle steering engine control method, device, equipment and storage medium |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111256703A (en) * | 2020-05-07 | 2020-06-09 | 江苏方天电力技术有限公司 | Multi-rotor unmanned aerial vehicle inspection path planning method |
CN112286203A (en) * | 2020-11-11 | 2021-01-29 | 大连理工大学 | Multi-agent reinforcement learning path planning method based on ant colony algorithm |
JP2021034050A (en) * | 2019-08-21 | 2021-03-01 | 哈爾浜工程大学 | Auv action plan and operation control method based on reinforcement learning |
CN113485453A (en) * | 2021-08-20 | 2021-10-08 | 中国华能集团清洁能源技术研究院有限公司 | Method and device for generating inspection flight path of offshore unmanned aerial vehicle and unmanned aerial vehicle |
-
2022
- 2022-03-21 CN CN202210274635.5A patent/CN114610070B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2021034050A (en) * | 2019-08-21 | 2021-03-01 | 哈爾浜工程大学 | Auv action plan and operation control method based on reinforcement learning |
CN111256703A (en) * | 2020-05-07 | 2020-06-09 | 江苏方天电力技术有限公司 | Multi-rotor unmanned aerial vehicle inspection path planning method |
CN112286203A (en) * | 2020-11-11 | 2021-01-29 | 大连理工大学 | Multi-agent reinforcement learning path planning method based on ant colony algorithm |
CN113485453A (en) * | 2021-08-20 | 2021-10-08 | 中国华能集团清洁能源技术研究院有限公司 | Method and device for generating inspection flight path of offshore unmanned aerial vehicle and unmanned aerial vehicle |
Non-Patent Citations (1)
Title |
---|
汤翔;杨源: "海上风电场智能无人巡检方案研究", 电工技术, no. 020, 31 December 2021 (2021-12-31) * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115933787A (en) * | 2023-03-14 | 2023-04-07 | 西安英图克环境科技有限公司 | Indoor multi-terminal intelligent control system based on indoor environment monitoring |
CN115933787B (en) * | 2023-03-14 | 2023-05-16 | 西安英图克环境科技有限公司 | Indoor multi-terminal intelligent control system based on indoor environment monitoring |
CN116878518A (en) * | 2023-09-06 | 2023-10-13 | 滨州市华亿电器设备有限公司 | Unmanned aerial vehicle inspection path planning method for urban power transmission line maintenance |
CN116878518B (en) * | 2023-09-06 | 2023-11-21 | 滨州市华亿电器设备有限公司 | Unmanned aerial vehicle inspection path planning method for urban power transmission line maintenance |
CN118819175B (en) * | 2024-09-20 | 2024-11-19 | 云南视限科技有限公司 | Unmanned aerial vehicle steering engine control method, device, equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN114610070B (en) | 2024-06-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Chung et al. | Placement and routing optimization for automated inspection with unmanned aerial vehicles: A study in offshore wind farm | |
CN110110943B (en) | Fleet energy efficiency comprehensive intelligent optimization management system and optimization method based on big data | |
CN108764560A (en) | Aircraft scene trajectory predictions method based on shot and long term Memory Neural Networks | |
KR102089564B1 (en) | Tethered gyroglider control systems | |
CN114610070B (en) | Unmanned aerial vehicle-coordinated intelligent inspection method for wind farm | |
CN112734970B (en) | Automatic inspection system and method for unmanned aerial vehicle in wind farm based on LoRaWAN positioning technology | |
Liu et al. | Study on UAV parallel planning system for transmission line project acceptance under the background of industry 5.0 | |
CN113705982A (en) | Scheduling decision method for vehicle-mounted machine cooperative power patrol | |
CN107304746A (en) | Wind generating set and operation control method and device thereof | |
CN113342041A (en) | Method and system for realizing automatic inspection of fan blade power generation equipment by unmanned aerial vehicle | |
CN114153224A (en) | Micro-meteorology-based unmanned aerial vehicle flight path intelligent planning control system and method | |
CN117575115A (en) | Intelligent inspection optimization method and system for power transmission line | |
CN115326075A (en) | Path planning method for realizing wind field global automatic inspection based on unmanned aerial vehicle | |
CN111488974A (en) | Deep learning neural network-based ocean wind energy downscaling method | |
CN114548228A (en) | Unmanned aerial vehicle power grid inspection method and system based on SOC and MESH networking | |
CN118210328B (en) | Unmanned aerial vehicle autonomous navigation system based on model light weight technology | |
Huang et al. | Study on a boat-assisted drone inspection scheme for the modern large-scale offshore wind farm | |
Liu et al. | Mission capability estimation of multicopter uav for low-altitude remote sensing | |
CN115574826B (en) | National park unmanned aerial vehicle patrol path optimization method based on reinforcement learning | |
CN117455030A (en) | Unmanned aerial vehicle nest bionic planning method, device and storage medium | |
CN113110558B (en) | Hybrid propulsion unmanned aerial vehicle demand power prediction method | |
CN117170408A (en) | Photovoltaic panel site inspection path intelligent planning system and method based on unmanned aerial vehicle | |
Changxin et al. | UAV electric patrol path planning based on improved ant colony optimization-A* algorithm | |
CN116400728A (en) | Unmanned plane path planning method based on depth deterministic strategy gradient algorithm | |
Pan et al. | Track Planning for Damage Detection on the Building Enclosure by UAV Considering Image Detection. |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |