CN104616033A - Fault diagnosis method for rolling bearing based on deep learning and SVM (Support Vector Machine) - Google Patents

Fault diagnosis method for rolling bearing based on deep learning and SVM (Support Vector Machine) Download PDF

Info

Publication number
CN104616033A
CN104616033A CN201510080839.5A CN201510080839A CN104616033A CN 104616033 A CN104616033 A CN 104616033A CN 201510080839 A CN201510080839 A CN 201510080839A CN 104616033 A CN104616033 A CN 104616033A
Authority
CN
China
Prior art keywords
mrow
msub
layer
training
msubsup
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510080839.5A
Other languages
Chinese (zh)
Inventor
刘嘉敏
刘军委
刘亦哲
罗甫林
彭玲
黄鸿
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University
Original Assignee
Chongqing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University filed Critical Chongqing University
Priority to CN201510080839.5A priority Critical patent/CN104616033A/en
Publication of CN104616033A publication Critical patent/CN104616033A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention provides a fault diagnosis method for a rolling bearing based on a deep learning and SVM (Support Vector Machine). The method comprises using a manure learning algorithm in a deep belief network theory to complete a characteristic extraction task needed by fault diagnosis; automatically extracting the substantive characteristics of data input independent of manual selection from simple to complicate, from low to high, and automatically digging abundant information concealed in known data; in addition, classifying and identifying a test sample by adopting an SVM classification method, seeking and finding a global minimum of a target function through an effective method previously designed, so as to solve the problem that a deep belief network may be trapped into a locally optimal solution. According to the fault diagnosis method for the rolling bearing based on the deep learning and SVM provided by the invention, the accuracy and effectiveness of the fault diagnosis method for a rolling bearing can be improved, and a new effective way can be provided to solve the accuracy and effectiveness of the fault diagnosis method, therefore the fault diagnosis method can be extensively applied complex systems in chemistry, metallurgy, electric power, aviation fields and the like.

Description

Rolling bearing fault diagnosis method based on deep learning and support vector machine
Technical Field
The invention belongs to the technical field of mechanical fault diagnosis and computer artificial intelligence, and particularly relates to a rolling bearing fault diagnosis method based on deep learning and a support vector machine.
Background
The rolling bearing is one of the most important mechanical parts in rotary machinery, is widely applied to various important departments of chemical industry, metallurgy, electric power, aviation and the like, and is one of the most vulnerable elements. The performance of the bearing and the working condition directly affect the performance of the shaft associated with the bearing and the gear arranged on the rotating shaft and even the performance of the whole machine equipment, and the defects can cause abnormal vibration and noise of the equipment and even cause the damage of the equipment. Therefore, diagnosing the rolling bearing fault, especially analyzing the early fault, and avoiding the accident is particularly important in the practical production.
Currently, the commonly used mechanical fault feature extraction methods mainly include Fast Fourier Transform (FFT), wavelet transform and Empirical Mode Decomposition (EMD), artificial intelligence, and the like. The FFT method cannot simultaneously take into account the global and localized problems of the signal in the time domain and the frequency domain. During wavelet transformation, wavelet bases are different, decomposition results are different, and the wavelet bases are difficult to select. The EMD method can decompose a signal into a plurality of IMF (intrinsic mode functions) components, and Hilbert transform is performed on all the IMF components to obtain time-frequency distribution of the signal, but theoretically, problems such as mode confusion, under-envelope, over-envelope, endpoint effect and the like in the EMD method are still in research. In the fault diagnosis method based on artificial intelligence, currently, an artificial neural network is mainly used for finishing classification of diagnosis targets through continuous learning and information feedback on a system; however, the method has the disadvantages that the reasoning process is poor in interpretability, and when a sample to be diagnosed is incomplete (data is missing), the neural network cannot carry out effective reasoning work, and the bearing cannot be correspondingly diagnosed by utilizing early characteristics of faults.
Disclosure of Invention
Aiming at the problems in the prior art, the invention provides a rolling bearing fault diagnosis method based on deep learning and a support vector machine.
In order to achieve the purpose, the invention adopts the following technical means:
the rolling bearing fault diagnosis method based on deep learning and a support vector machine comprises the following steps:
1) when the rolling bearings under four different working conditions work in a rotating mode, vibration acceleration signals of the rolling bearings under each working condition working at different rotating speeds are collected through the acceleration sensors respectively, denoising pretreatment is carried out, working condition labels are added, and vibration acceleration signal data under various working conditions after pretreatment and working condition labels are added are used as training samples; the four working conditions are normal operation, bearing inner ring fault operation, bearing rolling element fault operation and bearing outer ring fault operation respectively;
2) establishing a deep belief network model, training the deep belief network model by adopting a training sample, inputting the training sample into the deep belief network model, and performing layer-by-layer training and optimization by adopting an unsupervised greedy layer-by-layer training method to obtain a connection weight and an offset parameter of the deep belief network model;
3) respectively taking training samples under various working conditions as input of a deep belief network model for determining a connection weight and a bias parameter, carrying out deep learning on the training samples, and respectively reconstructing each training sample under each working condition by adopting the deep belief network model for determining the connection weight and the bias parameter to obtain a training sample reconstruction signal corresponding to each training sample under each working condition;
4) acquiring vibration acceleration signal data of a rolling bearing to be tested during rotation work through an acceleration sensor, and performing denoising pretreatment to obtain a test sample;
5) taking the test sample as the input of a depth belief network model for determining a connection weight and a bias parameter, carrying out depth learning on the test sample, and reconstructing the test sample by adopting the depth belief network model for determining the connection weight and the bias parameter to obtain a test sample reconstruction signal;
6) and taking the test sample reconstruction signal as the matching characteristic of the test sample, taking the training sample reconstruction signal corresponding to each training sample under each working condition as the matching reference, matching the test sample with the training sample by adopting a support vector machine classification method, and judging the working condition class to which the training sample most matched with the test sample belongs as the working condition class of the test sample, thereby obtaining the fault diagnosis result of the rolling bearing to be tested.
In the rolling bearing fault diagnosis method based on deep learning and support vector machine, specifically, the joint distribution function of the deep belief network model established in the step 2) is:
<math> <mrow> <mi>E</mi> <mrow> <mo>(</mo> <mi>v</mi> <mo>,</mo> <mi>h</mi> <mo>|</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> <mo>=</mo> <mo>-</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>I</mi> </munderover> <msub> <mi>a</mi> <mi>i</mi> </msub> <msub> <mi>v</mi> <mi>i</mi> </msub> <mo>-</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>j</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>J</mi> </munderover> <msub> <mi>b</mi> <mi>j</mi> </msub> <msub> <mi>h</mi> <mi>j</mi> </msub> <mo>-</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>I</mi> </munderover> <munderover> <mi>&Sigma;</mi> <mrow> <mi>j</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>J</mi> </munderover> <msub> <mi>w</mi> <mi>ij</mi> </msub> <msub> <mi>v</mi> <mi>i</mi> </msub> <msub> <mi>h</mi> <mi>j</mi> </msub> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein θ ═ wij,ai,bj) For deep belief network model parameters, wijVisible layer ith node v representing deep belief networkiWith the jth node h of the hidden layerjThe connection weight value between aiAnd bjRespectively represent ith nodes v of visible layeriBias parameters and the jth node h of the hidden layerjThe bias parameter of (2);
an unsupervised greedy layer-by-layer training method is adopted to carry out layer-by-layer training and optimization on the deep belief network model, and the specific mode is as follows:
21) training the limited Boltzmann machines in each layer of the deep belief network model in a layer-by-layer training mode, wherein the hidden layer output of the limited Boltzmann machine at the lower layer is used as the visible layer input of the limited Boltzmann machine at the upper layer until the output of the last hidden layer of the deep belief network model is obtained; the method specifically comprises the following steps:
the joint probability distribution probability p (v, h | θ) of the visible layer node and the hidden layer node is:
p(v,h|θ)=e-E(v,h|θ)/Z(θ); (2)
wherein, <math> <mrow> <mi>Z</mi> <mrow> <mo>(</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mi>&Sigma;</mi> <mi>v</mi> </munder> <munder> <mi>&Sigma;</mi> <mi>h</mi> </munder> <msup> <mi>e</mi> <mrow> <mo>-</mo> <mi>E</mi> <mrow> <mo>(</mo> <mi>v</mi> <mo>,</mo> <mi>h</mi> <mo>|</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> </mrow> </msup> </mrow> </math> is a distribution function;
when the state of the visible layer node is given, the jth node h of the hidden layerjThe activation probability of (c) is:
<math> <mrow> <mi>p</mi> <mrow> <mo>(</mo> <msub> <mi>h</mi> <mi>j</mi> </msub> <mo>=</mo> <mn>1</mn> <mo>|</mo> <mi>v</mi> <mo>,</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> <mo>=</mo> <mi>&sigma;</mi> <mrow> <mo>(</mo> <msub> <mi>b</mi> <mi>j</mi> </msub> <mo>+</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>I</mi> </munderover> <msub> <mi>v</mi> <mi>i</mi> </msub> <msub> <mi>w</mi> <mi>ji</mi> </msub> <mo>)</mo> </mrow> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>3</mn> <mo>)</mo> </mrow> </mrow> </math>
given the state of the hidden layer node, the ith node v of the visible layeriThe activation probability of (c) is:
<math> <mrow> <mi>p</mi> <mrow> <mo>(</mo> <msub> <mi>v</mi> <mi>i</mi> </msub> <mo>=</mo> <mn>1</mn> <mo>|</mo> <mi>h</mi> <mo>,</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> <mo>=</mo> <mi>&sigma;</mi> <mrow> <mo>(</mo> <msub> <mi>a</mi> <mi>j</mi> </msub> <mo>+</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>I</mi> </munderover> <msub> <mi>h</mi> <mi>i</mi> </msub> <msub> <mi>w</mi> <mi>ji</mi> </msub> <mo>)</mo> </mrow> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>4</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein σ (x) is 1/(1+ e)-x) Is sigmoid function;
according to the activation probability, when a given training sample is input to the visible layer node, after all nodes of the hidden layer are excited by adopting a joint distribution function of a deep belief network model, the nodes of the next hidden layer are excited, so that the visible layer node is obtained again; then, calculating the condition distribution of the visible layer data by adopting a contrast divergence algorithm to obtain hidden layer data, calculating the visible layer data by calculating the condition distribution of the hidden layer data, reconstructing the visible layer data, adjusting and updating the parameters of the deep belief network model by utilizing a gradient descent method, wherein the updating difference values of the connection weight and the offset parameter are respectively as follows:
Δwij=(<vihj>data-<vihj>recon); (6)
Δai=(<vi>data-<vi>recon); (7)
Δbj=(<hj>data-<hj>recon); (8)
wherein, Δ wijVisible layer ith node v representing deep belief networkiWith the jth node h of the hidden layerjThe connection weight value w betweenijUpdate difference, Δ a, for performing an updateiAnd Δ bjRespectively represent ith nodes v of visible layeriBias parameter a ofiUpdating difference value and jth node h of hidden layer for updatingjBias parameter b ofjUpdating the updated difference value; in order to train the learning rate of the training,<·>datarepresenting a mathematical expectation over a distribution defined by the training data set,<·>reconrepresenting a mathematical expectation on the distribution of the reconstructed deep belief network model output;
training layer by layer until the output of the last hidden layer of the deep belief network model is obtained;
22) training a back propagation network for the output of the last hidden layer of the deep belief network model obtained in the step 21, propagating the classification error of the classification result output by training prediction and the actual classification result of the training sample backward layer by layer, and adjusting the connection weight of each layer of the deep belief network model; the method specifically comprises the following steps:
the training of the back propagation network is divided into a forward propagation process and a back propagation process; in the forward propagation process, the output of the last hidden layer of the deep belief network model obtained in the step 21 is propagated to an output layer by layer as input to obtain a predicted classification category, the actual classification result of the training sample is determined according to the working condition label of the training sample, then the predicted classification result is compared with the actual classification result of the training sample to obtain a classification error, and the classification error is transmitted back layer by layer, so that the parameters of the deep training network are adjusted; in the backward propagation process, the value of the sensitivity of each layer needs to be calculated, and the sensitivity is transmitted from top to bottom to correct the connection weight of the deep belief network model;
for the output layer, assume the actual output of the ith node is oiThe desired output is diThen, the calculation expression of the sensitivity of the ith node is:
i=oi(1-oi)(di-oi); (9)
for the mth hidden layer, the computational expression of the sensitivity of the ith node is:
<math> <mrow> <msubsup> <mi>&delta;</mi> <mi>i</mi> <mi>m</mi> </msubsup> <mo>=</mo> <msubsup> <mi>y</mi> <mi>i</mi> <mi>m</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>y</mi> <mi>i</mi> <mi>m</mi> </msubsup> <mo>)</mo> </mrow> <munder> <mi>&Sigma;</mi> <mi>j</mi> </munder> <msubsup> <mi>w</mi> <mi>ij</mi> <mi>m</mi> </msubsup> <msubsup> <mi>&delta;</mi> <mi>j</mi> <mrow> <mi>m</mi> <mo>+</mo> <mn>1</mn> </mrow> </msubsup> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>10</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein,indicating the sensitivity of the ith node of the mth hidden layer,indicating the sensitivity of the ith node of the (m + 1) th hidden layer,represents the output of the ith node of the mth hidden layer,representing the connection weight between the ith node of the mth hidden layer and the jth node of the (m + 1) th hidden layer;
after the sensitivity of each node in each hidden layer is obtained, the connection weight of the deep belief network model is updated and adjusted according to the following formula:
<math> <mrow> <msubsup> <mi>w</mi> <mi>ij</mi> <mi>m</mi> </msubsup> <mo>&LeftArrow;</mo> <msubsup> <mi>w</mi> <mi>ij</mi> <mi>m</mi> </msubsup> <mo>+</mo> <msub> <mi>&epsiv;</mi> <mrow> <mi>fine</mi> <mo>-</mo> <mi>tuning</mi> </mrow> </msub> <mo>&times;</mo> <msubsup> <mi>y</mi> <mi>i</mi> <mi>m</mi> </msubsup> <msubsup> <mi>&delta;</mi> <mi>j</mi> <mrow> <mi>m</mi> <mo>+</mo> <mn>1</mn> </mrow> </msubsup> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>11</mn> <mo>)</mo> </mrow> </mrow> </math>
<math> <mrow> <msubsup> <mi>b</mi> <mi>j</mi> <mi>m</mi> </msubsup> <mo>&LeftArrow;</mo> <msubsup> <mi>b</mi> <mi>j</mi> <mi>m</mi> </msubsup> <mo>+</mo> <msub> <mi>&epsiv;</mi> <mrow> <mi>fine</mi> <mo>-</mo> <mi>tuning</mi> </mrow> </msub> <mo>&times;</mo> <msubsup> <mi>&delta;</mi> <mi>j</mi> <mrow> <mi>m</mi> <mo>+</mo> <mn>1</mn> </mrow> </msubsup> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>12</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein,fine-tuningthe learning rate is shown to be adjusted and optimized,representing the bias parameter of the jth node of the mth hidden layer;
and finally determining the connection weight and the bias parameter of the whole depth belief network model after adjusting and optimizing the connection weight of each layer of the depth belief network model.
In the rolling bearing fault diagnosis method based on deep learning and support vector machine, as an optimal solution, the specific way of matching the test sample and the training sample in step 6) by using the support vector machine classification method is as follows:
61) in the training samples of four working conditions, aiming at the situation that the kth class training sample is regarded as a positive class, k belongs to {1,2,3,4}, other 3 classes of training samples are regarded as negative classes, and a classification decision function f of the kth class is obtained by a classification method of two classes of support vector machinesk(x):
<math> <mrow> <msub> <mi>f</mi> <mi>k</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>=</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>n</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>N</mi> </munderover> <msubsup> <mi>&alpha;</mi> <mi>n</mi> <mi>k</mi> </msubsup> <msub> <mi>y</mi> <mi>n</mi> </msub> <mi>K</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <msub> <mi>x</mi> <mi>n</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msub> <mi>b</mi> <mi>k</mi> </msub> <mo>;</mo> </mrow> </math>
Wherein,classifying a decision function f for class kk(x) N-th training sample reconstructed signal xnA corresponding Lagrange coefficient; bkClassifying a decision function f for class kk(x) The optimal hyperplane position coefficient; y isnRepresenting the classification mark corresponding to the nth training sample, y when the nth training sample belongs to the positive classn1, y when the nth training sample belongs to the negative classn-1; n belongs to {1,2, …, N }, wherein N is the total number of training samples of four working conditions; k (x, x)n) Representing a classification decision function fk(x) Is compared with the n-th training sample to reconstruct the signal xnGaussian radial basis kernel function of (1);
thus obtaining a classification decision function corresponding to each working condition in the four working conditions;
62) and respectively taking the test sample reconstruction signals as input quantities of the classification decision functions corresponding to the four working conditions, calculating the four classification decision function values of the test sample reconstruction signals as the input quantities, and judging the working condition type corresponding to the largest classification decision function value as the working condition type of the test sample to obtain a fault diagnosis result of the rolling bearing to be tested.
Compared with the prior art, the invention has the following beneficial effects:
1. the rolling bearing fault diagnosis method based on deep learning and the support vector machine completes the feature extraction task required by fault diagnosis by utilizing a mature learning algorithm in a deep belief network theory, can automatically extract the essential features of input data from low level to high level without depending on manual selection from simple to complex, can save a large amount of manpower, can automatically dig out rich information hidden in known data, and is particularly suitable for noisy, uncertain and dynamic systems.
2. In the rolling bearing fault diagnosis method based on deep learning and the support vector machine, the support vector machine classification method is adopted to classify and identify the test samples, and the learning process in the support vector machine classification method can be regarded as a process for optimally searching an optimal solution, so that a previously designed effective method can be adopted to search and find a global minimum value of a target function, and the problem that a deep belief network possibly falls into a local optimal solution is solved.
3. Compared with the prior art, the rolling bearing fault diagnosis method can improve the accuracy and effectiveness of rolling bearing fault diagnosis, provides a new effective way for solving the problem of rolling bearing fault diagnosis, and can be widely applied to complex systems in the fields of chemical industry, metallurgy, electric power, aviation and the like.
Drawings
Fig. 1 is a flow chart of a rolling bearing fault diagnosis method based on deep learning and a support vector machine according to the invention.
Fig. 2 is an example graph of time domain distribution (time domain unit is ms) of an original vibration acceleration signal of a rolling bearing inner ring fault operation.
Fig. 3 is an example graph of the time domain distribution of the original vibration acceleration signal (time domain unit is ms) of the rolling element fault operation of the rolling bearing.
FIG. 4 is a model architecture diagram of the deep belief network model.
FIG. 5 is a schematic model diagram of a restricted Boltzmann machine.
FIG. 6 is a schematic diagram of the position relationship of the SVM normalized optimal classification hyperplane.
Detailed Description
In order to overcome the defects of the prior art, the rolling bearing fault diagnosis method based on deep learning and the support vector machine firstly adopts a deep belief network to learn the essential characteristics of training sample data, and then adopts the support vector machine classification method to classify and identify the test sample, so that the fault working condition category of the rolling bearing is determined, and the accuracy and the effectiveness of the fault diagnosis of the rolling bearing are improved.
The Deep Belief Network (DBN) has strong function expression capability and shows the excellent characteristic of learning essential characteristics of data from a few samples. Research shows that a deep network structure composed of multiple nonlinear mapping layers is more effective than a shallow structure, and has good effect and efficiency on complex function representation and complex classification. The deep belief network explains that the multi-hidden-layer neural network has excellent feature learning capability, and the learned features are more essential to data, so that the classification and visualization are facilitated.
The core idea of a Support Vector Machine (SVM) classifier is to map an input Vector to a high-dimensional feature space through some nonlinear mapping (kernel function) and construct an optimal classification hyperplane, thereby realizing classification and identification. The method has unique advantages in solving small sample, nonlinear and high-dimensional pattern recognition, can well limit over-learning, is particularly suitable for data processing of small sample sets, and can be applied to fault diagnosis and fault prediction.
Based on the above advantages of the deep belief network and the support vector machine, the rolling bearing fault diagnosis method of the invention integrates the above advantages of the deep belief network and the support vector machine, and utilizes the deep learning and the support vector machine to classify the fault conditions of the rolling bearing, so as to realize the identification and diagnosis of the rolling bearing fault, the specific operation flow is shown in fig. 1, and the method comprises the following steps:
1) when the rolling bearings under four different working conditions work in a rotating mode, vibration acceleration signals of the rolling bearings under each working condition working at different rotating speeds are collected through the acceleration sensors respectively, denoising pretreatment is carried out, working condition labels are added, and vibration acceleration signal data under various working conditions after pretreatment and working condition labels are added are used as training samples; the four working conditions are normal operation, bearing inner ring fault operation, bearing rolling element fault operation and bearing outer ring fault operation respectively.
For example, fig. 2 and fig. 3 respectively show time domain distribution diagrams (time domain unit is ms) of original vibration acceleration signals of the rolling bearing under the working conditions of inner ring fault operation and rolling body fault operation, and the signal difference is obvious. Therefore, the fault condition of the rolling bearing can be identified based on the vibration acceleration signal data of the rolling bearing under different working conditions.
2) Establishing a deep belief network model, training the deep belief network model by adopting a training sample, inputting the training sample into the deep belief network model, and performing layer-by-layer training and optimization by adopting an unsupervised greedy layer-by-layer training method to obtain a connection weight and an offset parameter of the deep belief network model.
A model architecture diagram of the deep belief network model is shown in fig. 4, and structurally, the deep belief network model is composed of a multilayer unsupervised Boltzmann machine (RBM) network and a layer of supervised back-propagation (BP) network.
The training of the deep training network model comprises 2 steps of "pre-training" and "tuning-tuning". The pre-training stage deep training network trains RBMs in each layer in a layer-by-layer (layerwise) training mode, and the hidden layer output of the RBMs in the lower layer is used as the visible layer input of the RBMs in the upper layer. And in the fine-tuning stage, the last layer of BP network is trained in a supervised learning mode, errors between actual output and expected output are propagated backwards layer by layer, and the weight of the whole deep training network is optimized. The training process of the RBM network can be actually regarded as the initialization of the weight of the deep BP network, so that the deep training network overcomes the defects of long training time and easy falling into a local optimal solution of the BP network due to the random initialization of the weight parameter.
A restricted boltzmann machine.
The constrained boltzmann machine is a typical energy-based model (EBM) which is composed of a visible layer (visible layer) and a hidden layer (hidden layer), and a schematic diagram of the model is shown in fig. 5. Wherein v and h respectively represent a visible layer and a hidden layer, and W represents a connection weight between the two layers. For the visible layer and the hidden layer, the connection relationship is that the neurons between the layers are fully connected, and no neuron in the layers is connected.
Assuming that the visible layer and the hidden layer are binary variables, the numbers of neurons in the visible layer and the hidden layer are I, J and v respectivelyiAnd hjRespectively representing the states of the ith visible layer neuron and the jth hidden layer neuron. For a particular set of (v, h), the joint distribution function of the deep belief network model is:
<math> <mrow> <mi>E</mi> <mrow> <mo>(</mo> <mi>v</mi> <mo>,</mo> <mi>h</mi> <mo>|</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> <mo>=</mo> <mo>-</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>I</mi> </munderover> <msub> <mi>a</mi> <mi>i</mi> </msub> <msub> <mi>v</mi> <mi>i</mi> </msub> <mo>-</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>j</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>J</mi> </munderover> <msub> <mi>b</mi> <mi>j</mi> </msub> <msub> <mi>h</mi> <mi>j</mi> </msub> <mo>-</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>I</mi> </munderover> <munderover> <mi>&Sigma;</mi> <mrow> <mi>j</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>J</mi> </munderover> <msub> <mi>w</mi> <mi>ij</mi> </msub> <msub> <mi>v</mi> <mi>i</mi> </msub> <msub> <mi>h</mi> <mi>j</mi> </msub> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein θ ═ wij,ai,bj) For deep belief networksModel parameter, wijVisible layer ith node v representing deep belief networkiWith the jth node h of the hidden layerjThe connection weight value between aiAnd bjRespectively represent ith nodes v of visible layeriBias parameters and the jth node h of the hidden layerjThe bias parameter of (2);
and (3) obtaining the joint probability distribution of (v, h) by using the joint distribution function of the deep belief network model:
p(v,h|θ)=e-E(v,h|θ)/Z(θ); (2)
wherein,is a distribution function. In practical terms, we are most concerned with the distribution (likelihood function) defined by the RBM about the observed variable, i.e., p (v | θ), which is the edge distribution of the joint probability p (v, h | θ).
Due to the special structure of the RBM (no connection of neurons in the layers), the activation states of the various hidden layer nodes are independent of each other when given the state of a visible layer node. At this time, the jth node h of the hidden layerjThe activation probability of (c) is:
<math> <mrow> <mi>p</mi> <mrow> <mo>(</mo> <msub> <mi>h</mi> <mi>j</mi> </msub> <mo>=</mo> <mn>1</mn> <mo>|</mo> <mi>v</mi> <mo>,</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> <mo>=</mo> <mi>&sigma;</mi> <mrow> <mo>(</mo> <msub> <mi>b</mi> <mi>j</mi> </msub> <mo>+</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>I</mi> </munderover> <msub> <mi>v</mi> <mi>i</mi> </msub> <msub> <mi>w</mi> <mi>ji</mi> </msub> <mo>)</mo> </mrow> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>3</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein σ (x) is 1/(1+ e)-x) Is sigmoid function. Similarly, given the state of the hidden layer node, the ith node v of the visible layeriThe activation probability of (c) is:
<math> <mrow> <mi>p</mi> <mrow> <mo>(</mo> <msub> <mi>v</mi> <mi>i</mi> </msub> <mo>=</mo> <mn>1</mn> <mo>|</mo> <mi>h</mi> <mo>,</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> <mo>=</mo> <mi>&sigma;</mi> <mrow> <mo>(</mo> <msub> <mi>a</mi> <mi>j</mi> </msub> <mo>+</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>I</mi> </munderover> <msub> <mi>h</mi> <mi>i</mi> </msub> <msub> <mi>w</mi> <mi>ji</mi> </msub> <mo>)</mo> </mrow> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>4</mn> <mo>)</mo> </mrow> </mrow> </math>
the RBM is trained in an iterative mode, and the training aims at learning a parameter theta (w)ij,ai,bj) To fit given training data. The parameter θ can be obtained by solving a maximum log likelihood function on a training sample set (assuming the total number of training samples as T), that is:
<math> <mrow> <msup> <mi>&theta;</mi> <mo>*</mo> </msup> <mo>=</mo> <mi>arg</mi> <munder> <mi>max</mi> <mi>&theta;</mi> </munder> <mi>L</mi> <munderover> <mi>&Sigma;</mi> <mrow> <mi>t</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>T</mi> </munderover> <mi>ln</mi> <mi>p</mi> <mrow> <mo>(</mo> <msup> <mi>v</mi> <mrow> <mo>(</mo> <mi>t</mi> <mo>)</mo> </mrow> </msup> <mo>|</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>5</mn> <mo>)</mo> </mrow> </mrow> </math>
according to the activation probability, when a training sample is input to the visible layer node, after all nodes of the hidden layer are excited by adopting a joint distribution function of a deep belief network model, the next hidden layer node is excited, and thus the visible layer node is obtained again; and then, calculating the conditional distribution of the visible layer data by adopting a contrast divergence algorithm to obtain hidden layer data, calculating the visible layer data by using the conditional distribution of the hidden layer data obtained by calculation, reconstructing the visible layer data, and adjusting and updating the parameters of the deep belief network model by utilizing a gradient descent method. By using the Contrast Divergence (CD) algorithm proposed by Hinton, the updated difference values of the connection weight and the bias parameter can be obtained as follows:
Δwij=(<vihj>data-<vihj>recon) (6);
Δai=(<vi>data-<vi>recon) (7);
Δbj=(<hj>data-<hj>recon) (8);
wherein, Δ wijVisible layer ith node v representing deep belief networkiWith the jth node h of the hidden layerjThe connection weight value w betweenijPerforming an updated updateDifference, Δ aiAnd Δ bjRespectively represent ith nodes v of visible layeriBias parameter a ofiUpdating difference value and jth node h of hidden layer for updatingjBias parameter b ofjUpdating the updated difference value; in order to train the learning rate of the training,<·>datarepresenting a mathematical expectation over a distribution defined by the training data set,<·>reconrepresenting a mathematical expectation over the distribution of the reconstructed deep belief network model output.
And training layer by layer until the output of the last hidden layer of the deep belief network model is obtained.
Secondly, the output of the last hidden layer of the deep belief network model is subjected to the training of a back propagation network, the classification error of the classification result output by the training prediction and the actual classification result of the training sample is propagated backwards layer by layer, and the connection weight of each layer of the deep belief network model is optimized.
The back propagation network is a supervised classifier, classifies the feature vectors obtained by the RBM through pre-training, and plays a role in adjusting and optimizing parameters of the whole deep training network. The training of the back propagation network is divided into a forward propagation process and a back propagation process; in the forward propagation process, the output of the RBM is used as an input feature vector and is propagated to an output layer by layer to obtain a predicted classification category, the actual classification result of the training sample is determined according to the working condition label of the training sample, then the predicted classification result is compared with the actual classification result of the training sample to obtain a classification error, and the classification error is transmitted back layer by layer to adjust and optimize the parameters of the deep training network; in the backward propagation process, the value of the sensitivity of each layer needs to be calculated, and the sensitivity is transmitted from top to bottom to correct the connection weight of the deep belief network model.
For the output layer, assume the actual output of the ith node is oiThe desired output is diThen, the calculation expression of the sensitivity of the ith node is:
i=oi(1-oi)(di-oi); (9)
for the mth hidden layer, the computational expression of the sensitivity of the ith node is:
<math> <mrow> <msubsup> <mi>&delta;</mi> <mi>i</mi> <mi>m</mi> </msubsup> <mo>=</mo> <msubsup> <mi>y</mi> <mi>i</mi> <mi>m</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>y</mi> <mi>i</mi> <mi>m</mi> </msubsup> <mo>)</mo> </mrow> <munder> <mi>&Sigma;</mi> <mi>j</mi> </munder> <msubsup> <mi>w</mi> <mi>ij</mi> <mi>m</mi> </msubsup> <msubsup> <mi>&delta;</mi> <mi>j</mi> <mrow> <mi>m</mi> <mo>+</mo> <mn>1</mn> </mrow> </msubsup> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>10</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein,indicating the sensitivity of the ith node of the mth hidden layer,indicating the sensitivity of the ith node of the (m + 1) th hidden layer,represents the output of the ith node of the mth layer,representing the weight between the ith node of the mth layer and the jth node of the next layer;
after the sensitivity of each node in each hidden layer is obtained, the connection weight of the deep belief network model is updated and adjusted according to the following formula:
<math> <mrow> <msubsup> <mi>w</mi> <mi>ij</mi> <mi>m</mi> </msubsup> <mo>&LeftArrow;</mo> <msubsup> <mi>w</mi> <mi>ij</mi> <mi>m</mi> </msubsup> <mo>+</mo> <msub> <mi>&epsiv;</mi> <mrow> <mi>fine</mi> <mo>-</mo> <mi>tuning</mi> </mrow> </msub> <mo>&times;</mo> <msubsup> <mi>y</mi> <mi>i</mi> <mi>m</mi> </msubsup> <msubsup> <mi>&delta;</mi> <mi>j</mi> <mrow> <mi>m</mi> <mo>+</mo> <mn>1</mn> </mrow> </msubsup> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>11</mn> <mo>)</mo> </mrow> </mrow> </math>
<math> <mrow> <msubsup> <mi>b</mi> <mi>j</mi> <mi>m</mi> </msubsup> <mo>&LeftArrow;</mo> <msubsup> <mi>b</mi> <mi>j</mi> <mi>m</mi> </msubsup> <mo>+</mo> <msub> <mi>&epsiv;</mi> <mrow> <mi>fine</mi> <mo>-</mo> <mi>tuning</mi> </mrow> </msub> <mo>&times;</mo> <msubsup> <mi>&delta;</mi> <mi>j</mi> <mrow> <mi>m</mi> <mo>+</mo> <mn>1</mn> </mrow> </msubsup> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>12</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein,fine-tuningthe learning rate is shown to be adjusted and optimized,represents the jth node of the mth layerOf (1);
and finally determining the connection weight and the bias parameter of the whole depth belief network model after adjusting and optimizing the connection weight of each layer of the depth belief network model.
3) And respectively taking the training samples under various working conditions as the input of a deep belief network model for determining a connection weight and a bias parameter, carrying out deep learning on the training samples, and respectively reconstructing each training sample under each working condition by adopting the deep belief network model for determining the connection weight and the bias parameter to obtain a training sample reconstruction signal corresponding to each training sample under each working condition.
According to the characteristics of the deep belief network, the original data sample with lower error can be reconstructed by using the deep belief network model which is trained, adjusted and optimized and then determined to be connected with the weight and the offset parameter, so that the reconstructed signal of the data sample obtained by reconstruction can embody and depict the essential characteristics of the original data sample, and the essential characteristics can be used as the reference characteristics for classification and identification. Therefore, in the step, the deep belief network model for determining the connection weight and the offset parameter is used for reconstructing each training sample under each working condition, so that the training sample reconstruction signal corresponding to each training sample under each working condition is used as a matching reference for performing classification and identification on the test sample at a later stage.
4) And acquiring vibration acceleration signal data of the rolling bearing to be tested during rotation operation through the acceleration sensor, and performing denoising pretreatment to obtain a test sample.
5) And taking the test sample as the input of the depth belief network model for determining the connection weight and the offset parameter, carrying out depth learning on the test sample, and reconstructing the test sample by adopting the depth belief network model for determining the connection weight and the offset parameter to obtain a test sample reconstruction signal.
Similarly, the step reconstructs the test sample by using a depth belief network model for determining the connection weight and the offset parameter, and the obtained reconstructed signal of the test sample describes the essential characteristics contained in the vibration acceleration signal number of the rolling bearing to be tested, so that the intrinsic characteristics are matched with the essential characteristics embodied by the reconstructed signal of the training sample under various working conditions, and the identification of the fault working condition category of the rolling bearing to be tested is realized.
6) And taking the test sample reconstruction signal as the matching characteristic of the test sample, taking the training sample reconstruction signal corresponding to each training sample under each working condition as the matching reference, matching the test sample with the training sample by adopting a support vector machine classification method, and judging the working condition class to which the training sample most matched with the test sample belongs as the working condition class of the test sample, thereby obtaining the fault diagnosis result of the rolling bearing to be tested.
Support Vector Machines (SVMs) were proposed by Vapnik et al in AT & TBell laboratories in 1963, which are based on VC dimension theory and structure risk minimization principle in statistics, and seek an optimal compromise between model complexity (i.e., learning accuracy of a specific training sample) and learning ability (i.e., ability to correctly recognize an arbitrary sample) according to limited sample information to obtain the best popularization ability. The SVM maps the vectors into a higher-dimensional space, a maximum separation hyperplane is established in the high-dimensional space, two hyperplanes which are parallel to each other are established on two sides capable of separating the data hyperplanes, the separation hyperplane enables the distance between the two parallel hyperplanes to be maximized, and the larger the distance is, the smaller the error of the classification result is.
Fig. 6 is a schematic diagram of the position relationship of the normalized optimal hyperplane under two-dimensional two-class conditions, where H is a separating hyperplane H1, H2 are two mutually parallel hyperplanes, and H1, H2 are classification intervals d ═ 2/| | | w |. In order to ensure the linearization of the data, the data needs to be mapped to a kernel function space; meanwhile, in order to effectively separate the two classes, it should be ensured that the two classes are correctly separated to maximize the classification interval, i.e. the objective function is:
<math> <mrow> <munder> <mi>min</mi> <mrow> <mi>w</mi> <mo>,</mo> <mi>b</mi> </mrow> </munder> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <msup> <mrow> <mo>|</mo> <mo>|</mo> <mi>w</mi> <mo>|</mo> <mo>|</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <mi>C</mi> <mrow> <mo>(</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>N</mi> </munderover> <msub> <mi>&epsiv;</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <mrow> <mo>(</mo> <mi>w</mi> <mo>&CenterDot;</mo> <mi>w</mi> <mo>)</mo> </mrow> <mo>+</mo> <mi>C</mi> <mrow> <mo>(</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>N</mi> </munderover> <msub> <mi>&epsiv;</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>13</mn> <mo>)</mo> </mrow> </mrow> </math>
for the hyperplane H1, H2, there are:
equation (14) may be equivalent to:
wherein:denotes w andthe inner product of the two phases is,the table maps xi to kernel function space; b and C are constants;ithe variable > 0 is the relaxation variable,the error degree of the training sample is represented, and the larger the value of the error degree, the more the error sample is represented.
Applying Lagrange multiplier method to equations (13) and (15) yields:
wherein: alpha is alphai>0,βiAnd the Lagrangian coefficient is more than 0, and the Lagrangian function is L (w, b, alpha).
The formula (16) is for w,iAnd the partial derivative of b is zero, yielding:
equation (17) is substituted for equation (16), and the solution of the optimal hyperplane is equivalent to the solution of the following dual problem.
<math> <mrow> <mfenced open='' close=''> <mtable> <mtr> <mtd> <mi>max</mi> <mi>Q</mi> <mrow> <mo>(</mo> <mi>&alpha;</mi> <mo>)</mo> </mrow> <mo>=</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>N</mi> </munderover> <msub> <mi>&alpha;</mi> <mi>i</mi> </msub> <mo>-</mo> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>,</mo> <mi>j</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>N</mi> </munderover> <msub> <mi>&alpha;</mi> <mi>i</mi> </msub> <msub> <mi>&alpha;</mi> <mi>j</mi> </msub> <msub> <mi>y</mi> <mi>i</mi> </msub> <msub> <mi>y</mi> <mi>j</mi> </msub> <mi>K</mi> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>&CenterDot;</mo> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> </mtd> </mtr> <mtr> <mtd> <mi>s</mi> <mo>.</mo> <mi>t</mi> <mo>.</mo> <mi></mi> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>N</mi> </munderover> <msub> <mi>y</mi> <mi>i</mi> </msub> <msub> <mi>&alpha;</mi> <mi>i</mi> </msub> <mo>=</mo> <mn>0</mn> <mo>,</mo> <mi>C</mi> <msub> <mrow> <mo>></mo> <mi>&alpha;</mi> </mrow> <mi>i</mi> </msub> <mo>></mo> <mn>0</mn> </mtd> </mtr> </mtable> </mfenced> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>18</mn> <mo>)</mo> </mrow> </mrow> </math>
Wherein:
using Lagrange multiplier method, the solution is obtained as:
the classification rule function for obtaining the optimal classification surface from equation (19) is:
the invention selects a Gaussian Radial Basis (RBF) kernel function:
<math> <mrow> <mi>K</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>&CenterDot;</mo> <msub> <mi>x</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mi>exp</mi> <mrow> <mo>(</mo> <mo>-</mo> <mfrac> <msup> <mrow> <mo>|</mo> <mo>|</mo> <mi>x</mi> <mo>-</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>|</mo> <mo>|</mo> </mrow> <mn>2</mn> </msup> <msup> <mi>&sigma;</mi> <mn>2</mn> </msup> </mfrac> <mo>)</mo> </mrow> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>21</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein: σ is a parameter of the RBF kernel function.
The SVM algorithm seeks an optimal classification surface among data on the basis of statistics, and nonlinear data are mapped to a kernel function space to be linearized, so that the computational complexity is simplified, and a better classification effect is achieved.
In step 6), there are many specific classification methods that can be applied to the support vector machine classification method, such as one-to-one classification (OVO-SVM), one-to-many classification (also called one-to-many classification, OVR-SVM), directed acyclic graph classification (DAG-SVMs), decision tree classification, error correction output coding classification, and the like. However, considering that the method only relates to the fault working condition classification recognition of four working conditions, namely normal operation, bearing inner ring fault operation, bearing rolling element fault operation and bearing outer ring fault operation, the recognition categories are not many, and the factors of recognition efficiency and accuracy are comprehensively considered, the one-to-many SVM classification method is more suitable because the decision functions of the SVM classifiers needing to be established and subjected to recognition operation by the one-to-many classification method are only four (one for each fault working condition category), and after the vibration acceleration signal data of the rolling bearing under the four different working conditions is reconstructed by the deep belief network model, the difference of the essential characteristics is enough to be recognized, and the recognition accuracy can be better ensured.
In the method of the present invention, the step 6) preferably adopts a one-to-many support vector machine classification method to match the test sample with the training sample in the following specific way:
61) in the training samples of four working conditions, aiming at the situation that the kth class training sample is regarded as a positive class, k belongs to {1,2,3,4}, other 3 classes of training samples are regarded as negative classes, and a classification decision function f of the kth class is obtained by a classification method of two classes of support vector machinesk(x):
<math> <mrow> <msub> <mi>f</mi> <mi>k</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>=</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>n</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>N</mi> </munderover> <msubsup> <mi>&alpha;</mi> <mi>n</mi> <mi>k</mi> </msubsup> <msub> <mi>y</mi> <mi>n</mi> </msub> <mi>K</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <msub> <mi>x</mi> <mi>n</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msub> <mi>b</mi> <mi>k</mi> </msub> <mo>;</mo> </mrow> </math>
Wherein,classifying a decision function f for class kk(x) N-th training sample reconstructed signal xnA corresponding Lagrange coefficient; bkClassifying a decision function f for class kk(x) The optimal hyperplane position coefficient; y isnRepresenting the classification mark corresponding to the nth training sample, y when the nth training sample belongs to the positive classn1, y when the nth training sample belongs to the negative classn-1; n belongs to {1,2, …, N }, wherein N is the total number of training samples of four working conditions; k (x, x)n) Representing a classification decision function fk(x) Is compared with the n-th training sample to reconstruct the signal xnGaussian radial basis kernel function of (1);
thus obtaining a classification decision function corresponding to each working condition in the four working conditions;
62) and respectively taking the test sample reconstruction signals as input quantities of the classification decision functions corresponding to the four working conditions, calculating the four classification decision function values of the test sample reconstruction signals as the input quantities, and judging the working condition type corresponding to the largest classification decision function value as the working condition type of the test sample to obtain a fault diagnosis result of the rolling bearing to be tested.
Through experimental data verification, the rolling bearing fault diagnosis method based on deep learning and the support vector machine carries out fault diagnosis according to the process, and under the condition of 600 training samples (150 training samples in each working condition), rolling bearing fault diagnosis and identification are carried out by random sampling for 800 times, the identification accuracy reaches 94.3 percent (the average fault identification accuracy in the industry is only about 83 percent), and the actual application requirements can be completely met.
In conclusion, the rolling bearing fault diagnosis method based on deep learning and the support vector machine of the invention utilizes the mature learning algorithm in the deep belief network theory to complete the feature extraction task required by fault diagnosis, can automatically extract the essential features of input data from simple to complex and from low level to high level without depending on manual selection, can save a large amount of manpower, can automatically dig out rich information hidden in known data, and is particularly suitable for noisy, uncertain and dynamic systems; in addition, because the support vector machine classification method is adopted to classify and identify the test samples, the learning process in the support vector machine classification method can be regarded as a process for optimally searching an optimal solution, so that a previously designed effective method can be adopted to search and find a global minimum value of a target function, and the problem that a deep belief network possibly falls into a local optimal solution is solved; compared with the prior art, the rolling bearing fault diagnosis method can improve the accuracy and effectiveness of rolling bearing fault diagnosis, provides a new effective way for solving the problem of rolling bearing fault diagnosis, and can be widely applied to complex systems in the fields of chemical industry, metallurgy, electric power, aviation and the like.
Finally, the above embodiments are only for illustrating the technical solutions of the present invention and not for limiting, although the present invention has been described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications or equivalent substitutions may be made to the technical solutions of the present invention without departing from the spirit and scope of the technical solutions of the present invention, and all of them should be covered in the claims of the present invention.

Claims (3)

1. The rolling bearing fault diagnosis method based on deep learning and a support vector machine is characterized by comprising the following steps:
1) when the rolling bearings under four different working conditions work in a rotating mode, vibration acceleration signals of the rolling bearings under each working condition working at different rotating speeds are collected through the acceleration sensors respectively, denoising pretreatment is carried out, working condition labels are added, and vibration acceleration signal data under various working conditions after pretreatment and working condition labels are added are used as training samples; the four working conditions are normal operation, bearing inner ring fault operation, bearing rolling element fault operation and bearing outer ring fault operation respectively;
2) establishing a deep belief network model, training the deep belief network model by adopting a training sample, inputting the training sample into the deep belief network model, and performing layer-by-layer training and optimization by adopting an unsupervised greedy layer-by-layer training method to obtain a connection weight and an offset parameter of the deep belief network model;
3) respectively taking training samples under various working conditions as input of a deep belief network model for determining a connection weight and a bias parameter, carrying out deep learning on the training samples, and respectively reconstructing each training sample under each working condition by adopting the deep belief network model for determining the connection weight and the bias parameter to obtain a training sample reconstruction signal corresponding to each training sample under each working condition;
4) acquiring vibration acceleration signal data of a rolling bearing to be tested during rotation work through an acceleration sensor, and performing denoising pretreatment to obtain a test sample;
5) taking the test sample as the input of a depth belief network model for determining a connection weight and a bias parameter, carrying out depth learning on the test sample, and reconstructing the test sample by adopting the depth belief network model for determining the connection weight and the bias parameter to obtain a test sample reconstruction signal;
6) and taking the test sample reconstruction signal as the matching characteristic of the test sample, taking the training sample reconstruction signal corresponding to each training sample under each working condition as the matching reference, matching the test sample with the training sample by adopting a support vector machine classification method, and judging the working condition class to which the training sample most matched with the test sample belongs as the working condition class of the test sample, thereby obtaining the fault diagnosis result of the rolling bearing to be tested.
2. The rolling bearing fault diagnosis method based on deep learning and support vector machine according to claim 1, wherein the joint distribution function of the deep belief network model established in the step 2) is:
<math> <mrow> <mi>E</mi> <mrow> <mo>(</mo> <mi>v</mi> <mo>,</mo> <mi>h</mi> <mo>|</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> <mo>=</mo> <mo>-</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>I</mi> </munderover> <msub> <mi>a</mi> <mi>i</mi> </msub> <msub> <mi>v</mi> <mi>i</mi> </msub> <mo>-</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>j</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>J</mi> </munderover> <msub> <mi>b</mi> <mi>j</mi> </msub> <msub> <mi>h</mi> <mi>j</mi> </msub> <mo>-</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>I</mi> </munderover> <munderover> <mi>&Sigma;</mi> <mrow> <mi>j</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>J</mi> </munderover> <msub> <mi>w</mi> <mi>ij</mi> </msub> <msub> <mi>v</mi> <mi>i</mi> </msub> <msub> <mi>h</mi> <mi>j</mi> </msub> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein θ ═ wij,ai,bj) For deep belief network model parameters, wijVisible layer ith node v representing deep belief networkiWith the jth node h of the hidden layerjThe connection weight value between aiAnd bjRespectively represent ith nodes v of visible layeriBias parameters and the jth node h of the hidden layerjThe bias parameter of (2);
an unsupervised greedy layer-by-layer training method is adopted to carry out layer-by-layer training and optimization on the deep belief network model, and the specific mode is as follows:
21) training the limited Boltzmann machines in each layer of the deep belief network model in a layer-by-layer training mode, wherein the hidden layer output of the limited Boltzmann machine at the lower layer is used as the visible layer input of the limited Boltzmann machine at the upper layer until the output of the last hidden layer of the deep belief network model is obtained; the method specifically comprises the following steps:
the joint probability distribution probability p (v, h | θ) of the visible layer node and the hidden layer node is:
p(v,h|θ)=e-E(v,h|θ)/Z(θ);(2)
wherein, <math> <mrow> <mi>Z</mi> <mrow> <mo>(</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mi>&Sigma;</mi> <mi>v</mi> </munder> <munder> <mi>&Sigma;</mi> <mi>h</mi> </munder> <msup> <mi>e</mi> <mrow> <mo>-</mo> <mi>E</mi> <mrow> <mo>(</mo> <mi>v</mi> <mo>,</mo> <mi>h</mi> <mo>|</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> </mrow> </msup> </mrow> </math> is a distribution function;
when the state of the visible layer node is given, the jth node h of the hidden layerjThe activation probability of (c) is:
<math> <mrow> <mi>p</mi> <mrow> <mo>(</mo> <msub> <mi>h</mi> <mi>j</mi> </msub> <mo>=</mo> <mn>1</mn> <mo>|</mo> <mi>v</mi> <mo>,</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> <mo>=</mo> <mi>&sigma;</mi> <mrow> <mo>(</mo> <msub> <mi>b</mi> <mi>j</mi> </msub> <mo>+</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>I</mi> </munderover> <msub> <mi>v</mi> <mi>i</mi> </msub> <msub> <mi>w</mi> <mi>ji</mi> </msub> <mo>)</mo> </mrow> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>3</mn> <mo>)</mo> </mrow> </mrow> </math>
given the state of the hidden layer node, the ith node v of the visible layeriThe activation probability of (c) is:
<math> <mrow> <mi>p</mi> <mrow> <mo>(</mo> <msub> <mi>v</mi> <mi>i</mi> </msub> <mo>=</mo> <mn>1</mn> <mo>|</mo> <mi>h</mi> <mo>,</mo> <mi>&theta;</mi> <mo>)</mo> </mrow> <mo>=</mo> <mi>&sigma;</mi> <mrow> <mo>(</mo> <msub> <mi>a</mi> <mi>i</mi> </msub> <mo>+</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>I</mi> </munderover> <msub> <mi>h</mi> <mi>j</mi> </msub> <msub> <mi>w</mi> <mi>ji</mi> </msub> <mo>)</mo> </mrow> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>4</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein σ (x) is 1/(1+ e)-x) Is sigmoid function;
according to the activation probability, when a given training sample is input to the visible layer node, after all nodes of the hidden layer are excited by adopting a joint distribution function of a deep belief network model, the nodes of the next hidden layer are excited, so that the visible layer node is obtained again; then, calculating the condition distribution of the visible layer data by adopting a contrast divergence algorithm to obtain hidden layer data, calculating the visible layer data by calculating the condition distribution of the hidden layer data, reconstructing the visible layer data, adjusting and updating the parameters of the deep belief network model by utilizing a gradient descent method, wherein the updating difference values of the connection weight and the offset parameter are respectively as follows:
Δwij=(<vihj>data-<vihj>recon);(6)
Δai=(<vi>data-<vi>recon);(7)
Δbj=(<hj>data-<hj>recon);(8)
wherein, Δ wijVisible layer ith node v representing deep belief networkiWith the jth node h of the hidden layerjThe connection weight value w betweenijUpdate difference, Δ a, for performing an updateiAnd Δ bjRespectively represent ith nodes v of visible layeriBias parameter a ofiUpdating difference value and jth node h of hidden layer for updatingjBias parameter b ofjUpdating the updated difference value; in order to train the learning rate of the training,<·>datarepresenting a mathematical expectation over a distribution defined by the training data set,<·>reconrepresenting a mathematical expectation on the distribution of the reconstructed deep belief network model output;
training layer by layer until the output of the last hidden layer of the deep belief network model is obtained;
22) training a back propagation network for the output of the last hidden layer of the deep belief network model obtained in the step 21, propagating the classification error of the classification result output by training prediction and the actual classification result of the training sample backward layer by layer, and adjusting the connection weight of each layer of the deep belief network model; the method specifically comprises the following steps:
the training of the back propagation network is divided into a forward propagation process and a back propagation process; in the forward propagation process, the output of the last hidden layer of the deep belief network model obtained in the step 21 is propagated to an output layer by layer as input to obtain a predicted classification category, the actual classification result of the training sample is determined according to the working condition label of the training sample, then the predicted classification result is compared with the actual classification result of the training sample to obtain a classification error, and the classification error is transmitted back layer by layer, so that the parameters of the deep training network are adjusted; in the backward propagation process, the value of the sensitivity of each layer needs to be calculated, and the sensitivity is transmitted from top to bottom to correct the connection weight of the deep belief network model;
for the output layer, assume the actual output of the ith node is oiThe desired output is diThen, the calculation expression of the sensitivity of the ith node is:
i=oi(1-oi)(di-oi);(9)
for the mth hidden layer, the computational expression of the sensitivity of the ith node is:
<math> <mrow> <msubsup> <mi>&delta;</mi> <mi>i</mi> <mi>m</mi> </msubsup> <mo>=</mo> <msubsup> <mi>y</mi> <mi>i</mi> <mi>m</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>y</mi> <mi>i</mi> <mi>m</mi> </msubsup> <mo>)</mo> </mrow> <munder> <mi>&Sigma;</mi> <mi>j</mi> </munder> <msubsup> <mi>w</mi> <mi>ij</mi> <mi>m</mi> </msubsup> <msubsup> <mi>&delta;</mi> <mi>j</mi> <mrow> <mi>m</mi> <mo>+</mo> <mn>1</mn> </mrow> </msubsup> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>10</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein,indicating the sensitivity of the ith node of the mth hidden layer,indicating the sensitivity of the ith node of the (m + 1) th hidden layer,represents the output of the ith node of the mth hidden layer,representing the connection weight between the ith node of the mth hidden layer and the jth node of the (m + 1) th hidden layer;
after the sensitivity of each node in each hidden layer is obtained, the connection weight of the deep belief network model is updated and adjusted according to the following formula:
<math> <mrow> <msubsup> <mi>w</mi> <mi>ij</mi> <mi>m</mi> </msubsup> <mo>&LeftArrow;</mo> <msubsup> <mi>w</mi> <mi>ij</mi> <mi>m</mi> </msubsup> <mo>+</mo> <msub> <mi>&epsiv;</mi> <mrow> <mi>fine</mi> <mo>-</mo> <mi>tuning</mi> </mrow> </msub> <mo>&times;</mo> <msubsup> <mi>y</mi> <mi>i</mi> <mi>m</mi> </msubsup> <msubsup> <mi>&delta;</mi> <mi>j</mi> <mrow> <mi>m</mi> <mo>+</mo> <mn>1</mn> </mrow> </msubsup> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>11</mn> <mo>)</mo> </mrow> </mrow> </math>
<math> <mrow> <msubsup> <mi>b</mi> <mi>j</mi> <mi>m</mi> </msubsup> <mo>&LeftArrow;</mo> <msubsup> <mi>b</mi> <mi>j</mi> <mi>m</mi> </msubsup> <mo>+</mo> <msub> <mi>&epsiv;</mi> <mrow> <mi>fine</mi> <mo>-</mo> <mi>tuning</mi> </mrow> </msub> <mo>&times;</mo> <msubsup> <mi>&delta;</mi> <mi>j</mi> <mrow> <mi>m</mi> <mo>+</mo> <mn>1</mn> </mrow> </msubsup> <mo>;</mo> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>12</mn> <mo>)</mo> </mrow> </mrow> </math>
wherein,fine-tuningthe learning rate is shown to be adjusted and optimized,representing the bias parameter of the jth node of the mth hidden layer;
and finally determining the connection weight and the bias parameter of the whole depth belief network model after adjusting and optimizing the connection weight of each layer of the depth belief network model.
3. The rolling bearing fault diagnosis method based on deep learning and support vector machine according to claim 1, wherein the specific way of matching the test sample and the training sample in step 6) by using the support vector machine classification method is as follows:
61) in the training samples of four working conditions, aiming at the situation that the kth class training sample is regarded as a positive class, k belongs to {1,2,3,4}, other 3 classes of training samples are regarded as negative classes, and a classification decision function f of the kth class is obtained by a classification method of two classes of support vector machinesk(x):
<math> <mrow> <msub> <mi>f</mi> <mi>k</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>=</mo> <munderover> <mi>&Sigma;</mi> <mrow> <mi>n</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>N</mi> </munderover> <msubsup> <mi>&alpha;</mi> <mi>n</mi> <mi>k</mi> </msubsup> <msub> <mi>y</mi> <mi>n</mi> </msub> <mi>K</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <msub> <mi>x</mi> <mi>n</mi> </msub> <mo>)</mo> </mrow> <mo>+</mo> <msub> <mi>b</mi> <mi>k</mi> </msub> <mo>;</mo> </mrow> </math>
Wherein,classifying a decision function f for class kk(x) N-th training sample reconstructed signal xnA corresponding Lagrange coefficient; bkClassifying a decision function f for class kk(x) The optimal hyperplane position coefficient; y isnRepresenting the classification mark corresponding to the nth training sample, y when the nth training sample belongs to the positive classn1, y when the nth training sample belongs to the negative classn-1; n belongs to {1,2, …, N }, wherein N is the total number of training samples of four working conditions; k (x, x)n) Representing a classification decision function fk(x) Is compared with the n-th training sample to reconstruct the signal xnGaussian radial basis kernel function of (1);
thus obtaining a classification decision function corresponding to each working condition in the four working conditions;
62) and respectively taking the test sample reconstruction signals as input quantities of the classification decision functions corresponding to the four working conditions, calculating the four classification decision function values of the test sample reconstruction signals as the input quantities, and judging the working condition type corresponding to the largest classification decision function value as the working condition type of the test sample to obtain a fault diagnosis result of the rolling bearing to be tested.
CN201510080839.5A 2015-02-13 2015-02-13 Fault diagnosis method for rolling bearing based on deep learning and SVM (Support Vector Machine) Pending CN104616033A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510080839.5A CN104616033A (en) 2015-02-13 2015-02-13 Fault diagnosis method for rolling bearing based on deep learning and SVM (Support Vector Machine)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510080839.5A CN104616033A (en) 2015-02-13 2015-02-13 Fault diagnosis method for rolling bearing based on deep learning and SVM (Support Vector Machine)

Publications (1)

Publication Number Publication Date
CN104616033A true CN104616033A (en) 2015-05-13

Family

ID=53150470

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510080839.5A Pending CN104616033A (en) 2015-02-13 2015-02-13 Fault diagnosis method for rolling bearing based on deep learning and SVM (Support Vector Machine)

Country Status (1)

Country Link
CN (1) CN104616033A (en)

Cited By (51)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105300693A (en) * 2015-09-25 2016-02-03 东南大学 Bearing fault diagnosis method based on transfer learning
CN105760839A (en) * 2016-02-22 2016-07-13 重庆大学 Bearing fault diagnosis method based on multi-feature manifold learning and support vector machine
CN105823634A (en) * 2016-05-10 2016-08-03 东莞理工学院 Bearing damage identification method based on time frequency relevance vector convolution Boltzmann machine
CN105930861A (en) * 2016-04-13 2016-09-07 西安西拓电气股份有限公司 Adaboost algorithm based transformer fault diagnosis method
CN105973594A (en) * 2016-04-25 2016-09-28 西北工业大学 Rolling bearing fault prediction method based on continuous deep belief network
CN105973595A (en) * 2016-04-27 2016-09-28 清华大学深圳研究生院 Diagnosis method of rolling bearing fault
CN106092574A (en) * 2016-05-30 2016-11-09 西安工业大学 The Method for Bearing Fault Diagnosis selected with sensitive features is decomposed based on improving EMD
CN106291233A (en) * 2016-07-29 2017-01-04 武汉大学 A kind of fault phase-selecting method based on convolutional neural networks
CN106291234A (en) * 2016-07-29 2017-01-04 武汉大学 A kind of transmission line of electricity internal fault external fault based on convolutional neural networks judges and fault phase-selecting method
CN106409027A (en) * 2016-11-14 2017-02-15 贵州师范学院 Deep learning based brain-wave intelligent learning machine and program control method thereof
CN106441888A (en) * 2016-09-07 2017-02-22 广西大学 High-speed train rolling bearing fault diagnosis method
CN106598917A (en) * 2016-12-07 2017-04-26 国家海洋局第二海洋研究所 Upper ocean thermal structure prediction method based on deep belief network
CN106874957A (en) * 2017-02-27 2017-06-20 苏州大学 A kind of Fault Diagnosis of Roller Bearings
CN106886660A (en) * 2017-03-23 2017-06-23 哈尔滨理工大学 EEMD Hilbert envelope spectrums are with Rolling Bearing Status recognition methods under the varying load that DBN is combined
CN106898354A (en) * 2017-03-03 2017-06-27 清华大学 Speaker number estimation method based on DNN models and supporting vector machine model
CN106980873A (en) * 2017-03-09 2017-07-25 南京理工大学 Fancy carp screening technique and device based on deep learning
WO2017124336A1 (en) * 2016-01-20 2017-07-27 Sensetime Group Limited Method and system for adapting deep model for object representation from source domain to target domain
CN107036816A (en) * 2016-11-17 2017-08-11 重庆工商大学 A kind of Aero-engine Bearing method for diagnosing faults
CN107247231A (en) * 2017-07-28 2017-10-13 南京航空航天大学 A kind of aerogenerator fault signature extracting method based on OBLGWO DBN models
CN107463938A (en) * 2017-06-26 2017-12-12 南京航空航天大学 A kind of aerial engine air passage unit failure detection method based on space correction SVMs
CN107634857A (en) * 2017-09-22 2018-01-26 安徽师范大学 Fault Model structure and appraisal procedure based on SVM
CN107657634A (en) * 2017-09-06 2018-02-02 上海电力学院 Shale digital cores three-dimensional reconstruction method based on deep learning and SVMs
CN107657088A (en) * 2017-09-07 2018-02-02 南京工业大学 Rolling bearing fault diagnosis method based on MCKD algorithm and support vector machine
CN107917805A (en) * 2017-10-16 2018-04-17 铜仁职业技术学院 Fault Diagnosis of Roller Bearings based on depth belief network and support vector machines
CN108241589A (en) * 2016-12-26 2018-07-03 大唐移动通信设备有限公司 A kind of data processing method and device
CN108304927A (en) * 2018-01-25 2018-07-20 清华大学 Bearing fault modality diagnostic method and system based on deep learning
CN108519768A (en) * 2018-03-26 2018-09-11 华中科技大学 A kind of method for diagnosing faults analyzed based on deep learning and signal
CN108734617A (en) * 2018-04-25 2018-11-02 黄冈职业技术学院 A kind of intelligence municipal works cost information storage system
CN108875281A (en) * 2018-08-08 2018-11-23 佛山科学技术学院 A kind of hybrid intelligent health status diagnostic method and device based on deep learning
CN108959498A (en) * 2018-06-26 2018-12-07 北京航空航天大学 A kind of big data processing platform and its design method for health monitoring
CN109102005A (en) * 2018-07-23 2018-12-28 杭州电子科技大学 Small sample deep learning method based on shallow Model knowledge migration
CN109188026A (en) * 2018-10-25 2019-01-11 北京航空航天大学 The deep learning method of automatic Calibration suitable for mems accelerometer
CN109490814A (en) * 2018-09-07 2019-03-19 广西电网有限责任公司电力科学研究院 Metering automation terminal fault diagnostic method based on deep learning and Support Vector data description
CN109520611A (en) * 2018-11-08 2019-03-26 温州大学 A kind of monitoring method of earthquake simulation shaking table operating condition
CN109669017A (en) * 2017-10-17 2019-04-23 中国石油化工股份有限公司 Refinery's distillation tower top based on deep learning cuts water concentration prediction technique
CN109685331A (en) * 2018-12-06 2019-04-26 中国软件与技术服务股份有限公司 A kind of high-speed rail bogie sensor fault diagnosis method based on machine learning
CN109791401A (en) * 2016-09-16 2019-05-21 西门子股份公司 Generate the fault model for being used for embedded analysis and diagnosis/Forecast reasoning
CN110009161A (en) * 2019-04-15 2019-07-12 成都四方伟业软件股份有限公司 Water supply forecast method and device
CN110268350A (en) * 2017-03-03 2019-09-20 松下知识产权经营株式会社 Deteriorate the additional learning method of diagnostic system
CN110341986A (en) * 2019-07-16 2019-10-18 哈尔滨工业大学 Aircraft auxiliary power plant performance parameter multistep forecasting method based on RBM optimization ELM
CN110456732A (en) * 2019-08-06 2019-11-15 武汉恒力华振科技有限公司 A kind of punching machine fault monitoring system with learning functionality
CN110622043A (en) * 2017-05-12 2019-12-27 株式会社普利司通 Road surface state discrimination method and road surface state discrimination device
CN111044285A (en) * 2019-11-22 2020-04-21 军事科学院系统工程研究院军用标准研究中心 Method for diagnosing faults of mechanical equipment under complex conditions
CN111539381A (en) * 2020-05-18 2020-08-14 中车永济电机有限公司 Construction method of wind turbine bearing fault classification diagnosis model
CN111734772A (en) * 2020-06-16 2020-10-02 大连理工大学 Magnetorheological vibration suppression method in thin-wall part machining
CN111827982A (en) * 2019-04-17 2020-10-27 中国石油天然气集团有限公司 Method and device for predicting overflow and leakage working conditions of drilling well
CN111860429A (en) * 2020-07-30 2020-10-30 科大讯飞股份有限公司 Blast furnace tuyere abnormality detection method, device, electronic apparatus, and storage medium
CN112379659A (en) * 2020-11-12 2021-02-19 西安石油大学 Petroleum drilling machine fault prediction system
CN113469217A (en) * 2021-06-01 2021-10-01 桂林电子科技大学 Unmanned automobile navigation sensor abnormity detection method based on deep learning
CN114004374A (en) * 2021-10-22 2022-02-01 南京信息工程大学 Air conditioning unit sensor fault diagnosis method based on Boltzmann machine voting
CN114016998A (en) * 2020-07-16 2022-02-08 中国石油天然气集团有限公司 Prediction method and device for logging encountering block

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100116060A1 (en) * 2007-03-26 2010-05-13 Tsunemi Murayama Method and system for abnormality diagnosis of very low speed rotating machine
CN104091181A (en) * 2014-07-15 2014-10-08 中国科学院合肥物质科学研究院 Injurious insect image automatic recognition method and system based on deep restricted Boltzmann machine
CN104198184A (en) * 2014-08-11 2014-12-10 中国人民解放军空军工程大学 Bearing fault diagnostic method based on second generation wavelet transform and BP neural network

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100116060A1 (en) * 2007-03-26 2010-05-13 Tsunemi Murayama Method and system for abnormality diagnosis of very low speed rotating machine
CN104091181A (en) * 2014-07-15 2014-10-08 中国科学院合肥物质科学研究院 Injurious insect image automatic recognition method and system based on deep restricted Boltzmann machine
CN104198184A (en) * 2014-08-11 2014-12-10 中国人民解放军空军工程大学 Bearing fault diagnostic method based on second generation wavelet transform and BP neural network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
吕启等: "基于DBN模型的遥感图像分类", 《计算机研究与发展》 *
胥永刚等: "基于双树复小波包变换和SVM的滚动轴承故障诊断方法", 《航天动力学报》 *

Cited By (74)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105300693B (en) * 2015-09-25 2016-10-12 东南大学 A kind of Method for Bearing Fault Diagnosis based on transfer learning
CN105300693A (en) * 2015-09-25 2016-02-03 东南大学 Bearing fault diagnosis method based on transfer learning
WO2017124336A1 (en) * 2016-01-20 2017-07-27 Sensetime Group Limited Method and system for adapting deep model for object representation from source domain to target domain
CN108604304A (en) * 2016-01-20 2018-09-28 商汤集团有限公司 For adapting the depth model indicated for object from source domain to the method and system of aiming field
CN105760839A (en) * 2016-02-22 2016-07-13 重庆大学 Bearing fault diagnosis method based on multi-feature manifold learning and support vector machine
CN105930861A (en) * 2016-04-13 2016-09-07 西安西拓电气股份有限公司 Adaboost algorithm based transformer fault diagnosis method
CN105930861B (en) * 2016-04-13 2019-08-06 西安西拓电气股份有限公司 A kind of Diagnosis Method of Transformer Faults based on Adaboost algorithm
CN105973594A (en) * 2016-04-25 2016-09-28 西北工业大学 Rolling bearing fault prediction method based on continuous deep belief network
CN105973594B (en) * 2016-04-25 2018-05-29 西北工业大学 A kind of rolling bearing fault Forecasting Methodology based on continuous depth confidence network
CN105973595A (en) * 2016-04-27 2016-09-28 清华大学深圳研究生院 Diagnosis method of rolling bearing fault
CN105823634A (en) * 2016-05-10 2016-08-03 东莞理工学院 Bearing damage identification method based on time frequency relevance vector convolution Boltzmann machine
CN105823634B (en) * 2016-05-10 2018-04-13 东莞理工学院 Damage of the bearing recognition methods based on time-frequency interconnection vector convolution Boltzmann machine
CN106092574A (en) * 2016-05-30 2016-11-09 西安工业大学 The Method for Bearing Fault Diagnosis selected with sensitive features is decomposed based on improving EMD
CN106092574B (en) * 2016-05-30 2018-04-17 西安工业大学 Based on the Method for Bearing Fault Diagnosis for improving EMD decomposition and sensitive features selection
CN106291234B (en) * 2016-07-29 2019-07-23 武汉大学 A kind of judgement of transmission line of electricity internal fault external fault and fault phase-selecting method based on convolutional neural networks
CN106291233B (en) * 2016-07-29 2019-07-23 武汉大学 A kind of fault phase-selecting method based on convolutional neural networks
CN106291234A (en) * 2016-07-29 2017-01-04 武汉大学 A kind of transmission line of electricity internal fault external fault based on convolutional neural networks judges and fault phase-selecting method
CN106291233A (en) * 2016-07-29 2017-01-04 武汉大学 A kind of fault phase-selecting method based on convolutional neural networks
CN106441888A (en) * 2016-09-07 2017-02-22 广西大学 High-speed train rolling bearing fault diagnosis method
CN109791401A (en) * 2016-09-16 2019-05-21 西门子股份公司 Generate the fault model for being used for embedded analysis and diagnosis/Forecast reasoning
CN106409027A (en) * 2016-11-14 2017-02-15 贵州师范学院 Deep learning based brain-wave intelligent learning machine and program control method thereof
CN107036816A (en) * 2016-11-17 2017-08-11 重庆工商大学 A kind of Aero-engine Bearing method for diagnosing faults
CN107036816B (en) * 2016-11-17 2019-06-11 重庆工商大学 A kind of Aero-engine Bearing method for diagnosing faults
CN106598917B (en) * 2016-12-07 2019-08-06 国家海洋局第二海洋研究所 A kind of upper ocean heat structure prediction technique based on deepness belief network
CN106598917A (en) * 2016-12-07 2017-04-26 国家海洋局第二海洋研究所 Upper ocean thermal structure prediction method based on deep belief network
CN108241589A (en) * 2016-12-26 2018-07-03 大唐移动通信设备有限公司 A kind of data processing method and device
CN106874957A (en) * 2017-02-27 2017-06-20 苏州大学 A kind of Fault Diagnosis of Roller Bearings
CN106898354A (en) * 2017-03-03 2017-06-27 清华大学 Speaker number estimation method based on DNN models and supporting vector machine model
CN110268350A (en) * 2017-03-03 2019-09-20 松下知识产权经营株式会社 Deteriorate the additional learning method of diagnostic system
CN106898354B (en) * 2017-03-03 2020-05-19 北京华控智加科技有限公司 Method for estimating number of speakers based on DNN model and support vector machine model
CN110268350B (en) * 2017-03-03 2022-12-20 松下知识产权经营株式会社 Additional learning method for deterioration diagnosis system
CN106980873A (en) * 2017-03-09 2017-07-25 南京理工大学 Fancy carp screening technique and device based on deep learning
CN106886660A (en) * 2017-03-23 2017-06-23 哈尔滨理工大学 EEMD Hilbert envelope spectrums are with Rolling Bearing Status recognition methods under the varying load that DBN is combined
CN110622043A (en) * 2017-05-12 2019-12-27 株式会社普利司通 Road surface state discrimination method and road surface state discrimination device
CN107463938B (en) * 2017-06-26 2021-02-26 南京航空航天大学 Aero-engine gas circuit component fault detection method based on interval correction support vector machine
CN107463938A (en) * 2017-06-26 2017-12-12 南京航空航天大学 A kind of aerial engine air passage unit failure detection method based on space correction SVMs
CN107247231A (en) * 2017-07-28 2017-10-13 南京航空航天大学 A kind of aerogenerator fault signature extracting method based on OBLGWO DBN models
CN107657634A (en) * 2017-09-06 2018-02-02 上海电力学院 Shale digital cores three-dimensional reconstruction method based on deep learning and SVMs
CN107657634B (en) * 2017-09-06 2020-11-27 上海电力学院 Shale digital core three-dimensional reconstruction method based on deep learning and support vector machine
CN107657088B (en) * 2017-09-07 2021-06-11 南京工业大学 Rolling bearing fault diagnosis method based on MCKD algorithm and support vector machine
CN107657088A (en) * 2017-09-07 2018-02-02 南京工业大学 Rolling bearing fault diagnosis method based on MCKD algorithm and support vector machine
CN107634857A (en) * 2017-09-22 2018-01-26 安徽师范大学 Fault Model structure and appraisal procedure based on SVM
CN107634857B (en) * 2017-09-22 2020-09-08 安徽师范大学 Fault detection model construction and evaluation method based on SVM
CN107917805A (en) * 2017-10-16 2018-04-17 铜仁职业技术学院 Fault Diagnosis of Roller Bearings based on depth belief network and support vector machines
CN109669017A (en) * 2017-10-17 2019-04-23 中国石油化工股份有限公司 Refinery's distillation tower top based on deep learning cuts water concentration prediction technique
CN109669017B (en) * 2017-10-17 2021-04-27 中国石油化工股份有限公司 Refinery distillation tower top cut water ion concentration prediction method based on deep learning
CN108304927A (en) * 2018-01-25 2018-07-20 清华大学 Bearing fault modality diagnostic method and system based on deep learning
CN108519768B (en) * 2018-03-26 2019-10-08 华中科技大学 A kind of method for diagnosing faults analyzed based on deep learning and signal
CN108519768A (en) * 2018-03-26 2018-09-11 华中科技大学 A kind of method for diagnosing faults analyzed based on deep learning and signal
CN108734617A (en) * 2018-04-25 2018-11-02 黄冈职业技术学院 A kind of intelligence municipal works cost information storage system
CN108959498A (en) * 2018-06-26 2018-12-07 北京航空航天大学 A kind of big data processing platform and its design method for health monitoring
CN109102005A (en) * 2018-07-23 2018-12-28 杭州电子科技大学 Small sample deep learning method based on shallow Model knowledge migration
CN108875281A (en) * 2018-08-08 2018-11-23 佛山科学技术学院 A kind of hybrid intelligent health status diagnostic method and device based on deep learning
CN109490814A (en) * 2018-09-07 2019-03-19 广西电网有限责任公司电力科学研究院 Metering automation terminal fault diagnostic method based on deep learning and Support Vector data description
CN109490814B (en) * 2018-09-07 2021-02-26 广西电网有限责任公司电力科学研究院 Metering automation terminal fault diagnosis method based on deep learning and support vector data description
CN109188026A (en) * 2018-10-25 2019-01-11 北京航空航天大学 The deep learning method of automatic Calibration suitable for mems accelerometer
CN109520611A (en) * 2018-11-08 2019-03-26 温州大学 A kind of monitoring method of earthquake simulation shaking table operating condition
CN109685331A (en) * 2018-12-06 2019-04-26 中国软件与技术服务股份有限公司 A kind of high-speed rail bogie sensor fault diagnosis method based on machine learning
CN110009161A (en) * 2019-04-15 2019-07-12 成都四方伟业软件股份有限公司 Water supply forecast method and device
CN111827982A (en) * 2019-04-17 2020-10-27 中国石油天然气集团有限公司 Method and device for predicting overflow and leakage working conditions of drilling well
CN110341986A (en) * 2019-07-16 2019-10-18 哈尔滨工业大学 Aircraft auxiliary power plant performance parameter multistep forecasting method based on RBM optimization ELM
CN110341986B (en) * 2019-07-16 2021-07-20 哈尔滨工业大学 Multi-step prediction method for performance parameters of airplane auxiliary power device based on RBM (radial basis function) optimization ELM (electric field model)
CN110456732A (en) * 2019-08-06 2019-11-15 武汉恒力华振科技有限公司 A kind of punching machine fault monitoring system with learning functionality
CN110456732B (en) * 2019-08-06 2021-09-28 武汉恒力华振科技有限公司 Punch press fault monitoring system with learning function
CN111044285A (en) * 2019-11-22 2020-04-21 军事科学院系统工程研究院军用标准研究中心 Method for diagnosing faults of mechanical equipment under complex conditions
CN111539381A (en) * 2020-05-18 2020-08-14 中车永济电机有限公司 Construction method of wind turbine bearing fault classification diagnosis model
CN111539381B (en) * 2020-05-18 2022-04-05 中车永济电机有限公司 Construction method of wind turbine bearing fault classification diagnosis model
CN111734772A (en) * 2020-06-16 2020-10-02 大连理工大学 Magnetorheological vibration suppression method in thin-wall part machining
CN114016998A (en) * 2020-07-16 2022-02-08 中国石油天然气集团有限公司 Prediction method and device for logging encountering block
CN111860429A (en) * 2020-07-30 2020-10-30 科大讯飞股份有限公司 Blast furnace tuyere abnormality detection method, device, electronic apparatus, and storage medium
CN111860429B (en) * 2020-07-30 2024-02-13 科大讯飞股份有限公司 Blast furnace tuyere abnormality detection method, device, electronic equipment and storage medium
CN112379659A (en) * 2020-11-12 2021-02-19 西安石油大学 Petroleum drilling machine fault prediction system
CN113469217A (en) * 2021-06-01 2021-10-01 桂林电子科技大学 Unmanned automobile navigation sensor abnormity detection method based on deep learning
CN114004374A (en) * 2021-10-22 2022-02-01 南京信息工程大学 Air conditioning unit sensor fault diagnosis method based on Boltzmann machine voting

Similar Documents

Publication Publication Date Title
CN104616033A (en) Fault diagnosis method for rolling bearing based on deep learning and SVM (Support Vector Machine)
Tao et al. An unsupervised fault diagnosis method for rolling bearing using STFT and generative neural networks
Zhang et al. Machinery fault diagnosis with imbalanced data using deep generative adversarial networks
Shao et al. Highly accurate machine fault diagnosis using deep transfer learning
Li et al. Multiscale local features learning based on BP neural network for rolling bearing intelligent fault diagnosis
Jia et al. Deep normalized convolutional neural network for imbalanced fault classification of machinery and its understanding via visualization
Grezmak et al. Interpretable convolutional neural network through layer-wise relevance propagation for machine fault diagnosis
Liu et al. Artificial intelligence for fault diagnosis of rotating machinery: A review
AlThobiani et al. An approach to fault diagnosis of reciprocating compressor valves using Teager–Kaiser energy operator and deep belief networks
CN106874957A (en) A kind of Fault Diagnosis of Roller Bearings
CN106886660A (en) EEMD Hilbert envelope spectrums are with Rolling Bearing Status recognition methods under the varying load that DBN is combined
CN105760839A (en) Bearing fault diagnosis method based on multi-feature manifold learning and support vector machine
CN206504869U (en) A kind of rolling bearing fault diagnosis device
CN108256556A (en) Wind-driven generator group wheel box method for diagnosing faults based on depth belief network
Chen et al. Machine fault classification using deep belief network
Thoppil et al. Deep learning algorithms for machinery health prognostics using time-series data: A review
CN113033309A (en) Fault diagnosis method based on signal downsampling and one-dimensional convolution neural network
CN110503025B (en) Analog circuit early fault diagnosis method based on semi-supervised cooperative training
Zilvan et al. Denoising convolutional variational autoencoders-based feature learning for automatic detection of plant diseases
Sadik et al. Autismnet: Recognition of autism spectrum disorder from facial expressions using mobilenet architecture
CN115392323A (en) Bearing fault monitoring method and system based on cloud edge cooperation
Tripathi et al. Multivariate time series classification with an attention-based multivariate convolutional neural network
Aljemely et al. Wise-local response convolutional neural network based on Naïve Bayes theorem for rotating machinery fault classification
Bækkegaard et al. Target classification using kinematic data and a recurrent neural network
Du et al. Fault diagnosis of plunger pump in truck crane based on relevance vector machine with particle swarm optimization algorithm

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20150513