[go: up one dir, main page]

CN109800875A - Chemical industry fault detection method based on particle group optimizing and noise reduction sparse coding machine - Google Patents

Chemical industry fault detection method based on particle group optimizing and noise reduction sparse coding machine Download PDF

Info

Publication number
CN109800875A
CN109800875A CN201910016558.1A CN201910016558A CN109800875A CN 109800875 A CN109800875 A CN 109800875A CN 201910016558 A CN201910016558 A CN 201910016558A CN 109800875 A CN109800875 A CN 109800875A
Authority
CN
China
Prior art keywords
training
layer
fault detection
model
matrix
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910016558.1A
Other languages
Chinese (zh)
Inventor
苏堪裂
李秀喜
旷天亮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN201910016558.1A priority Critical patent/CN109800875A/en
Publication of CN109800875A publication Critical patent/CN109800875A/en
Pending legal-status Critical Current

Links

Landscapes

  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

本发明公开了一种基于粒子群优化和降噪稀疏编码机的化工故障检测方法,该方法主要对标准化和白化后的训练数据利用多个堆叠式的降噪稀疏自编码机进行无监督特征学习,再通过有监督的方式进行Softmax分类器模型训练,最后通过有监督微调整个网络的权重参数,同时对于关键的可调超参数引入了粒子群优化算法自动调优,得到训练好的化工过程故障检测智能模型从而用于过程实时数据的故障检测。本发明采用深度神经网络的贪婪逐层训练方法自适应地智能学习化工过程原始数据所隐含的知识,从而能够更准确地提取出过程故障的信息,所述方法较传统方法而言更具智能性,能够提高故障检测的性能,且由于加入自动优化算法,比人工参数调优节省了很多时间。

The invention discloses a chemical fault detection method based on particle swarm optimization and noise reduction sparse coding machine. The method mainly uses a plurality of stacked noise reduction sparse automatic coding machines to perform unsupervised feature learning on the training data after normalization and whitening. , and then train the Softmax classifier model in a supervised way, and finally fine-tune the weight parameters of the entire network through supervision. At the same time, the particle swarm optimization algorithm is introduced to automatically tune the key adjustable hyperparameters, and the trained chemical process faults are obtained. Detect smart models for fault detection of process real-time data. The invention adopts the greedy layer-by-layer training method of the deep neural network to adaptively and intelligently learn the knowledge implicit in the original data of the chemical process, so that the information of the process failure can be extracted more accurately, and the method is more intelligent than the traditional method. It can improve the performance of fault detection, and because of adding automatic optimization algorithm, it saves a lot of time than manual parameter tuning.

Description

Chemical fault detection method based on particle swarm optimization and noise reduction sparse coding machine
Technical Field
The invention relates to the field of chemical process fault detection and diagnosis, in particular to a chemical fault detection method based on particle swarm optimization and a noise reduction sparse coding machine.
Background
The chemical process fault detection is one of the most powerful tools for managing abnormal working conditions of the chemical process, provides certain guarantee for process safety early warning, and saves great economic loss. According to the estimation of the national security administration, the abnormal working condition causes economic loss of at least about $ 200 million per year to the American petroleum and chemical industry, and in the United kingdom, the loss caused by the abnormal working condition is as high as $ 270 million per year, so that the development of a chemical process fault detection method with better performance is crucial to the actual chemical production.
The chemical process data has the characteristics of nonlinearity, high dimension, non-Gaussian distribution and the like, so that the extraction of fault information in the fault detection process is more complicated, a plurality of traditional chemical process fault detection methods based on the process data are developed, the data-based methods do not need to acquire a large amount of expert knowledge in advance, the conditions of the current system can be predicted only by acquiring data acquired in the chemical process and establishing a proper fault detection model, and the methods are wide in scientific research and industrial application at present. Although the conventional methods such as PCA, ICA, KPCA, KICA, MICA, etc. which are popular at present can effectively detect some faults, the detection rate of some disturbing faults is very low, which indicates that the conventional methods still cannot completely and accurately extract the information of the faults, so that new methods need to be developed to improve the detection rate of the faults. The characteristics of the complex chemical process such as nonlinearity, high noise, non-Gaussian distribution and the like enable the traditional chemical process fault detection method not to show excellent diagnostic performance, so that the development of a fault monitoring method suitable for the complex nonlinear chemical process is very necessary. And the characteristics implied by the original data of the chemical process can be more accurately learned by the aid of a layer-by-layer greedy learning mode of the deep neural network, so that the chemical process can be monitored by applying a proper classification model.
With the development of deep learning, many students in the computer field are turning to research deep neural networks, and the deep neural networks and shallow neural networks are mostly distinguished in that the deep neural networks are trained layer by layer in an unsupervised learning manner, so that the deep neural networks have excellent feature learning ability, and the learned features are more essential in describing data, thereby being beneficial to classification. Research in the direction of an automatic coding machine is a popular research direction in the field of deep learning in recent years, and is widely applied to handwriting recognition, image classification, audio feature extraction, and the like because of its excellent performance in feature learning, and is also gradually applied to other fields, for example, in mechanical failure diagnosis, a trainee starts to apply an automatic coding machine or a modified automatic coding machine to perform feature learning. In 2006, Hiton et al, a famous deep learning learner, proposed a deep neural network algorithm of an Automatic Encoder (AE), and applied to image recognition to obtain a good classification effect, and then many scholars of automatic encoders based on this basis proposed improved algorithms, for example, Bengio et al, proposed a Sparse Automatic Encoder (SAE), added with sparsity restriction, so that the learned features can more accurately restore original data; vincent et al proposed a noise reduction Auto Encoder (DAE) in 2008, and added a noise reduction strategy in unsupervised feature learning, so that the learned features are more robust. The stack Denoising Sparse automatic coding machine (SDSA) is an improved automatic coding machine which is researched more enthusiastically in recent three years, and the stack Denoising Sparse automatic coding machine means that a plurality of Denoising Sparse automatic coding machines are Stacked to perform layer-by-layer training and layer-by-layer learning. The algorithm takes a noise reduction sparse automatic coding machine as a base body, each layer of weight parameters and bias items of the algorithm are trained layer by layer in an unsupervised mode by stacking a plurality of similar noise reduction sparse automatic coding machines, so that a deep neural network is built, the learned characteristics can deal with high-noise data and the learned characteristics have sparseness, and the training method is carried out in an unsupervised mode.
Because the method disclosed by the invention relates to more manually adjustable hyper-parameters and the method for selecting the hyper-parameters by adopting a manual test has randomness and time consumption, single-target discrete optimization is carried out on multiple parameters to realize the automatic adjusting function of the method, so that the method is more intelligent and achieves better performance. Common parameter optimization algorithms include genetic algorithms, artificial fish swarm algorithms, particle swarm algorithms and the like. The Particle Swarm Optimization (PSO) algorithm is a bionic algorithm which is proposed by Kennedy et al in 1995 and simulates foraging behavior of a bird flock, and the algorithm is widely applied in the fields of science and engineering due to the advantages of high convergence rate, simple implementation process, easy understanding of optimization process and the like, and becomes one of the fastest-developing parameter intelligent optimization algorithms.
Disclosure of Invention
The invention aims to provide a chemical fault detection method based on particle swarm optimization and a noise reduction sparse coding machine, aiming at the defects of the prior art. The method applies a stack type noise reduction sparse automatic coding system (SDSA) algorithm in deep learning to feature learning in a chemical process, conducts model training of a Softmax classifier in a supervision mode, finally fine-tunes weight parameters of the whole network through a BP algorithm, and introduces a particle swarm optimization algorithm to automatically tune key adjustable hyper-parameters. The method can adopt a greedy layer-by-layer training method of the deep neural network to adaptively and intelligently learn the implicit knowledge of the original data, so that the fault information is extracted.
The purpose of the invention can be realized by the following technical scheme:
a chemical industry fault detection method based on particle swarm optimization and a noise reduction sparse coding machine comprises the following steps:
step one, data acquisition:
taking historical time sequence data acquired by a simulation system or a DCS (distributed control System) as a training sample set XtrainChemical process real-time data from DCS system as test sample set XtestWherein a training sample set X is collectedtrainThe method comprises time sequence data under normal working conditions and various fault working conditions, is used for establishing an intelligent fault detection model of the method, and tests a sample set XtestThe method is real-time working condition data monitored on line, also comprises time sequence data under normal working conditions and various faults, and is used for verifying the diagnosis precision of the method or realizing fault detection by applying a model established by the method in actual industry;
step two, data preprocessing:
firstly, a training sample set X is calculatedtrainMean value X of monitoring variables of data under normal working conditionmeanAnd standard deviation XstdThen training sample set XtrainAnd test sample set XtestAll using the mean value XmeanAnd standard deviation XstdCarrying out standardization preprocessing, and training sample set X after standardization preprocessingtrainstdAnd test sample set XteststdThen carrying out whitening pretreatment to obtain a whitened training sample set XtrainwhiteAnd test sample set XtestwhiteCompleting the pretreatment of the training sample and the test sample set;
step three, off-line training:
the off-line training aims at establishing an intelligent fault detection model in the chemical process by adopting a preprocessed training sample, and the process can be divided into four major parts, namely an unsupervised pre-training stack type noise reduction sparse automatic coding machine, a supervised pre-training Softmax classifier, a BP algorithm global fine-tuning network parameter and a particle swarm optimization adjustable super parameter; firstly, carrying out unsupervised pre-training stack type noise reduction sparse automatic coding machine, adopting N noise reduction sparse coding machines to code a sample set into a feature space layer by layer for a training set after preprocessing, adopting a layer-by-layer training method, optimizing the loss function minimization of each layer in the training process of each layer so as to obtain model parameters of each layer, and finally obtaining the final feature h of the training sample learned by N hidden layersN(ii) a Secondly, carrying out supervised pre-training on a Softmax classifier, and dividing h into hNAs the input of the Softmax classifier model, adding corresponding working condition labels to all the learned training sample characteristicsyi1 indicates that the sample is normal, yi2 represents that the sample is a fault, and the cost function of Softmax is optimized to obtain the pre-trained Softmax model parameters; then carrying out BP algorithm global microAdjusting network parameters, carrying out supervised fine adjustment on all model parameters in the whole training process, taking SDSA (software development architecture) model parameters subjected to unsupervised pre-training and Softmax model parameters subjected to supervised pre-training as initial values, taking a pre-processed training sample as an input layer, firstly, carrying out feature learning layer by layer through the SDSA parameters to obtain a final feature matrix of a hidden layer, calculating a loss function value of the feature matrix through a Softmax classifier, and then optimizing global parameters by using a feedback propagation (BP) algorithm to minimize and converge the loss function, thereby obtaining a preliminarily trained fault detection model; finally, performing particle swarm optimization adjustable hyper-parameters, wherein because the artificial adjustable hyper-parameters in the whole intelligent fault detection model are not optimal, if the performance of the SDSA is not optimal due to improper setting of the hyper-parameters, a Particle Swarm Optimization (PSO) algorithm is called to perform hyper-parameter optimization, and finally, the values of the optimized hyper-parameters and all model parameters corresponding to the optimized hyper-parameters are determined, so that a finally trained fault detection model is obtained, and the model can be used for online monitoring of the process state;
step four, online monitoring:
for the chemical process of real-time continuous production, whether the current process is in a fault state can be effectively predicted by using the fault detection model trained in the third step; with pretreated test specimen XtestwhiteAnd as an input layer, obtaining finally learned test characteristics for the determined neural network structure by adopting a layer-by-layer characteristic learning method according to a trained fault detection model, calculating the membership probability value of a Softmax prediction function according to the learned final characteristics, indicating that the working condition is normal when the obtained membership class is 1, indicating that an abnormal working condition occurs when the membership class is 2, immediately sending an early warning by a connected warning device, indicating fault detection, and simultaneously notifying a technician or engineer to check the system safety and remove faults in time, thereby realizing fault monitoring on the current working condition.
Further, in the second step, the training sample set X is realized through the following stepstrainAnd test sample set XtestThe normalization pretreatment:
(1) training sample set XtrainIs a matrix of n multiplied by m, wherein n is the number of samples and m is the number of observation variables, and the training sample set X after the standardization processing is solved through the following formulatrainstdAnd test sample set Xteststd
Wherein, XijRepresenting a training sample set XtrainAnd test sample set XtestValue of the jth variable of the ith sample, Xmean,jFor training sample XtrainMean value of j variable of data under normal working condition, Xstd,jTraining sample XtrainAnd standard deviation of the jth variable of the data under the medium normal working condition.
Further, in the second step, the training sample set X is realized through the following stepstrainAnd test sample set XtestNormalization and whitening pre-processing:
(1) the training sample set X after the standardization treatmenttrainstdAnd test sample set XteststdPerforming whitening pretreatment, performing eigenvalue decomposition on the covariance matrix of the training sample by the following formula to obtain an orthogonal matrix of the eigenvector of the covariance matrix and a diagonal matrix of the eigenvalue of the orthogonal matrix, thereby obtaining a whitening matrix Wwhite
Cov=VDVT(2)
Wwhite=VD-1/2VT(3)
Wherein Cov is a training sample set X after the standardization processingtrainstdV is an orthogonal matrix of eigenvectors of the covariance matrix, D is a diagonal matrix of eigenvalues of the covariance matrix;
(2) then whiten the processed training sample XtrainwhiteAnd test sample XtestwhiteAre all WwhiteCalculating to obtain:
wherein, WwhiteFor whitening the preprocessed whitening matrix, XtrainwhiteIs a whitened training sample set, XtestwhiteIs the whitened set of test samples.
Further, in step three, the unsupervised pre-training stacked noise reduction sparse automatic coding machine in the offline training is specifically performed by the following steps:
(1) for the training sample X after the pretreatmenttrainwhiteSetting a deep neural network structure of a matrix of nxm, wherein N is the number of samples, m is the number of observation variables, and the number of hidden layers of the SDSA is assumed to be N, namely, the number HL of nodes of each hidden layer network is mainly determined1——HLNThe deep neural network has the following layers: an input layer (preprocessed training samples), N hidden layers (learning N layers of features), and a classification layer (output prediction probability), so that the global network can be regarded as a neural network with N +2 layers; firstly, initializing the weight matrix parameter of the first hidden layer, and dividing XtrainwhiteTraining a noise reduction sparse automatic coding (DSA) machine with a first hidden layer as an input layer of the SDSA1) Training sample X bytrainwhiteAdding Gaussian noise which is partially obeyed to normal distribution, and changing the training set into a data set Xc mixed with noise:
Xc=X+le*G (5)
wherein X is a training set matrix when no noise is added, and X is a training set X after preprocessing when a first hidden layer is formedtrainwhite(ii) a le is the noise level and can be set artificially (the value is 0-1, and generally 0.1); g is the generated Gaussian noise with the same dimension as the training set matrix when the noise is not added, and Xc is a data set containing artificial noise;
(2) after the interference noise is added, using Xc as input to carry out coding learning and decoding reconstruction, wherein the coding stage is the stage of characteristic learning essentially, and the decoding is the reconstruction of characteristic data, and the coding formula and the decoding formula are as follows:
wherein h is the characteristic of learning, Y is the information reconstructed by the characteristic h, and the closer the value of the H is to the original data X without noise, the better the model parameter training is; w is a coding weight parameter matrix, b1,b2As an offset vector, WTIs a decoding weight parameter matrix, and the feature h of the first hidden layer can be learned by the formula1And a model parameter W of the first hidden layer1 11,b21
(3) However, it can be noted that these model parameters are not optimal, so the first hidden layer features learned cannot deeply express the information contained in the original data; the optimal capability of the model can be played only by defining a reasonable loss function and then continuously optimizing the loss function to reach the minimum value, wherein the loss function is as follows:
wherein L istotal(W,WT,b1,b2) Representing the loss function value of the noise-reduction sparse automatic coding machine of the current hidden layer, wherein n is the number of samples and Y is the number of samplesiFor the reconstructed information vector value of the ith sample, XiFor the raw data vector value when no noise is added to the ith sample,is the value of the ith row and jth column of the weighting parameter of the ith layer, lambdarThe regularization weight attenuation hyperparameter is used for adjusting the weight of the equation, sl represents the node number of the l layer of a single noise reduction automatic coding machine, s (l +1) represents the node number of the l +1 layer, β is the weight hyperparameter for controlling the sparsity penalty term and needs to be manually adjusted to a proper value, s2The number of nodes of the current hidden layer learning features, namely the number of features learned by a single sample;is relative entropy, representing the difference between the current average activity and the sparsity constraint; rho is a sparsity parameter, represents sparsity constraint and is a small value generally set by human, for example, rho is 0.1;is the average activation of the jth variable in the feature matrix h,representing the value of the jth variable of the ith sample in the feature matrix h of the current hidden layer;
the matlab toolbox minFunc can be called to minimize equation (8) during offline trainingA loss function, i.e. the optimized weight parameter W after the pre-training of the first hidden layer can be obtained1,And bias term b11,b21Then, using the trained model parameters for feature learning of the first hidden layer training sample;
(4) and the first three steps are equivalent to the completion of the off-line training of the model of the first hidden layer, and because the stacked noise reduction sparse automatic coding machine of the method comprises N hidden layers, the step (1-3) needs to be repeatedly called to train the model parameters of the multiple hidden layer network until the completion of the pre-training of the SDSA of the N hidden layers.
Further, the step (4) specifically includes:
first, a noise reduction sparse automatic coding (DSA) machine for the first hidden layer1) After training is completed, its training sample XtrainwhiteIs encoded as h by1
h1=f(W1X+b11) (11)
Wherein, W1,b11For a trained DSA1X is a training set matrix when no noise is added, and X is a training set X after preprocessing when a first hidden layer is formedtrainwhiteThe 2-N hidden layer represents the feature matrix learned by the previous layer; then the well-learned first hidden layer characteristic h is used1Sparse Autocoder (DSA) with noise reduction as second hidden layer2) Repeatedly calling the step (1-3) to train the DSA2Obtaining the model weight parameter and the bias term W of the second hidden layer2,b12,b22And the characteristics of the second hidden layer are coded into h through the above formula (11)2This process is repeated until the DSANAfter training is finished, the final characteristics h learned by N hidden layers are obtainedNWill pre-trainThe learned features are used for a subsequent classification layer, and model parameters of the Softmax classifier can be trained.
Further, in step three, the supervised pretraining Softmax classifier in the offline training is specifically performed by the following steps:
(1) the Nth hidden layer characteristic h learned after the SDSA containing N hidden layers is trainedNAs input to the Softmax classifier model, corresponding condition labels are then appendedyi1 means that the ith sample is normal, yi2 represents that the ith sample is a fault; firstly, initializing a model parameter matrix theta, and predicting probability values of samples belonging to various classes according to the following prediction functions of a Softmax classifier:
wherein,for the probability values that the ith sample belongs to each class, θ is the model parameter matrix of the Softmax classifier, consisting ofVector composition; k is the number of classes defined by the classifier, where k is 2;the feature vector of the Nth hidden layer of the ith sample is obtained;
(2) because various membership probability values predicted by the prediction function are inaccurate, a classification loss function needs to be constructed so as to obtain optimized model parameters, and the adopted Softmax classifier takes the regularization item into consideration so as to effectively avoid the overfitting of the classification model training and define the loss function as follows:
wherein, the meaning of the 1{ } part is that if an index function in brackets is true, 1 is returned, otherwise, 0 is returned; lambda [ alpha ]smIs the weighted decay term coefficient of Softmax, λsm>0, is an artificially adjustable hyper-parameter; for the loss function, the matlab toolbox minFunc can be called to optimize the loss function minimum, so that the optimized parameter matrix theta of the Softmax classifier model can be obtained.
Further, in step three, the global fine-tuning network parameters of the BP algorithm in the offline training are specifically performed by the following steps:
with all model parameters pre-trained { (W)1,b11),(W2,b12),…,(WN,b1N) Theta as initial value to pre-process the training sample XtrainwhiteFor an input layer, firstly, performing layer-by-layer feature learning by using an equation (11) through initial parameters of a model of the SDSA to obtain a final hidden layer feature matrix, calculating a loss value of the equation (13) through a classification layer by using the feature matrix, then optimizing global parameters by using a feedback propagation algorithm, updating all model parameters once per iteration, and thus converging the loss value of the equation (13) to the minimum, and finishing a fine tuning process.
Further, in step three, the particle swarm optimization adjustable hyper-parameter in the offline training is specifically performed through the following steps:
(1) the parameters of the method can be divided into two categories of model parameters and adjustable parameters, and the method optimizes three key adjustable parameters which are respectively: regularization weight attenuation hyperparameter lambdarWeight hyperparameter β for controlling sparsity penalty term, weight attenuation term coefficient lambda of Softmaxsm(ii) a The three parameters form a PSO optimized single particle (lambda)r,β,λsm) Thus the dimensions of each particle are 3-dimensional; statorYi NpThe method comprises the following steps of carrying out simultaneous optimization on each particle, setting K iterations, initializing the initial position and the speed of each particle, taking a fitness function value as the overall Accuracy of a training sample, taking the three key adjustable hyper-parameters as independent variables, training the fine-tuned fault detection model to obtain the overall Accuracy of the training sample, wherein the formula is defined as the following formula:
wherein p isiRepresenting the category, y, of the ith training sample predicted by the fault detection modeliRepresenting the category of the ith training sample in the artificially given label, if the category of the ith training sample is equal to the category of the ith training sample, returning to be 1, and if the category of the ith training sample is not equal to the category of the ith training sample, indicating that the current sample is diagnosed by the model incorrectly, and returning to be 0; calculating the optimal fitness and position of each particle and the global optimal fitness and position of the particle swarm through the formula (14);
(2) since the global optimal fitness and the optimal position after the one iteration are not the most accurate, the speed and the position of each particle need to be updated, and the position of the particle under the t-th iteration can be expressed asThe flight velocity of each particle at the t-th iteration can be expressed asThe PSO algorithm updates the velocity and position of each particle by the following equation:
wherein WpIs coefficient of inertia, C1Tracking an acceleration coefficient of a self historical optimal value for the particles, and expressing the cognition of the particles to the particles; c2C is generally set for tracking the acceleration coefficient of the optimal value of the group for the particle, representing the cognition of the particle on the group knowledge, namely social knowledge1=C22, t is the current iteration number, ξ and η are at 0,1]Uniformly distributed random numbers within the interval;the position of the optimal fitness experienced by the particle i by the t-th iteration; gbtThe best position that the particle swarm has experienced by the time of the tth iteration;
(3) retraining the overall fine-tuning network parameters of the unsupervised pretrained SDSA, the supervised pretrained Softmax classifier and the BP algorithm of the fault detection model to obtain a new overall optimal fitness value and position, continuously updating the position and speed of the particles by adopting a formula (15-16), and stopping optimizing when the iteration times exceed K iterations defined by the formula, so as to obtain the positions of the particles under the overall optimal fitness, namely under the optimized adjustable hyper-parameter, the detection accuracy of the training samples is highest, thereby completing the automatic optimization hyper-parameter; and setting the optimized adjustable hyper-parameter value as the optimized adjustable hyper-parameter of the SDSA, and re-training all model parameters of the SDSA under the determined network structure and the optimized hyper-parameter, thereby obtaining a trained fault detection model, wherein the model parameters can be used for feature learning and classification of a test sample.
Further, in step four, the online monitored feature learning and membership probability prediction are performed by the following steps:
for the real-time chemical process of continuous production, the fault detection model trained in the third step is utilized to effectively predict whether the current process is in a fault state or not so as to preprocess the test sample XtestwhiteAs an input layer, adopting the determined neural network structure according to the trained fault detection modelUsing a layer-by-layer feature learning method, and obtaining weight parameters (W) of each trained noise reduction sparse automatic coding machine1,b11),(W2,b12),…,(WN,b1N) Using a formula (11) to perform feedforward learning to obtain the characteristics of the current hidden layer of a test sample, then using the learned characteristics of the current hidden layer as input, combining DSA model parameters of the next hidden layer to perform learning characteristics by using the same formula, learning all hidden layers by analogy to obtain finally learned characteristics, then using a formula (12) to perform prediction of each membership class, wherein the classes of the single sample only have two types, namely a fault state and a normal state, so that the prediction probability vector of the single sample only has two values, selecting the prediction class with the maximum probability value as the predicted membership class of the whole fault detection model, when the calculated membership class is 1, indicating that the working condition belongs to the normal state, when the membership class is 2, indicating that an abnormal working condition appears, immediately sending an early warning by a connected warning device, indicating fault detection, and simultaneously notifying a technician or engineer to check the safety of the system and eliminate the fault in time, therefore, fault monitoring can be carried out on the current working condition.
Compared with the prior art, the invention has the following advantages and beneficial effects:
1. the invention adopts a stack type noise reduction sparse self-encoder algorithm in a deep learning algorithm and a particle swarm parameter optimization algorithm to develop a novel fault monitoring method suitable for a complex nonlinear chemical process, does not need to mark data during feature learning, adopts a deep neural network, can realize self-adaptive intelligent learning of the features of original data compared with the traditional shallow neural network method and an artificial neural network, saves a great deal of energy and time compared with manual feature and knowledge extraction, and can more deeply distinguish normal data from abnormal data by the learned features, so that the clustering performance of samples of various categories is more obvious, and the algorithm is more intelligent.
2. The method avoids the randomness and the time consumption of manual test parameter selection on the basis of manual adjustable hyper-parameter selection, adopts the technical scheme of automatic parameter adjustment, and performs key adjustable hyper-parameter automatic adjustment by fusing a particle swarm optimization algorithm in the method, thereby saving a large amount of time and enabling the method to be more automatic.
3. Compared with the traditional chemical process fault detection technology (such as PCA, ICA, KPCA, MICA and the like), the method adopts a layer-by-layer training method of a deep neural network, can adaptively learn the information of nonlinear process data during characteristic learning, and simultaneously adds a noise reduction strategy, so that the robustness of an algorithm to high-noise data is enhanced, and the established fault detection model is more accurate.
Drawings
Fig. 1 is a flowchart of a chemical fault detection method based on particle swarm optimization and a noise reduction sparse coding machine according to an embodiment of the present invention.
FIG. 2 is a process flow diagram of a Tennessee-Iseman (TE) chemical process employed in an embodiment of the present invention.
Fig. 3 is a graph of the average failure detection rate and the false alarm rate of ten repeated tests according to the embodiment of the present invention.
Fig. 4 is a comparison graph of the fault detection rates of the methods with a large detection rate improvement according to the embodiment of the present invention.
Fig. 5(a), 5(b), 5(c), 5(d), 5(e), 5(f), 5(g), 5(h), and 5(i) are graphs of the results of fault monitoring of fault 1, fault 2, fault 4, fault 6, fault 7, fault 8, fault 12, fault 13, and fault 17, respectively.
Fig. 6(a) is a comparison graph of the first three principal component analyses of the test sample after the pretreatment, and fig. 6(b) is a graph of the first three principal component analyses of the test feature learned by the PSO-SDSA method.
Detailed Description
The present invention will be described in further detail with reference to examples and drawings, but the present invention is not limited thereto.
Example (b):
the embodiment provides a Chemical fault detection method based on a particle swarm optimization and noise reduction sparse coding machine, a flow chart of the method is shown in fig. 1, and the method provided by the embodiment is applied to a Tennessee-Isemann (TE) benchmark Chemical process to further explain the method of the embodiment, wherein the TE process is a computer simulation of an actual Chemical process published by downloads and Vogel in 1993 in the SCI journal of Computers & Chemical Engineering, and the process is mainly developed into a performance evaluation process monitoring method, and a process flow chart of the process is shown in fig. 2. The TE process mainly includes 5 operation units, namely: the device comprises a reactor, a condenser, a vapor-liquid separator, a circulating compressor and a stripping tower. In the simulated data, a total of 41 observed variables were monitored, 22 continuous process variables, 19 compositional variables, respectively. The TE process further includes 21 preset faults, and the embodiment uses the first 20 faults for monitoring, and the 20 preset faults are shown in table 1 below.
TABLE 1 TE Process 20 Preset failures
The chemical process fault detection method comprises the following steps:
step one, data acquisition:
collecting data under normal working conditions and 20 faults in the TE process, and dividing the data into a training sample set XtrainAnd test sample set Xtest. The training sample set comprises 13480 normal working condition samples and 480 samples under each fault. The test sample set contains 960 normal samples, 960 fault samples each, but the fault samples are all in a fault state starting at the 161 st sample. The process monitored 41 variables, so the training sample set formed an 23080 × 41 matrix and the test sample set formed an 20160 × 41 matrix.
Step two, data preprocessing:
firstly, the mean value X of each variable is obtained from 13480 normal samples in training datameanAnd standard deviation Xstd. Then using the mean value XmeanAnd standard deviation XstdAdopting formula (1) to carry out standardization processing on the training sample set and the test sample set, and then utilizing formula (2-3) to obtain whitening matrix W of the training sample setwhiteObtaining the whitened training sample set X by using the formula (4-5)trainwhiteAnd test sample set Xtestwhite
Step three, off-line training:
firstly, defining the hidden layer number to be 2, the network structure to be 41-130-20-2, initializing the weight parameter matrix and the bias item of the SDSA, and using the whitened training sample set XtrainwhiteAs an input layer of the SDSA, performing greedy training layer by layer on 2 hidden layers, optimizing a loss function of the SDSA by using a matlab toolbox minFunc to obtain optimized model parameters, and performing feedforward learning by adopting a formula (11) to obtain a characteristic h of a training sample of the SDSA2. Then the training sample characteristics h2The model is input as the input of a Softmax classifier model, and all learned training sample characteristics are added with corresponding working condition labelsyi1 indicates that the sample is normal, yi2 stands forThe sample is a fault, and a loss function of Softmax is optimized through a matlab toolbox minFunc, so that a pre-trained Softmax model parameter is obtained; and then carrying out global fine tuning on network parameters by using a BP algorithm, carrying out supervised fine tuning on all model parameters in the whole training process, taking the SDSA model parameters subjected to unsupervised pre-training and the Softmax model parameters subjected to supervised pre-training as initial values, taking the preprocessed training samples as input layers, and optimizing the global parameters by using a feedback propagation (BP) algorithm so as to minimize and converge the loss function, thereby obtaining a preliminarily trained fault detection model. And finally, performing particle swarm optimization on three key adjustable hyper-parameters, and finally determining the value of the optimized hyper-parameter and all model parameters corresponding to the optimized hyper-parameter, thereby obtaining a finally trained fault detection model which can be used for on-line monitoring of the process state.
Step four, online monitoring:
for the real-time chemical process of continuous production, whether the current process is in a fault state can be effectively predicted by using the fault detection model trained in the third step. With pretreated test specimen XtestwhiteAs an input layer, according to a trained fault detection model, for a determined neural network structure, adopting a layer-by-layer feature learning method, and using the trained weight parameters { (W) of each noise reduction sparse automatic coding machine1,b11),(W2,b12),…,(WN,b1N) Performing feedforward learning by adopting a formula (11) to obtain the characteristics of the current hidden layer of the test sample, then taking the learned characteristics of the current hidden layer as input, learning the characteristics by adopting the same formula in combination with DSA model parameters of the next hidden layer, repeating the learning process to obtain the finally learned characteristics of all the hidden layers, and predicting each membership class by adopting a formula (12), wherein the classes of the single sample only have two types, namely a fault state and a normal state, so that the prediction probability vector of the single sample only has two values, the prediction class with the maximum probability value is selected as the predicted membership class of the whole fault detection model, when the calculated membership class is 1, the working condition is normal, and when the membership class is 1, the working condition is normalIf the number is 2, indicating that an abnormal working condition occurs, and finally counting the detection rate of each fault;
and the test sample set is regarded as real-time working condition data, and fault diagnosis can be performed on the real-time data acquired in the actual chemical process through the steps.
Firstly, ten repeated tests are carried out on the selected network structure of 41-130-20-2 and the optimized superparameter, and the optimal superparameter obtained by the TE process is as follows: regularization weight attenuation hyperparameter lambdar0.002, 2.628255 is the weight hyperparameter β controlling the sparsity penalty term, and λ is the weight attenuation factor of Softmaxsm0.0001, and 0.9998. The average failure detection rate and the average false alarm rate of the test obtained in each test are shown in fig. 3, and it can be seen from the graph that the average value of the Failure Detection Rate (FDR) of the test sample of the ten-time average test is 82.42%, the standard deviation is 0.857%, the average value of the False Alarm Rate (FAR) is 0.64%, and the standard deviation is 1.194%, which indicates that the overall FDR of the method is higher and the FAR is lower, and the diagnostic performance of the method is better according to the method for evaluating the failure detection performance. In addition, the difference between the FDR and the FAR of the ten tests is not large, which shows that the method has the effect of improving the stability of the model by fusing the PSO optimization, and the training is difficult and the model is trapped in local optimization due to improper parameters under the condition that the model is not optimized, so that the PSO optimization is selected, the problems can be avoided, the model can achieve better diagnostic performance under the optimal parameters, and meanwhile, the PSO optimization is used as a more mature unconstrained discrete optimization method, has the advantages of high convergence speed, easiness in programming realization and the like, and in conclusion, the PSO parameter optimization is necessary and is easy to realize.
The result (9 th test) with the best performance (highest fault detection rate and lowest false alarm rate) of the method is selected to show the detection rate of each fault of the test sample and compared with other methods as shown in table 2. As can be seen from table 2, the proposed PSO-SDSA method has a high diagnostic accuracy for faults 1,2, 4,6,7,8, 10, 11, 12, 13, 14, 16, 17, 18, 19, 20. It can be seen that comparing the PCA method, the MICA method and the KPCA method, the proposed method of mean failure detection rate (PSO-SDSA) is the best, with an average FDR of 83.48% for all failures and an FAR of only 0.21%. In the method developed in the past, faults 3, 9 and 15 are faults which are difficult to detect in the TE process, the difference between the faults and normal sample data is not large, so that the very low diagnosis rate (mostly lower than 10%) occurs when a plurality of methods detect the faults, the difference between the deep neural network and the modeling of the traditional method is that the modeling data adopted is a large number of normal samples and fault samples, in addition, the method can deeply train and add noise modeling layer by layer due to greedy, the loss of information is strictly limited, and the learned fault characteristics can deeply mine the difference between the micro-disturbance faults and the normal faults, so that the fault characteristics are better distinguished from the normal characteristics. Therefore, the fault detection rates of faults 3, 9 and 15 are greatly improved in the proposed method, wherein the fault 3 reaches 50.5%, the fault 5 reaches 44.755%, the fault 15 reaches 44.875%, and other methods are basically maintained below 10%, so that the proposed PSO-SDSA method effectively improves the performance of certain fault points which are difficult to detect to a certain extent. Although the PSO-SDSA method is somewhat lower than other methods at some faults (fault 1, 14), the method is generally modeled under consideration of global errors, and since the global faults are sought to be smaller, the model may be biased towards reducing the errors of the harder-to-detect faults, while somewhat ignoring some of the easier-to-detect fault samples, which may result in some of the easier-to-detect faults FDR being somewhat lower than the conventional method instead, and the resulting fault detection rate (fault 1 is 96.5%, fault 14 is 90.375%) is quite acceptable. In general, compared with the other methods, the method has greatly improved performance, and is an FDD method which is more advantageous in the current research. In addition, fig. 4 shows a comparison graph (faults 3,5,9,10,15,19,20) of the fault detection rate, where FDR is significantly improved compared with other methods, so that the superiority of the performance improvement of the method can be more intuitively shown, and it can be seen that each point of the PSO-SDSA curve is on the other method curves, which shows that the detection rates of the 7 given faults are greatly improved.
TABLE 2 Fault detectivity for various methods of TE Process
And (3) carrying out early warning speed analysis of online monitoring on the test result, selecting 9 faults with higher detectable rate to carry out detection speed analysis (faults 1,2, 4,6,7,8,12, 13 and 17), respectively drawing each fault sample and the prediction probability in the figure 5 according to the sequence, wherein a 0.5 warning line in the figure is a classification control limit, and when the sample prediction probability value exceeds the control limit, the model is identified as a fault state. As can be seen from fig. 5, for the selected 9 faults, most of the samples after the fault is detected can be detected, thus resulting in a high fault detection rate (the FDR of the faults is higher than 90%), and faults 1,4,6,7,8,12 are detected at the 161 st sample, delayed by 0, which indicates that the detection speed of the faults is extremely high. And the detection points of faults 2 and 17 are at the 162 th sample, although 1 point is delayed, the detection speed is still high, and the early warning effect is completely acceptable in practical application. Fault 13 was detected at a relatively slower rate, and at sample 167 a fault was detected, delaying the onset of alarm by 6 points. In addition, as can be seen from the given online fault monitoring results, after the faults are detected by the faults 2,4,6,7 and 12, the number of points at which the samples are lower than the control limit is very small, so that the fault detection rates are all over 99%, which indicates that the method has strong capturing capability for detecting the faults. The faults 1,8,13 and 17 are occasionally stepped below the control limit after the faults are detected, so that the detection rates of the faults are lower, but the detection rates are all over 94%, and therefore the detection performance is better.
In order to further research the reason why the fault detection rate and the false alarm rate of the method are good, principal component analysis is carried out on the characteristics learned by the method. We selected the first 3 PCs of the data of normal state and faults 1,2,6,7 for comparison, fig. 6(a) is a comparison graph of the first three principal component analyses of the test sample after preprocessing, and fig. 6(b) is a graph of the first three principal component analyses of the test features learned by the PSO-SDSA method. As can be seen from a comparison between fig. 6(a) and fig. 6(b), in the test data after the preprocessing, the data cross-over overlap of faults 1,2,7 and normal samples is serious, and the linear inseparable characteristic is serious, the TE process has the characteristic of complex nonlinearity, correct classification cannot be realized by directly feeding the TE process into a classification model, and the TE process can be known through characteristic principal component analysis after being learned by a PSO-SDSA method, the faults 1,2,6 and 7 have better distinguishability, and the sample aggregation of each fault is also better, meanwhile, the clustering performance between normal samples is good, and the normal samples are not very dispersive, so that the faults can be completely distinguished from the normal samples, the fault diagnosis rate is high, and the false alarm rate is very low. Under the characteristic learning of the stack type noise reduction sparse self-coding machine, fault information and normal information can be deeply distinguished, so that the learned characteristics have excellent clustering performance, and the fault detection performance is improved.
The above description is only for the preferred embodiments of the present invention, but the protection scope of the present invention is not limited thereto, and any person skilled in the art can substitute or change the technical solution of the present invention and the inventive concept within the scope of the present invention, which is disclosed by the present invention, and the equivalent or change thereof belongs to the protection scope of the present invention.

Claims (9)

1.一种基于粒子群优化和降噪稀疏编码机的化工故障检测方法,其特征在于,所述方法包括以下步骤:1. a chemical fault detection method based on particle swarm optimization and noise reduction sparse coding machine, is characterized in that, described method comprises the following steps: 步骤一、数据采集:Step 1. Data collection: 对于模拟系统或DCS系统采集到的历史时序数据作为训练样本集Xtrain,来自DCS系统的化工过程实时数据作为测试样本集Xtest,其中采集到的训练样本集Xtrain包含正常工况下和各种故障工况下的时序数据,用于建立本方法的智能故障检测模型,测试样本集Xtest是在线监测的实时工况数据,也包含正常工况下和各种故障下的时序数据,用于验证本方法的诊断精度或实际工业中应用本方法建立的模型来实现故障检测;The historical time series data collected by the simulation system or DCS system is used as the training sample set X train , and the real-time chemical process data from the DCS system is used as the test sample set X test , wherein the collected training sample set X train includes The time series data under various fault conditions is used to establish the intelligent fault detection model of this method. The test sample set X test is the real-time operating condition data monitored online, and also includes the time series data under normal conditions and various faults. To verify the diagnostic accuracy of the method or apply the model established by the method in the actual industry to realize fault detection; 步骤二、数据预处理:Step 2: Data preprocessing: 首先计算出训练样本集Xtrain中正常工况下的数据的各监测变量的均值Xmean和标准差Xstd,然后将训练样本集Xtrain和测试样本集Xtest都利用该均值Xmean和标准差Xstd进行标准化预处理,标准化预处理后的训练样本集Xtrainstd和测试样本集Xteststd再进行白化预处理得到白化后的训练样本集Xtrainwhite和测试样本集Xtestwhite,至此完成训练样本和测试样本集的预处理;First, calculate the mean value X mean and standard deviation X std of each monitoring variable of the data in the training sample set X train under normal working conditions, and then use the mean value X mean and the standard The difference X std is subjected to standardization preprocessing, and the preprocessed training sample set X trainstd and test sample set X teststd are subjected to whitening preprocessing to obtain the whitened training sample set X trainwhite and test sample set X testwhite , so far the training samples and Preprocessing of the test sample set; 步骤三、离线训练:Step 3. Offline training: 采用预处理好的训练样本建立化工过程智能故障检测模型,其过程可分为无监督预训练栈式降噪稀疏自动编码机、有监督预训练Softmax分类器、BP算法全局微调网络参数、粒子群优化可调超参数四大部分,首先进行无监督预训练栈式降噪稀疏自动编码机,对预处理后的训练集采用N个降噪稀疏编码机将样本集逐层编码为特征空间,采用逐层训练的方法,在每一层训练过程中优化其各层的损失函数最小化,从而得到各层模型参数,最终得到其N个隐层学习到的训练样本的最终特征hN;其次进行有监督预训练Softmax分类器,将hN作为Softmax分类器模型的输入,将所有学习到的训练样本特征附加对应的工况标签yi=1代表该样本是正常的,yi=2代表该样本是故障,通过最优化Softmax的成本函数从而得到其预训练后的Softmax模型参数;然后进行BP算法全局微调网络参数,对整个训练过程的所有模型参数进行有监督地微调,以无监督预训练好的SDSA模型参数和有监督预训练Softmax模型参数作为初始值,以预处理好的训练样本为输入层,首先通过这些SDSA参数逐层特征学习,从而得到最终隐层的特征矩阵,将该特征矩阵通过Softmax分类器计算其损失函数值,然后使用反馈传播(BP)算法优化全局参数从而使得损失函数最小化收敛即可,从而得到初步训练好的故障检测模型;最后进行粒子群优化可调超参数,由于整个智能故障检测模型中的人工可调超参数并非最优的,如果这些超参数设置不当可能会导致SDSA的性能并非最优,因此将调用粒子群PSO优化算法进行超参数调优,最后确定出最优化超参数的值和其对应的全部模型参数,从而得到最终训练好的故障检测模型用于过程状态的在线监测;The pre-processed training samples are used to establish an intelligent fault detection model for chemical processes. The process can be divided into unsupervised pre-training stack noise reduction and sparse automatic encoder, supervised pre-training Softmax classifier, BP algorithm global fine-tuning network parameters, particle swarm The four major parts of the adjustable hyperparameters are optimized. First, the unsupervised pre-training stack-type noise reduction sparse auto-encoder is performed, and N noise reduction sparse encoders are used for the preprocessed training set to encode the sample set into feature space layer by layer. Using the method of layer-by-layer training, in the training process of each layer, the loss function of each layer is optimized to minimize the loss function of each layer, so as to obtain the model parameters of each layer, and finally obtain the final feature h N of the training samples learned by its N hidden layers; secondly Perform supervised pre-training of the Softmax classifier, use h N as the input of the Softmax classifier model, and attach all the learned training sample features to the corresponding working condition labels y i =1 represents that the sample is normal, y i =2 represents that the sample is faulty, and the pre-trained Softmax model parameters are obtained by optimizing the cost function of Softmax; then the BP algorithm is used to globally fine-tune the network parameters. All model parameters in the training process are fine-tuned in a supervised manner. The unsupervised pre-trained SDSA model parameters and the supervised pre-trained Softmax model parameters are used as initial values, and the pre-processed training samples are used as the input layer. First, pass these SDSA parameters. Layer-by-layer feature learning is used to obtain the feature matrix of the final hidden layer. The feature matrix is used to calculate its loss function value through the Softmax classifier, and then the feedback propagation (BP) algorithm is used to optimize the global parameters so that the loss function can be minimized and converged. Obtain the fault detection model that has been initially trained; finally, adjust the hyperparameters by particle swarm optimization. Since the artificially adjustable hyperparameters in the entire intelligent fault detection model are not optimal, if these hyperparameters are set improperly, the performance of SDSA may not be the same. Therefore, the particle swarm PSO optimization algorithm will be called for hyperparameter tuning, and finally the optimal hyperparameter value and all corresponding model parameters will be determined, so as to obtain the final trained fault detection model for online monitoring of process status. ; 步骤四、在线监控:Step 4. Online monitoring: 以预处理后的测试样本Xtestwhite作为输入层,根据训练好的故障检测模型,对确定的神经网络结构,采用逐层特征学习法得到最终学习到的测试特征,再通过所学习的最终特征计算其Softmax预测函数的隶属概率值,当得出的隶属类别为1,则说明该工况属于正常,当隶属类别为2,则说明有异常工况出现,随即连接的报警装置发出预警,指示故障检出,同时通知工艺员或工程师及时地检查系统安全并排除故障,从而可以实现对当前工况进行故障监控。Taking the preprocessed test sample X testwhite as the input layer, according to the trained fault detection model, for the determined neural network structure, the layer-by-layer feature learning method is used to obtain the final learned test features, and then calculate the final learned features. The membership probability value of its Softmax prediction function, when the obtained membership category is 1, it means that the working condition is normal; when the membership category is 2, it means that there is an abnormal working condition, and then the connected alarm device issues an early warning to indicate a fault. It is detected, and at the same time, the technician or engineer is notified to check the safety of the system and eliminate the fault in time, so as to realize the fault monitoring of the current working condition. 2.根据权利要求1所述的基于粒子群优化和降噪稀疏编码机的化工故障检测方法,其特征在于,步骤二中,通过以下步骤实现训练样本集Xtrain和测试样本集Xtest的标准化预处理:2. the chemical industry fault detection method based on particle swarm optimization and noise reduction sparse coding machine according to claim 1, is characterized in that, in step 2, realizes the standardization of training sample set X train and test sample set X test by following steps Preprocessing: 训练样本集Xtrain为n×m的矩阵,其中n为样本数,m为观测变量数,通过下式求解标准化处理后的训练样本集Xtrainstd和测试样本集XteststdThe training sample set X train is an n×m matrix, where n is the number of samples and m is the number of observed variables. The standardized training sample set X trainstd and the test sample set X teststd are solved by the following formula: 其中,Xij表示训练样本集Xtrain和测试样本集Xtest中第i个样本的第j个变量的值,Xmean,j为训练样本Xtrain中正常工况下数据的第j个变量的均值,Xstd,j训练样本Xtrain中正常工况下数据的第j个变量的标准差。Among them, X ij represents the value of the jth variable of the ith sample in the training sample set X train and the test sample set X test , and X mean, j is the value of the jth variable of the data in the training sample X train under normal conditions. Mean, X std, j The standard deviation of the jth variable of the data under normal conditions in the training sample X train . 3.根据权利要求2所述的基于粒子群优化和降噪稀疏编码机的化工故障检测方法,其特征在于,步骤二中,通过以下步骤实现训练样本集Xtrain和测试样本集Xtest的白化预处理:3. the chemical industry fault detection method based on particle swarm optimization and noise reduction sparse coding machine according to claim 2, is characterized in that, in step 2, realizes the whitening of training sample set X train and test sample set X test by following steps Preprocessing: (1)、将经过标准化处理后的训练样本集Xtrainstd和测试样本集Xteststd进行白化预处理,通过下式先对训练样本的协方差矩阵进行特征值分解,得到协方差矩阵的特征向量的正交矩阵和其特征值的对角矩阵,从而得到白化矩阵Wwhite(1) Perform whitening preprocessing on the standardized training sample set X trainstd and the test sample set X teststd , first perform eigenvalue decomposition on the covariance matrix of the training sample by the following formula, and obtain the eigenvector of the covariance matrix. The orthogonal matrix and the diagonal matrix of its eigenvalues, resulting in the whitening matrix W white : Cov=VDVT (2)Cov=VDV T (2) Wwhite=VD-1/2VT (3)W white = VD -1/2 V T (3) 其中,Cov是标准化处理后的训练样本集Xtrainstd的协方差矩阵,V是协方差矩阵的特征向量的正交矩阵,D是协方差阵的特征值的对角矩阵;Among them, Cov is the covariance matrix of the standardized training sample set X trainstd , V is the orthogonal matrix of the eigenvectors of the covariance matrix, and D is the diagonal matrix of the eigenvalues of the covariance matrix; (2)、然后白化处理后的训练样本Xtrainwhite和测试样本Xtestwhite均由Wwhite计算得出:(2), then the training sample X trainwhite and the test sample X testwhite after whitening are calculated by W white : 其中,Wwhite为白化预处理的白化矩阵,Xtrainwhite是白化处理后的训练样本集,Xtestwhite是白化处理后的测试样本集。Among them, W white is the whitening matrix of whitening preprocessing, X trainwhite is the training sample set after whitening processing, and X testwhite is the testing sample set after whitening processing. 4.根据权利要求3所述的基于粒子群优化和降噪稀疏编码机的化工故障检测方法,其特征在于,步骤三中,所述离线训练中的无监督预训练栈式降噪稀疏自动编码机具体通过以下步骤进行:4. the chemical industry fault detection method based on particle swarm optimization and noise reduction sparse coding machine according to claim 3, is characterized in that, in step 3, the unsupervised pre-training stack type noise reduction sparse automatic coding in described offline training The machine is carried out through the following steps: (1)、对于预处理后的训练样本Xtrainwhite为n×m的一个矩阵,n为样本个数,m为观测变量个数,假设其栈式降噪稀疏自动编码机(SDSA)的隐层数为N个,设置其深度神经网络结构,即主要确定各隐层网络的节点数HL1———HLN,其深度神经网络的各层主要定义为:输入层、N个隐层、分类层,因此其全局网络可以看成是具有N+2层的神经网络,首先初始化第一隐层的权重矩阵参数,将Xtrainwhite作为SDSA的输入层,训练出第一隐层的降噪稀疏自动编码机DSA1,通过下式对训练样本Xtrainwhite加入部分服从正态分布的高斯噪声,将训练集变成混有噪声的数据集Xc:(1) For the preprocessed training sample X trainwhite is a matrix of n × m, n is the number of samples, m is the number of observation variables, assuming the hidden layer of the stack noise reduction sparse automatic encoder (SDSA) The number is N, and its deep neural network structure is set, that is, the number of nodes in each hidden layer network is mainly determined HL 1 --- HL N , and each layer of the deep neural network is mainly defined as: input layer, N hidden layers, classification Therefore, its global network can be regarded as a neural network with N+2 layers. First, initialize the weight matrix parameters of the first hidden layer, use X trainwhite as the input layer of SDSA, and train the noise reduction and sparse automatic of the first hidden layer. The encoder DSA 1 , adds part of the Gaussian noise that obeys the normal distribution to the training sample X trainwhite by the following formula, and turns the training set into a data set Xc mixed with noise: Xc=X+le*G (5)Xc=X+le*G (5) 其中,X为未加入噪声时的训练集矩阵,第一隐层时X即为预处理后的训练集Xtrainwhite;le为噪声等级,其值取0-1,一般取0.1即可;G为产生的与未加入噪声时的训练集矩阵同维度的高斯噪声,Xc为含有人为噪声的数据集;Among them, X is the training set matrix when no noise is added, and X is the preprocessed training set X trainwhite in the first hidden layer; le is the noise level, and its value is 0-1, generally 0.1; G is The generated Gaussian noise with the same dimension as the training set matrix when no noise is added, Xc is the data set containing artificial noise; (2)、添加完干扰噪声后使用Xc作为输入进行编码学习和解码重构,编码阶段实质上是特征学习的阶段,解码则是对特征数据的重构,其编码公式和解码公式为:(2) After adding the interference noise, use Xc as the input to perform coding learning and decoding reconstruction. The coding stage is essentially the stage of feature learning, and the decoding is the reconstruction of the feature data. The coding formula and decoding formula are: 其中,h为学习的特征,Y为通过特征h重构出的信息,其值与未加入噪声的原始数据X越接近说明模型参数训练得越好;W是编码权重参数矩阵,b1,b2为偏置向量,WT是解码权重参数矩阵,通过上式可学习到第一隐层的特征h1和第一隐层的模型参数W1b11,b21Among them, h is the learned feature, Y is the information reconstructed by the feature h, the closer the value is to the original data X without noise, the better the model parameters are trained; W is the encoding weight parameter matrix, b 1 , b 2 is the bias vector, W T is the decoding weight parameter matrix, the feature h 1 of the first hidden layer and the model parameter W 1 of the first hidden layer can be learned through the above formula, b 11 , b 21 ; (3)、定义合理的损失函数,然后不断优化该损失函数达到最小值即能发挥模型的最优能力,其损失函数如下:(3) Define a reasonable loss function, and then continuously optimize the loss function to reach the minimum value, which can exert the optimal ability of the model. The loss function is as follows: 其中,Ltotal(W,WT,b1,b2)代表当前隐层的降噪稀疏自动编码机的损失函数值,n为样本个数,Yi为第i个样本的重构信息向量值,xi为第i个样本未加入噪声时的原始数据向量值,为第l层的权重参数的第i行第j列的值,λr是正则化权重衰减超参数,用于调节该方程的权重;sl表示单个降噪自动编码机的第l层的节点数,s(l+1)表示第l+1层的节点数,β是控制稀疏性惩罚项的权重超参数,需要人工调节至一个较合适的值;s2为当前隐层学习特征的节点数,即单个样本所学习到的特征的个数;是相对熵,代表当前平均激活度与稀疏性约束的差异;ρ为稀疏性参数,表示稀疏性约束,一般是人为设置的较小值,例如ρ取0.1;为特征矩阵h中第j个变量的平均激活度,代表当前隐层的特征矩阵h中第i个样本的第j个变量的值;Among them, L total (W, W T , b 1 , b 2 ) represents the loss function value of the noise reduction sparse auto-encoder of the current hidden layer, n is the number of samples, and Y i is the reconstructed information vector of the ith sample value, x i is the original data vector value of the ith sample without adding noise, is the value of the i-th row and the j-th column of the weight parameter of the l-th layer, λ r is the regularization weight decay hyperparameter used to adjust the weight of the equation; sl represents the number of nodes in the l-th layer of a single noise reduction auto-encoder , s(l+1) represents the number of nodes in the l+1th layer, β is the weight hyperparameter that controls the sparsity penalty term, and needs to be manually adjusted to a more appropriate value; s 2 is the number of nodes of the current hidden layer learning feature , that is, the number of features learned by a single sample; is the relative entropy, which represents the difference between the current average activation degree and the sparsity constraint; ρ is the sparsity parameter, which represents the sparsity constraint, which is generally a small value set artificially, for example, ρ takes 0.1; is the average activation of the jth variable in the feature matrix h, Represents the value of the jth variable of the ith sample in the feature matrix h of the current hidden layer; 在离线训练时能够调用matlab工具箱minFunc对方程(8)进行最小化损失函数,即可求出第一隐层预训练后最优化的权重参数W1和偏置项b11,b21,然后将训练好的模型参数用于第一隐层训练样本的特征学习;During offline training, the matlab toolbox minFunc can be called to minimize the loss function of equation (8), and the optimized weight parameter W 1 after the first hidden layer pre-training can be obtained, and bias terms b 11 , b 21 , and then use the trained model parameters for feature learning of the training samples of the first hidden layer; (4)、重复调用步骤(1-3)进行多隐层网络的模型参数训练,直至完成N个隐层的SDSA预训练。(4) Repeat calling steps (1-3) to train the model parameters of the multi-hidden layer network until the SDSA pre-training of N hidden layers is completed. 5.根据权利要求4所述的基于粒子群优化和降噪稀疏编码机的化工故障检测方法,其特征在于,所述步骤(4)具体包括:5. the chemical industry fault detection method based on particle swarm optimization and noise reduction sparse coding machine according to claim 4, is characterized in that, described step (4) specifically comprises: 首先对于上述第一隐层的降噪稀疏自动编码机DSA1训练完成后,其训练样本Xtrainwhite通过下式被编码为h1First, after the above-mentioned first hidden layer noise reduction sparse auto-encoder DSA 1 is trained, its training sample X trainwhite is encoded as h 1 by the following formula: h1=f(W1X+b11) (11)h 1 =f(W 1 X+b 11 ) (11) 其中,W1,b11为训练好的DSA1的模型参数,X为未加入噪声时的训练集矩阵,第一隐层时X即预处理后的训练集Xtrainwhite,第2-N隐层时则表示为前一层所学习到的特征矩阵;Among them, W 1 , b 11 are the model parameters of the trained DSA 1 , X is the training set matrix when no noise is added, X is the preprocessed training set X trainwhite in the first hidden layer, and the second-N hidden layer is represented as the feature matrix learned by the previous layer; 然后再使用学习好的第一隐层特征h1作为第二隐层降噪稀疏自动编码机DSA2的输入,重复调用步骤(1-3)去训练DSA2,得到第二隐层的模型权重参数和偏置项W2b12,b22,并将其第二隐层的特征通过上式(11)编码为h2,重复这个过程直至DSAN训练完毕,得到其N个隐层学习到的最终特征hN,将预训练所学习到的特征用于后续分类层,可以训练Softmax分类器的模型参数。Then use the learned first hidden layer feature h 1 as the input of the second hidden layer noise reduction sparse auto-encoder DSA 2 , and repeat steps (1-3) to train DSA 2 to obtain the model weight of the second hidden layer parameter and bias term W 2 , b 12 , b 22 , encode the feature of the second hidden layer as h 2 through the above formula (11), repeat this process until the DSA N training is completed, and obtain the final feature h N learned by its N hidden layers, The features learned by pre-training are used for the subsequent classification layer, and the model parameters of the Softmax classifier can be trained. 6.根据权利要求5所述的基于粒子群优化和降噪稀疏编码机的化工故障检测方法,其特征在于,步骤三中,所述离线训练中的有监督预训练Softmax分类器具体通过以下两个步骤进行:6. the chemical industry fault detection method based on particle swarm optimization and noise reduction sparse coding machine according to claim 5, is characterized in that, in step 3, the supervised pre-training Softmax classifier in described off-line training specifically passes following two steps to proceed: (1)、将含有N个隐层的SDSA训练后所学习到的第N隐层特征hN作为Softmax分类器模型的输入,然后附加对应的工况标签yi=1代表第i个样本是正常的,yi=2代表第i个样本是故障:首先初始化其模型参数矩阵θ,根据以下Softmax分类器预测函数预测其样本属于各类的概率值:(1) The Nth hidden layer feature hN learned after SDSA training with N hidden layers is used as the input of the Softmax classifier model, and then the corresponding working condition label is attached y i =1 represents that the ith sample is normal, and y i =2 represents that the ith sample is faulty: first initialize its model parameter matrix θ, and predict the probability value of its sample belonging to each category according to the following Softmax classifier prediction function: 其中,为第i个样本隶属于各类别的概率值,θ是Softmax分类器的模型参数矩阵,由向量构成;k为分类器所定义的类别数,这里k=2;为第i个样本的第N隐层的特征向量;in, is the probability value of the i-th sample belonging to each category, and θ is the model parameter matrix of the Softmax classifier, which is represented by Vector composition; k is the number of categories defined by the classifier, where k=2; is the feature vector of the Nth hidden layer of the ith sample; (2)、构建一个分类损失函数,从而得到最优化的模型参数,采用的Softmax分类器考虑了正则化项,从而可以有效地避免分类模型训练过拟合,定义损失函数如下式:(2) Construct a classification loss function to obtain the optimal model parameters. The Softmax classifier used considers the regularization term, which can effectively avoid overfitting of the classification model training. The loss function is defined as follows: 其中,1{.}部分的含义为括号内的指标函数为真则返回1,否则返回0;λsm是Softmax的权重衰减项系数,λsm>0,是人工可调的超参数;对于上述损失函数,调用matlab工具箱minFunc优化其损失函数最小,从而求出Softmax分类器模型的最优化参数矩阵θ。Among them, the meaning of the 1{.} part is that if the index function in the parentheses is true, it will return 1, otherwise it will return 0; Loss function, call the matlab toolbox minFunc to optimize the minimum loss function, so as to obtain the optimal parameter matrix θ of the Softmax classifier model. 7.根据权利要求6所述的基于粒子群优化和降噪稀疏编码机的化工故障检测方法,其特征在于,步骤三中,所述离线训练中的BP算法全局微调网络参数具体通过以下步骤进行:7. the chemical industry fault detection method based on particle swarm optimization and noise reduction sparse coding machine according to claim 6, is characterized in that, in step 3, the BP algorithm in described off-line training global fine-tuning network parameter is specifically carried out by following steps : 以预训练好的所有模型参数{(W1,b11),(W2,b12),...,(WN,b1N),θ}作为初始值,以预处理好的训练样本Xtrainwhite为输入层,首先通过SDSA的模型初始参数采用方程(11)进行逐层特征学习,从而得到最终隐层的特征矩阵,将该特征矩阵通过分类层计算方程(13)的损失值,然后使用反馈传播算法优化全局参数,每次迭代则更新一次所有模型参数,从而使得方程(13)的损失值收敛到最小即可,至此则完成了微调过程。Take all pre-trained model parameters {(W 1 , b 11 ), (W 2 , b 12 ), ..., (W N , b 1N ), θ} as initial values, and take the pre-trained training samples as initial values X trainwhite is the input layer. First, use equation (11) to perform layer-by-layer feature learning through the initial parameters of SDSA model, so as to obtain the feature matrix of the final hidden layer, and calculate the loss value of equation (13) through the feature matrix through the classification layer, and then The feedback propagation algorithm is used to optimize the global parameters, and all model parameters are updated once in each iteration, so that the loss value of equation (13) can be converged to the minimum, and the fine-tuning process is completed. 8.根据权利要求7所述的基于粒子群优化和降噪稀疏编码机的化工故障检测方法,其特征在于,步骤三中,所述离线训练中的粒子群优化可调超参数具体通过以下步骤进行:8. The chemical industry fault detection method based on particle swarm optimization and noise reduction sparse coding machine according to claim 7, is characterized in that, in step 3, the particle swarm optimization adjustable hyperparameter in described offline training specifically passes following steps conduct: (1)、对三个关键的可调参数进行优化,分别为:正则化权重衰减超参数λr,控制稀疏性惩罚项的权重超参数β,Softmax的权重衰减项系数λsm,以所述三个可调参数构成PSO优化的单个粒子(λr,β,λsm),因此各个粒子的维度为3维;定义Np个粒子进行同时寻优,设置K次迭代,初始化各粒子的初始位置和速度,其适应度函数值为训练样本的总体准确率,计算方法是以这三个关键可调超参数为自变量,通过训练好上述微调后的故障检测模型可得出训练样本的总体准确率Accuracy,其公式定义为下式:(1) Optimizing three key adjustable parameters, namely: regularization weight decay hyperparameter λ r , weight hyperparameter β controlling sparsity penalty term, and Softmax weight decay term coefficient λ sm . Three adjustable parameters constitute a single particle (λ r , β, λ sm ) optimized by PSO, so the dimension of each particle is 3 dimensions; define N p particles for simultaneous optimization, set K iterations, and initialize the initial value of each particle Position and speed, the fitness function value of which is the overall accuracy of the training samples. The calculation method uses these three key adjustable hyperparameters as independent variables. After training the above fine-tuned fault detection model, the overall training sample can be obtained. Accuracy, its formula is defined as the following formula: 其中,pi代表本故障检测模型预测出的第i个训练样本的所属类别,yi代表人为给定的标签中第i个训练样本的所属类别,若二者相等,则返回为1,若不相等,意味着当前样本被模型诊断错误,返回值为0;通过上式(14)计算出各粒子的的最优适应度和位置以及粒子群的全局最优适应度和位置;Among them, pi represents the category of the i -th training sample predicted by the fault detection model, and yi represents the category of the i-th training sample in the artificially given label. If the two are equal, it will return 1. If they are not equal, it means that the current sample is wrongly diagnosed by the model, and the return value is 0; the optimal fitness and position of each particle and the global optimal fitness and position of the particle swarm are calculated through the above formula (14); (2)、更新各粒子的速度和位置,设粒子在第t次迭代下的位置可表示为各个粒子在第t次迭代的飞行速度可表示为PSO算法通过以下公式更新各粒子的速度和位置:(2) Update the velocity and position of each particle, and suppose that the position of the particle under the t-th iteration can be expressed as The flight speed of each particle in the t-th iteration can be expressed as The PSO algorithm updates the velocity and position of each particle by the following formulas: 其中Wp为惯性系数,C1为粒子跟踪自身历史最优值的加速系数,表示粒子对自身的认知;C2为粒子跟踪群体最优值的加速系数,表示粒子对群体知识的认知,即社会知识,通常设置C1=C2=2;t为当前迭代次数;ξ和η为在[0,1]区间内的均匀分布的随机数;为截止到第t次迭代时粒子i所经历过的最优适应度的位置;gbt为截止到第t次迭代时粒子群所经历过的最好的位置;Among them, W p is the inertia coefficient, C 1 is the acceleration coefficient of the particle tracking its own historical optimal value, indicating the particle's cognition of itself; C 2 is the acceleration coefficient of the particle tracking group optimal value, indicating the particle's cognition of the group knowledge , namely social knowledge, usually set C 1 =C 2 =2; t is the current number of iterations; ξ and η are uniformly distributed random numbers in the interval [0, 1]; is the position of the optimal fitness experienced by particle i up to the t-th iteration; gb t is the best position experienced by the particle swarm up to the t-th iteration; (3)、重新训练上述故障检测模型的无监督预训练SDSA、有监督预训练Softmax分类器、BP算法全局微调网络参数,得到新的全局最优适应度值和位置,并采用公式(15-16)不断更新粒子的位置和速度,当迭代次数超过其定义的K次迭代时则终止寻优,从而得到全局最优适应度下的粒子位置,即在该最优化可调超参数下,其训练样本的检测正确率最高,从而完成了自动调优超参数;再将其优化出的可调超参数值设置成SDSA的最优化可调超参数,并重新训练出该确定的网络结构和最优超参数下的SDSA的全部模型参数,从而得到训练好的故障检测模型,其模型参数可用于测试样本的特征学习和分类。(3) Retrain the unsupervised pre-training SDSA of the above fault detection model, the supervised pre-training Softmax classifier, and the BP algorithm to globally fine-tune the network parameters to obtain a new global optimal fitness value and position, and use the formula (15- 16) Continuously update the position and speed of the particle, and terminate the optimization when the number of iterations exceeds its defined K iterations, so as to obtain the particle position under the global optimal fitness, that is, under the optimal adjustable hyperparameter, its The detection accuracy rate of the training samples is the highest, thus completing the automatic tuning of hyperparameters; then set the optimized tunable hyperparameters as SDSA's optimized tunable hyperparameters, and retrain the determined network structure and the optimal tunable hyperparameters. All model parameters of SDSA under the optimal hyperparameters are obtained to obtain a trained fault detection model, whose model parameters can be used for feature learning and classification of test samples. 9.根据权利要求8所述的基于粒子群优化和降噪稀疏编码机的化工故障检测方法,其特征在于,步骤四中,所述在线监控的特征学习和隶属概率预测通过以下步骤进行:9. the chemical industry fault detection method based on particle swarm optimization and noise reduction sparse coding machine according to claim 8, is characterized in that, in step 4, the feature learning of described online monitoring and membership probability prediction are carried out by following steps: 以预处理后的测试样本Xtestwhite作为输入层,根据训练好的故障检测模型,对确定的神经网络结构,采用逐层特征学习法,通过训练好的各个降噪稀疏自动编码机的权重参数{(W1,b11),(W2,b12),...,(WN,b1n)},采用公式(11)进行前馈学习得到测试样本当前隐层的特征,然后将当前隐层学习到的特征作为输入,联合下一隐层的DSA模型参数采用同样的公式进行学习特征,以此类推学习完所有隐层得到最终学习到的特征,再采用式(12)进行各隶属类别的预测,在故障检测中其类别只有两类,即故障状态和正常状态,因此对于单个样本其预测概率向量只有两个值,选取概率值最大的预测类别作为整个故障检测模型预测出的隶属类别,当计算出的隶属类别为1,则说明该工况属于正常,当隶属类别为2,则说明有异常工况出现,随即连接的报警装置发出预警,指示故障检出,同时通知工艺员或工程师及时地检查系统安全并排除故障,从而可以实现对当前工况进行故障监控。Taking the preprocessed test sample X testwhite as the input layer, according to the trained fault detection model, for the determined neural network structure, the layer-by-layer feature learning method is adopted, and the weight parameters of the trained noise reduction sparse auto-encoder are passed { (W 1 , b 11 ), (W 2 , b 12 ), ..., (W N , b 1n )}, use formula (11) to perform feedforward learning to obtain the features of the current hidden layer of the test sample, and then use the current The features learned by the hidden layer are used as input, and the DSA model parameters of the next hidden layer are combined with the same formula to learn the features, and so on after all hidden layers are learned to obtain the final learned features, and then formula (12) is used for each membership. For category prediction, in fault detection, there are only two categories, namely fault state and normal state. Therefore, for a single sample, the predicted probability vector has only two values, and the predicted category with the largest probability value is selected as the membership predicted by the entire fault detection model. When the calculated affiliation category is 1, it means that the working condition is normal; when the affiliation category is 2, it means that there is an abnormal working condition, and then the connected alarm device will issue an early warning, indicating the fault is detected, and notify the technician at the same time. Or engineers can check the system safety and troubleshoot in time, so that the current working conditions can be monitored for faults.
CN201910016558.1A 2019-01-08 2019-01-08 Chemical industry fault detection method based on particle group optimizing and noise reduction sparse coding machine Pending CN109800875A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910016558.1A CN109800875A (en) 2019-01-08 2019-01-08 Chemical industry fault detection method based on particle group optimizing and noise reduction sparse coding machine

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910016558.1A CN109800875A (en) 2019-01-08 2019-01-08 Chemical industry fault detection method based on particle group optimizing and noise reduction sparse coding machine

Publications (1)

Publication Number Publication Date
CN109800875A true CN109800875A (en) 2019-05-24

Family

ID=66556848

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910016558.1A Pending CN109800875A (en) 2019-01-08 2019-01-08 Chemical industry fault detection method based on particle group optimizing and noise reduction sparse coding machine

Country Status (1)

Country Link
CN (1) CN109800875A (en)

Cited By (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110147648A (en) * 2019-06-20 2019-08-20 浙江大学 Automobile sensor fault detection method based on independent component analysis and sparse denoising self-encoding encoder
CN110210569A (en) * 2019-06-06 2019-09-06 南京工业大学 FCM-LSTM-based chemical storage tank anomaly detection algorithm research
CN110297480A (en) * 2019-06-17 2019-10-01 重庆仲澜科技有限公司 The TE process failure diagnosis method of deepness belief network model based on parameter optimization
CN110412872A (en) * 2019-07-11 2019-11-05 中国石油大学(北京) Reciprocating compressor fault diagnosis and optimization method and device
CN110689140A (en) * 2019-09-27 2020-01-14 广东毓秀科技有限公司 Method for intelligently managing rail transit alarm data through big data
CN110779745A (en) * 2019-10-12 2020-02-11 杭州安脉盛智能技术有限公司 Heat exchanger early fault diagnosis method based on BP neural network
CN110929843A (en) * 2019-10-29 2020-03-27 国网福建省电力有限公司 Abnormal electricity consumption behavior identification method based on improved deep self-coding network
CN111028512A (en) * 2019-12-31 2020-04-17 福建工程学院 A real-time traffic prediction method and device based on sparse BP neural network
CN111459142A (en) * 2020-04-22 2020-07-28 北京航空航天大学 Aircraft liquid cooling failure fault diagnosis method based on stacked sparse noise reduction self-encoder
CN111487950A (en) * 2020-04-24 2020-08-04 西安交通大学 "Prediction-Verification-Feedback-Optimization" closed-loop system for online early warning and offline diagnosis
CN111695631A (en) * 2020-06-12 2020-09-22 泽恩科技有限公司 Method, device, equipment and medium for extracting verification fault features based on SAE
CN111913461A (en) * 2019-09-07 2020-11-10 宁波大学 Distributed chemical process monitoring method based on regularization GCCA model
CN112001440A (en) * 2020-08-20 2020-11-27 苏州鸿哲智能科技有限公司 Fault diagnosis logic algorithm and system
CN112198842A (en) * 2020-08-24 2021-01-08 华东理工大学 Tracing diagnosis method for abnormal operation state of PTA (pure terephthalic acid) oxidation unit
CN112465042A (en) * 2020-12-02 2021-03-09 中国联合网络通信集团有限公司 Generation method and device of classification network model
CN113378468A (en) * 2021-06-18 2021-09-10 中国科学院地理科学与资源研究所 Weight optimization method and system of multidimensional geoscience parameters
CN113657025A (en) * 2021-07-23 2021-11-16 上海睿而维科技有限公司 A track structure multi-sensor dynamic matching system
CN114021469A (en) * 2021-11-15 2022-02-08 浙江大学 Method for monitoring one-stage furnace process based on mixed sequence network
CN114445671A (en) * 2021-12-22 2022-05-06 中国科学院信息工程研究所 A device type-based abnormal flow detection method and device
CN114550257A (en) * 2022-02-23 2022-05-27 上海商汤智能科技有限公司 Face recognition network training method and device, electronic equipment and storage medium
CN114692667A (en) * 2020-12-30 2022-07-01 华为技术有限公司 Model training method and related device
CN114818806A (en) * 2022-04-25 2022-07-29 重庆大学 Gearbox fault diagnosis method based on wavelet packet and depth self-encoder
CN114839531A (en) * 2022-05-25 2022-08-02 淮阴工学院 Motor fault detection method based on group type sparse self-coding and group intelligence
CN115166514A (en) * 2022-06-24 2022-10-11 淮阴工学院 Motor fault identification method and system based on self-adaptive spectrum segmentation and denoising
CN115184054A (en) * 2022-05-30 2022-10-14 深圳技术大学 Semi-supervised failure detection and analysis method, device, terminal and medium for mechanical equipment
CN115268418A (en) * 2022-10-01 2022-11-01 深圳市世坤科技实业有限公司 Electrical control equipment fault alarm system and method
CN115438592A (en) * 2022-11-08 2022-12-06 成都中科合迅科技有限公司 Industrial research and development design data modeling method based on system engineering
CN115708028A (en) * 2021-08-19 2023-02-21 中国石油化工股份有限公司 Chemical process fault diagnosis method and system, electronic device and storage medium
CN116381536A (en) * 2023-03-07 2023-07-04 华中科技大学 A method and system for predicting the state of health of lithium batteries based on regression element learning
CN116559702A (en) * 2023-05-05 2023-08-08 阳光储能技术有限公司 Method and device for predicting battery health state, storage medium and electronic equipment
CN116567719A (en) * 2023-07-05 2023-08-08 北京集度科技有限公司 Data transmission method, vehicle-mounted system, device and storage medium
CN117725480A (en) * 2023-12-16 2024-03-19 国网山东省电力公司青岛供电公司 An intelligent detection method and system for lightning arrester faults
CN118520355A (en) * 2024-06-24 2024-08-20 北京时代启程物联科技有限公司 Equipment abnormality diagnosis method, system and program product based on artificial intelligence
CN118549813A (en) * 2024-06-20 2024-08-27 蜂巢能源科技股份有限公司 Battery fault detection method and system
CN119474670A (en) * 2024-10-28 2025-02-18 湖南大学 An intelligent online test process data anomaly detection method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106323636A (en) * 2016-08-16 2017-01-11 重庆交通大学 Adaptive extraction and diagnosis method for degree features of mechanical fault through stack-type sparse automatic coding depth neural network
CN106682688A (en) * 2016-12-16 2017-05-17 华南理工大学 Pile-up noise reduction own coding network bearing fault diagnosis method based on particle swarm optimization
CN107436597A (en) * 2017-07-17 2017-12-05 华南理工大学 A kind of chemical process fault detection method based on sparse filtering and logistic regression

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106323636A (en) * 2016-08-16 2017-01-11 重庆交通大学 Adaptive extraction and diagnosis method for degree features of mechanical fault through stack-type sparse automatic coding depth neural network
CN106682688A (en) * 2016-12-16 2017-05-17 华南理工大学 Pile-up noise reduction own coding network bearing fault diagnosis method based on particle swarm optimization
CN107436597A (en) * 2017-07-17 2017-12-05 华南理工大学 A kind of chemical process fault detection method based on sparse filtering and logistic regression

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
旷天亮: "基于深度神经网络的复杂化工过程故障检测研究", 《中国优秀博硕士学位论文全文数据库(硕士) 工程科技1辑》 *

Cited By (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110210569A (en) * 2019-06-06 2019-09-06 南京工业大学 FCM-LSTM-based chemical storage tank anomaly detection algorithm research
CN110297480A (en) * 2019-06-17 2019-10-01 重庆仲澜科技有限公司 The TE process failure diagnosis method of deepness belief network model based on parameter optimization
CN110297480B (en) * 2019-06-17 2022-07-01 重庆仲澜科技有限公司 TE process fault diagnosis method of deep belief network model based on parameter optimization
CN110147648B (en) * 2019-06-20 2022-06-17 浙江大学 Automotive sensor fault detection method based on independent component analysis and sparse denoising autoencoder
CN110147648A (en) * 2019-06-20 2019-08-20 浙江大学 Automobile sensor fault detection method based on independent component analysis and sparse denoising self-encoding encoder
CN110412872A (en) * 2019-07-11 2019-11-05 中国石油大学(北京) Reciprocating compressor fault diagnosis and optimization method and device
CN111913461A (en) * 2019-09-07 2020-11-10 宁波大学 Distributed chemical process monitoring method based on regularization GCCA model
CN111913461B (en) * 2019-09-07 2022-03-18 宁波大学 Distributed chemical process monitoring method based on regularization GCCA model
CN110689140A (en) * 2019-09-27 2020-01-14 广东毓秀科技有限公司 Method for intelligently managing rail transit alarm data through big data
CN110779745A (en) * 2019-10-12 2020-02-11 杭州安脉盛智能技术有限公司 Heat exchanger early fault diagnosis method based on BP neural network
CN110779745B (en) * 2019-10-12 2021-07-06 杭州安脉盛智能技术有限公司 Heat exchanger early fault diagnosis method based on BP neural network
CN110929843A (en) * 2019-10-29 2020-03-27 国网福建省电力有限公司 Abnormal electricity consumption behavior identification method based on improved deep self-coding network
CN111028512A (en) * 2019-12-31 2020-04-17 福建工程学院 A real-time traffic prediction method and device based on sparse BP neural network
CN111028512B (en) * 2019-12-31 2021-06-04 福建工程学院 Real-time traffic prediction method and device based on sparse BP neural network
CN111459142A (en) * 2020-04-22 2020-07-28 北京航空航天大学 Aircraft liquid cooling failure fault diagnosis method based on stacked sparse noise reduction self-encoder
CN111487950A (en) * 2020-04-24 2020-08-04 西安交通大学 "Prediction-Verification-Feedback-Optimization" closed-loop system for online early warning and offline diagnosis
CN111487950B (en) * 2020-04-24 2021-11-16 西安交通大学 'prediction-verification-feedback-optimization' closed-loop system for online early warning and offline diagnosis
CN111695631A (en) * 2020-06-12 2020-09-22 泽恩科技有限公司 Method, device, equipment and medium for extracting verification fault features based on SAE
CN112001440A (en) * 2020-08-20 2020-11-27 苏州鸿哲智能科技有限公司 Fault diagnosis logic algorithm and system
CN112198842B (en) * 2020-08-24 2024-04-30 华东理工大学 Tracing diagnosis method for abnormal operation state of PTA oxidation unit
CN112198842A (en) * 2020-08-24 2021-01-08 华东理工大学 Tracing diagnosis method for abnormal operation state of PTA (pure terephthalic acid) oxidation unit
CN112465042A (en) * 2020-12-02 2021-03-09 中国联合网络通信集团有限公司 Generation method and device of classification network model
CN112465042B (en) * 2020-12-02 2023-10-24 中国联合网络通信集团有限公司 Method and device for generating classified network model
CN114692667A (en) * 2020-12-30 2022-07-01 华为技术有限公司 Model training method and related device
CN113378468A (en) * 2021-06-18 2021-09-10 中国科学院地理科学与资源研究所 Weight optimization method and system of multidimensional geoscience parameters
CN113378468B (en) * 2021-06-18 2024-03-29 中国科学院地理科学与资源研究所 Weight optimization method and system for multidimensional geoscience parameters
CN113657025A (en) * 2021-07-23 2021-11-16 上海睿而维科技有限公司 A track structure multi-sensor dynamic matching system
CN115708028A (en) * 2021-08-19 2023-02-21 中国石油化工股份有限公司 Chemical process fault diagnosis method and system, electronic device and storage medium
CN114021469A (en) * 2021-11-15 2022-02-08 浙江大学 Method for monitoring one-stage furnace process based on mixed sequence network
CN114445671A (en) * 2021-12-22 2022-05-06 中国科学院信息工程研究所 A device type-based abnormal flow detection method and device
CN114445671B (en) * 2021-12-22 2025-08-12 中国科学院信息工程研究所 Abnormal flow detection method and device based on equipment type
CN114550257A (en) * 2022-02-23 2022-05-27 上海商汤智能科技有限公司 Face recognition network training method and device, electronic equipment and storage medium
CN114818806A (en) * 2022-04-25 2022-07-29 重庆大学 Gearbox fault diagnosis method based on wavelet packet and depth self-encoder
CN114839531A (en) * 2022-05-25 2022-08-02 淮阴工学院 Motor fault detection method based on group type sparse self-coding and group intelligence
CN115184054B (en) * 2022-05-30 2022-12-27 深圳技术大学 Mechanical equipment semi-supervised fault detection and analysis method, device, terminal and medium
CN115184054A (en) * 2022-05-30 2022-10-14 深圳技术大学 Semi-supervised failure detection and analysis method, device, terminal and medium for mechanical equipment
CN115166514A (en) * 2022-06-24 2022-10-11 淮阴工学院 Motor fault identification method and system based on self-adaptive spectrum segmentation and denoising
CN115268418A (en) * 2022-10-01 2022-11-01 深圳市世坤科技实业有限公司 Electrical control equipment fault alarm system and method
CN115438592A (en) * 2022-11-08 2022-12-06 成都中科合迅科技有限公司 Industrial research and development design data modeling method based on system engineering
CN116381536B (en) * 2023-03-07 2024-03-19 华中科技大学 A method and system for predicting the health status of lithium batteries based on regression element learning
CN116381536A (en) * 2023-03-07 2023-07-04 华中科技大学 A method and system for predicting the state of health of lithium batteries based on regression element learning
CN116559702A (en) * 2023-05-05 2023-08-08 阳光储能技术有限公司 Method and device for predicting battery health state, storage medium and electronic equipment
CN116567719B (en) * 2023-07-05 2023-11-10 北京集度科技有限公司 Data transmission method, vehicle-mounted system, device and storage medium
CN116567719A (en) * 2023-07-05 2023-08-08 北京集度科技有限公司 Data transmission method, vehicle-mounted system, device and storage medium
CN117725480A (en) * 2023-12-16 2024-03-19 国网山东省电力公司青岛供电公司 An intelligent detection method and system for lightning arrester faults
CN117725480B (en) * 2023-12-16 2025-07-01 国网山东省电力公司青岛供电公司 A lightning arrester fault intelligent detection method and system
CN118549813A (en) * 2024-06-20 2024-08-27 蜂巢能源科技股份有限公司 Battery fault detection method and system
CN118520355A (en) * 2024-06-24 2024-08-20 北京时代启程物联科技有限公司 Equipment abnormality diagnosis method, system and program product based on artificial intelligence
CN119474670A (en) * 2024-10-28 2025-02-18 湖南大学 An intelligent online test process data anomaly detection method

Similar Documents

Publication Publication Date Title
CN109800875A (en) Chemical industry fault detection method based on particle group optimizing and noise reduction sparse coding machine
CN111476263B (en) Bearing defect identification method based on SDAE and improved GWO-SVM
CN108197648B (en) Hydroelectric generating set fault diagnosis method and system based on LSTM deep learning model
CN106323636A (en) Adaptive extraction and diagnosis method for degree features of mechanical fault through stack-type sparse automatic coding depth neural network
CN102707256B (en) Fault diagnosis method based on BP-Ada Boost nerve network for electric energy meter
CN114676742B (en) A method for detecting abnormal power consumption in power grid based on attention mechanism and residual network
CN109581339B (en) Sonar identification method based on automatic adjustment self-coding network of brainstorming storm
CN115828754A (en) Tool wear state monitoring method based on multi-scale spatio-temporal fusion network model
CN108875771A (en) A kind of failure modes model and method being limited Boltzmann machine and Recognition with Recurrent Neural Network based on sparse Gauss Bernoulli Jacob
CN114444620B (en) Indicator diagram fault diagnosis method based on generating type antagonistic neural network
CN113890109B (en) Generation method of multiple wind farm power daily scenarios with time-spatial correlation
CN112904810A (en) Process industry nonlinear process monitoring method based on effective feature selection
CN111191823A (en) A production logistics forecasting method based on deep learning
CN116400168A (en) A power grid fault diagnosis method and system based on deep feature clustering
CN112163474A (en) An Intelligent Diagnosis Method of Gearbox Based on Model Fusion
CN116911378A (en) Hydropower unit fault diagnosis method and system based on transfer learning and PSO algorithm
CN119150065A (en) Multi-scale space-time feature fusion chemical process fault diagnosis method
CN113033898A (en) Electrical load prediction method and system based on K-means clustering and BI-LSTM neural network
CN109901064B (en) ICA-LVQ-based high-voltage circuit breaker fault diagnosis method
CN117312938B (en) A fault diagnosis method based on graph isomorphism autoencoder
CN119668235B (en) Nuclear power plant fault diagnosis method based on interpretable deep migration learning
CN114372495A (en) Power Quality Disturbance Classification Method and System Based on Deep Spatial Residual Learning
CN120426117A (en) One-button start-stop operation monitoring method and system for gas-steam combined cycle units
CN112488202A (en) Distributed process monitoring method based on mutual information and projection dictionary pair learning
CN120011958A (en) A method for risk assessment of foundation pit deformation based on monitoring data

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20190524