999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

GA-1DLCNN method and its application in bearing fault diagnosis

2019-04-04 06:02:22YangZhenboJiaMinping

Yang Zhenbo Jia Minping

(School of Mechanical Engineering, Southeast University, Nanjing 211189, China)

Abstract:Due to the fact that the vibration signal of the rotating machine is one-dimensional and the large-scale convolution kernel can obtain a better perception field, on the basis of the classical convolution neural network model (LetNet-5), one-dimensional large-kernel convolution neural network(1DLCNN) is designed. Since the hyper-parameters of 1DLCNN have a greater impact on network performance, the genetic algorithm(GA) is used to optimize the hyper-parameters, and the method of optimizing the parameters of 1DLCNN by the genetic algorithm is named GA-1DLCNN. The experimental results show that the optimal network model based on the GA-1DLCNN method can achieve 99.9% fault diagnosis accuracy, which is much higher than those of other traditional fault diagnosis methods. In addition, the 1DLCNN is compared with one-dimencional small-kernel convolution neural network(1DSCNN) and the classical two-dimensional convolution neural network model. The input sample lengths are set to be 128, 256, 512, 1 024, and 2 048, respectively, and the final diagnostic accuracy results and the visual scatter plot show that the effect of 1DLCNN is optimal.

Key words:one-dimensional convolution neural network; large-size convolution kernel; hyper-parameter optimization; genetic algorithm

Bearings are the most common mechanical parts in rotating machinery. According to statistics, 30% of rotating machinery failures and 40% of motor failures are related to rolling bearings[1-3]. Obviously, bearings are one of the most fault-prone parts in mechanical equipment, so the fault diagnosis of the rolling bearings is necessary.

Traditional fault diagnosis methods are usually divided into three steps: 1) Fault signal acquisition; 2) Feature extraction; 3) Fault diagnosis. There are many research results based on traditional fault diagnosis methods. Abbasion et al.[4]firstly used wavelet analysis to preprocess the vibration signal of the motor bearing, which reduced the noise in the signal, and then they used the support vector machine to classify the faults. Han et al.[5]used LMD to decompose the original signal, then extracted the energy entropy and sample entropy of each product function, and finally used the support vector machine for fault diagnosis. Wu[6]applied wavelet packet decomposition to analyze multi-resolution bearing signals, and compared three diagnosis methods which are the BP, RBF and Elman network. However, manually extracting features requires prior knowledge to achieve better diagnostic results.

Convolution neural network (CNN) is one of the methods of deep learning. It has powerful information mining capability and can automatically extract features, and some achievements have been obtained by using CNN in the field of fault diagnosis. You et al.[7]used CNN to extract the features of vibration signals from rolling bearing, then input the extracted features into SVM for fault identification. Ding et al.[8]performed wavelet packet transform on the vibration signal to obtain its grayscale images, and then used them as the input of CNN to identify faults. Chen et al.[9]used 2DCNN to diagnose faults in the gearbox, whose input matrix is 16×16, reconstructed by 256 statistical features including RMS, standard deviation, skewness, kurtosis, rotation frequency and load. However, the method did not utilize the information mining capability of CNN. Guo et al.[10]proposed an adaptive CNN model, which can speed up the convergence rate to a certain extent and improve the recognition rate by changing the learning rate automatically. Janssens et al.[11]utilized 2DCNN to identify the patterns of four types of rotating machinery faults whose input was derived from DFT (discrete Fourier transform) of vibration acceleration signals acquired by two acceleration sensors. Wang et al.[12]applied the PSO algorithm to optimize the hyper-parameters of the CNN model, so that the network can adaptively learn more rules of the data set. Although some of the methods mentioned above have more or less achieved some good results, they still have the following deficiencies: 1) The traditional fault diagnosis method needs to use prior knowledge to extract features and how to exploit the ability of CNN to extract signal features automatically is worth studying; 2) The two-dimensional convolution kernel of a classical 2DCNN will disrupt the periodicity of fault signal, which renders the extracted features inconspicuous, resulting in inaccurate final diagnostic results; 3) There are many hyper-parameters in the CNN, and the selection of hyper-parameters affects the final effect of the algorithm. Therefore, it is necessary to optimize the hyper-parameters.

In this paper, combining the genetic algorithm and convolution neural network with one-dimensional large convolution kernel can solve the above problems, and we call the hybrid algorithm GA-1DLCNN. A comparative experiment is carried out to prove the effectiveness of the method. In addition, another experiment verifies that large convolution kernels have advantages over small convolution kernels and two-dimensional convolution kernels.

1 Basic Theory

1.1 Theory of CNN

The convolution layer and pool layer of CNN make it superior to traditional neural networks. Different convolution layers contain different numbers of convolution kernels. The convolution kernels continuously move on the feature maps passed on from the previous layer. During the movement, the convolution kernel is continuously convoluted with the overlapping data, and the output of the convolution layer is finally obtained. This constantly moving operation of the single convolution kernel also causes CNN to have the characteristic of weight sharing. Compared with the disadvantage of excessive weights in fully-connected networks, the sharing of the weights of CNN can reduce network parameters, which reduces computational complexity and avoids network overload caused by excessive weights. The following part describes the forward propagation algorithm of CNN.

Here, suppose that the input of the convolution layer isx∈RA×B, whereAis the number of samples,Bis the data dimension, and the output of the convolution layer is calculated as

(1)

In general,the pooling operation on the feature data is a nonlinear sampling method after the convolution operation. Through the pooling operation, the matrix size of the feature map can be effectively reduced, thereby reducing the parameters in the last fully connected layer. Its formula is

xl=down(xl-1)

(2)

wherexl-1denotes the output of the previous convolution layer; down( ) denotes the sampling operation, and the most commonly used sampling operation is the max-pooling, that is, taking the maximum value of the local area.

The CNN consists of pairs of convolution and pooling layers. Normally, the fully connected layer follows the last pooling layer. The fully connected layer of CNN is the same as that of traditional artificial neural networks. The number of full connection layers can be determined according to the specific conditions. The last layer of CNN is the softmax layer. The softmax layer is the classification layer of CNN. The number of neurons on this layer is equal to that of the categories that need to be classified. The formula on the softmax layer is

(3)

whereOrepresents the result of the output layer, which is the output category;Cis the number of failure categories.

Fig.1 shows the 1DLCNN model, and it can directly input the original one-dimensional vibration signal into the model. The model has two convolution layers and two pooling layers. From pooling layer 2 to the fully connected layer, each map of the pooling layer needs to be connected, and the output layer finally classifies the fault categories.

Fig.1 1DLCNN model construction

1.2 Hyper-parameter optimization by genetic algorithm

The CNN model has many hyper-parameters, and different hyper-parameter combinations can lead to varying model generalization performances. As there are many hyper-parameters in the model, different hyper-parameter combinations will vary each model performance. In order to obtain the optimal hyper-parameter combination, this paper proposes a new method combining the genetic algorithm with 1DLCNN,which is called the GA-1DLCNN method. On the one hand, the algorithm can play the role of 1DLCNN without manually extracting features (input raw data directly). On the other hand, it can exert the optimization characteristics of the genetic algorithm. The parameters needed to be optimized are the kernel number of the first convolution layerC1, the kernel number of the second convolution layerC2, the learning rateε, the batch sizeB, and the number of iterationsI.

The genetic algorithm is a method of searching for optimal solutions by simulating natural evolutionary processes, which uses the fitness function and probability transformation rules to guide the search direction. The basic operations of the genetic algorithm include selection, crossover, and mutation. The specific optimization processes are as follows:

1) Initializing the population and encoding. A groupXm×nis randomly generated, and each individualX1×nrepresents a hyper-parameter distribution. Here,nrepresents the number of hyper-parameters, andmis the initial population size. The encoding is shown in Fig.2.

Fig.2 Chromosome coding

2) Determining the fitness function. The squared reconstruction error of the training sample of 1DLCNN model is used as the fitness function, and its formula is as

(4)

whereo=(o1,o2,…,oC) is the actual output(obtained by Eq.(3)); the expected output isy=(y1,y2,…,yC). In this paper, one hot encoding is used to indicate expected output, for example, (0,0,0,1) means this is the fourth category.

3) Selecting. The roulette method based on the fitness ratio selection strategy is used. The probability of selectionpqfor each individualqis

(5)

(6)

whereFqis the fitness value of the individual;λis the coefficient.

4) Cross operation. Since the individual uses real number coding, the cross operation uses the real number crossing method, and the cross operation method of theu-th chromosomeauand thev-th chromosomeavin thew-th place is

(7)

wherebis a random number between [0, 1].

5) Variation. New individuals are generated based on the probability of mutation.

6) Calculate the fitness value and determine whether the maximum evolution algebra is reached, otherwise return to step 3).

A miller symbolizes greed, habitual and uncreative thinking as well as logic as a feeble protection against passion (Olderr 1986).Return to place in story.

After the optimization of the genetic algorithm is completed, the obtained optimal individual is utilized as the hyper-parameter combination of 1DCNN.

2 Rotating Machinery Fault Diagnosis Method Based on GA-1DLCNN

The flow chart of the fault diagnosis method based on GA-1DLCNN is shown in Fig.3. The specific steps are as follows:

1) Collect the fault signal, then intercept signals at equal length to obtain multiple sets of samples, and randomly separate the training set and the test set.

2) Generate an initial population. Set the number of populations and the max generation.

3) Train each 1DLCNN model, and the fitness function is obtained by Eq.(4).

4) Sequentially select, cross, and mutate operations to generate new populations, determine whether the max generation is reached. If the discriminant condition is met, output the optimal network hyper-parameter, otherwise, repeat steps 3) and 4) until the discriminant condition is satisfied.

Fig.3 Fault diagnosis process of rotating machinery based on GA-1DLCNN

5) Input the test set into the optimized model to obtain the fault classification results.

3 Experimental Verification and Discussion

3.1 Experimental data description

The experimental data comes from the bearing failure experiment of Case Western Reserve University. The sampling frequency is 12 kHz and the failures are divided into 3 types: Inner race failure, outer race failure and ball failure. Each failure type has three fault sizes of 0.007, 0.014, and 0.021 mm, respectively, and a total of 9 fault categories were obtained, namely, label1, label2, ... , label9. In this experiment, the data length of each sample was set to be 1 024, and 200 samples can be obtained for each fault. A total of 1 800 samples can be obtained, and 1 200 samples were randomly selected as training samples, and another 600 samples were used as test samples (test samples did not participate in training).

3.2 Test results and analysis

The simulation for 1DLCNN is performed in Tensorflow, a Google open deep learning platform. In addition,the NVIDIA GTX1050Ti GPU is used to accelerate algorithm training. In the evolution process of the genetic algorithm, the number of populations is set to be 5, the max generation is 10, the selection probability is 0.1, the crossover probability is 1, and the probability of mutation is 0.2[13].The results of hyper-parameters optimized by the genetic algorithm are shown in Tab.1. The results of each generation of the genetic algorithm are shown in Fig.4. As the generation continues to increase, the results become increasingly accurate and the recognition rate of the final model reaches 99.83%. Fig.5 shows a confusion matrix of misclassfication. TheX-axis is the target category, and theY-axis is the calculated category. When the target category is the same as the calculated category, the value at the intersection is increased by 1. So, according to Fig.5, the misclassification of each category can be obtained. Only one of 68 samples of category 1 is misclassified into category 1, and the other eight categories are all classified correctly.

Tab.1 Optimal hyper-parameters by genetic algorithm

Fig.4 Genetic algorithm optimization process

Fig.5 Confusion matrix of misclassification

3.2.2 Comparison with traditional methods

Feature extraction and the classifiers are two key points of traditional fault diagnosis methods, which affect the final result greatly. In this comparison experiment,the feature extraction methods used are: 1) Time-frequency mixed statistical indicators (TMSI); 2) Wavelet packet decomposition(WPD); 3) Empirical mode decomposition(EMD). The classifiers used contain BP, SVM and stacked auto-encoder (SAE). 10 time domain statistical indicators and 5 frequency domain statistical indicators are selected as time-frequency mixed statistical features[14]. The wavelet packet decomposition decomposed each fault signal into three layers, and a total of eight different frequency band signals were obtained. The energy value of each frequency band was taken as its feature[15], and therefore, 8 features were acquired in total. The EMD decomposed the sample signal into six IMF components[16], and extracted the time-frequency mixed statistical indicators[14]as features for each component. The comparison results are shown in Tab.2. The method proposed in this paper is more advantageous than the traditional methods of extracting features and identification.

3.2.3 Further comparison with classical CNN

In order to more comprehensively verify the effectiveness of a one-dimensional large-size convolution kernel, a large number of experimental comparisons have been set out, in which the parameters that need to be changed include the length of the input data and the number of iterations. In this comparison, a two-dimensional convolutional neural network(2DCNN), one-dimensional small-kernel convolution neural network (1DSCNN)are selected as the comparison algorithms. The input data length is set to be 128, 256, 512, 1 024, 2 048, respectively, and the number of iterations is set to be 50, 100, 200, 300, respectively. The number of training sets accounts for 2/3 of the total number of samples, and the number of test sets accounts for 1/3 of the total number of samples. The specific data set is divided as shown in Tab.3, and the convolution kernel parameter setting for each method is shown in Tab.4. In order to reduce the error of a single test, the average of five test results is calculated as the final result.

Tab.2 Diagnostic accuracy of different methods

Tab.3 Data set dividing

Tab.4 Corresponding convolution kernel parameter settings

According to Fig.6, the above experiments can verify that 1DLCNN is superior to the classic 2DCNN and 1DSCNN. The reasons are as follows: 1) In the field of image processing, the traditional two-dimensional convolution neural network directly inputs two-dimensional data into the network, and the convolution kernel used is also two-dimensional, but in the field of fault diagnosis, the fault signal is usually one-dimensional. If a one-dimensional signal is reconstructed into two-dimensional data, the two-dimensional convolution kernel is used to extract the signal characteristics, which will disturb the periodicity of the signal to some extent, so that the extracted features are not obvious, resulting in lower accuracy. 2) The larger-scale convolution kernels have a larger field of perception and the extracted features are better. Therefore, the final effect is better than that of the small convolution kernel.

(a)

(b)

(c)

(d)

The above comparison test also explains the effect of the input data length on the recognition rate. The above input data lengths are 128, 256, 512, 1 024, and 2 048, respectively. The general trend in Fig.6 is that as the input data length increases, the recognition rate of the model increases. Due to the fault signal having a certain periodicity, if the intercepted data length is too short relative to the period, the distribution of sample data divided into the same category will be different, reducing the classification accuracy eventually. Similarly, a too long intercepted data length will result in fewer samples, which also reduces the recognition rate. Therefore, in the field of fault diagnosis, the authors believe that when the data volume is sufficient, the input data length can be increased as much as possible, which can improve recognition.

In order to visualize the differences between these methods, we use a technique called t-SNE[17]. It should be noted that we take the data length of 2 048 as an example of feature visualization for analysis. First, the principal component analysis(PCA) is used to reduce the high-dimensional data of the soft max layer to 3-D, and then the scatter plot for each test data is obtained. Fig.7(a) is the feature visualization of 2DCNN, Fig.7(b) is the visualization of 1DSCNN, and Fig.7(c) is the feature visualization of 1DLCNN. Through comparisons, it can be clearly seen that the separation phenomenon of different categories in Fig.7(c) is obvious, while the aliasing phenomenon occurs in the other two cases, which also demonstrates that the classification effect of 1DLCNN is optimal.

(a)

(b)

(c)

4 Conclusions

1) When using the one-dimensional large-kernel convolution neural network for fault diagnosis, an original one-dimensional vibration signal is directly inputted into the network without the expert experience. The diagnosis accuracy of this method is more accurate than that of the traditional fault diagnosis method.

2) The genetic algorithm can optimize the hyper-parameters of the network to obtain a set of optimal solutions. In this way, artificially selecting hyper-parameters can be avoided.

3) The high-dimensional features obtained by 1DSCNN and 2DCNN are confused, while 1DLCNN can separate different features clearly, so the one-dimensional large kernel is more advantageous when dealing with one-dimensional mechanical fault signals.

主站蜘蛛池模板: 国产精品三级av及在线观看| 国产成人亚洲精品色欲AV| 亚洲福利片无码最新在线播放 | 一本大道在线一本久道| 人妻一本久道久久综合久久鬼色| 国产精品亚欧美一区二区三区| 丁香六月激情综合| 一本色道久久88| 精品乱码久久久久久久| 欧美三级不卡在线观看视频| 国产农村精品一级毛片视频| 97在线国产视频| 国产91精选在线观看| 国产一级α片| 中文字幕一区二区视频| 日韩 欧美 国产 精品 综合| 久操线在视频在线观看| 日韩第九页| 幺女国产一级毛片| 91色在线视频| 精品人妻一区无码视频| 欧美天堂在线| 亚洲国产精品一区二区第一页免| 亚洲国产天堂久久综合226114| 亚洲精品男人天堂| 狠狠做深爱婷婷久久一区| 亚洲欧美天堂网| 国产极品粉嫩小泬免费看| 麻豆精选在线| a亚洲视频| 国产精品成人免费视频99| 中文字幕色站| 国产一级视频久久| 四虎亚洲国产成人久久精品| 国产av剧情无码精品色午夜| 欧美综合区自拍亚洲综合绿色| 玖玖精品视频在线观看| 免费一级大毛片a一观看不卡| 亚洲国产成人自拍| 2021国产乱人伦在线播放| 在线观看热码亚洲av每日更新| 操美女免费网站| 97视频在线精品国自产拍| 无码AV动漫| 日韩一区二区三免费高清| 国产成人在线无码免费视频| 无码丝袜人妻| 四虎在线观看视频高清无码| 国产产在线精品亚洲aavv| 国产视频 第一页| 国产一区二区三区精品久久呦| 国产精品人莉莉成在线播放| 成人福利一区二区视频在线| 久操中文在线| 国产亚洲精品自在线| 久久96热在精品国产高清| 91国内外精品自在线播放| 人妻21p大胆| 国产91视频观看| 亚洲一区二区三区在线视频| 国产成人高清亚洲一区久久| 丁香婷婷在线视频| 国产精品自在线天天看片| 中国黄色一级视频| 久久久久久久97| 精品丝袜美腿国产一区| 亚洲欧美成人在线视频| 嫩草在线视频| 狠狠做深爱婷婷综合一区| 欧美成人国产| 国产老女人精品免费视频| 亚洲人在线| 亚洲欧美日韩中文字幕在线| 国产黄色片在线看| 亚洲国产精品成人久久综合影院| 欧美成人a∨视频免费观看| 无码aaa视频| 亚洲国产欧洲精品路线久久| 亚洲天堂网在线播放| 国产福利小视频在线播放观看| 成人综合久久综合| 无码电影在线观看|