999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

A Signal Recognition Algorithm Based on Compressive Sensing and Improved Residual Network at Airport Terminal Area

2021-09-26 02:40:16,,,

,,,

1.College of Civil Aviation,Nanjing University of Aeronautics and Astronautics,Nanjing 211106,P.R.China;

2.The 54th Research Institute of China Electronics Technology Group Corporation,Shijiazhuang 050081,P.R.China;

3.Transport Development Research Center,Zhejiang Scientific Research Institute of Transport,Hangzhou 310023,China

Abstract:It is particular important to identify the pattern of communication signal quickly and accurately at the airport terminal area with the increasing number of radio equipments.A signal modulation pattern recognition method based on compressive sensing and improved residual network is proposed in this work.Firstly,the compressive sensing method is introduced in the signal preprocessing process to discard the redundant components for sampled signals.And the compressed measurement signals are taken as the input of the network.Furthermore,based on a scaled exponential linear units activation function,the residual unit and the residual network are constructed in this work to solve the problem of long training time and indistinguishable sample similar characteristics.Finally,the global residual is introduced into the training network to guarantee the convergence of the network.Simulation results show that the proposed method has higher recognition efficiency and accuracy compared with the state-of-the-art deep learning methods.

Key words:compressed sensing;deep learning;residual network;modulation recognition

0 Introduction

The continuous increase of various civil aviation radio stations at the airport terminal area makes the electromagnetic environment more and more complex,and also leads to more serious frequency conflict interference[1].The United States(Next-Gen),Europe(SESAR)and International Civil Aviation Organization(ICAO)Doc9854 have proposed a vision for a new generation of navigation systems,including utilizing the digital communications to reduce the impact of interference in airspace[2].And the modulation recognition method is widely used in the field of digital communications.How to capture and interpret the communication signal effectively plays a major role in ground-to-air communication.

Traditionally,the decision theory is based on hypothesis likelihood ratio and the methods are based on higher-order statistics and feature extraction[3-4].However,the feature extraction is time-consuming and these methods have not been able to meet the demands of high precision and efficiency in the era of data explosion.So that state-of-the-art methods based on deep learning have emerged.The application in the field of communication brings brand-new opportunities and changes.A modulation recognition method based on deep belief network(DBN)was proposed in Ref[.5]to classify the twodimensional images of universal correlation functions of common digital modulation signals.Converting signal into constellation method and recog-nizing by convolution neural network(CNN)were proposed in Ref.[6]which obtained more accurate results than the support vector machine(SVM)and also avoided the manual selection of features.Instead of converting signals into an images,the inphase component and orthogonal component sampling point data proposed in Ref.[7]were directly used as the input of CNN and the stability of the deep learning method was proved at low signal-tonoise ratio(SNR).The reference data set of modulation recognition learning was given in Ref.[8].A simulation data set simulated in a more realistic radio environment was proposed in Ref[.9],which includes 24 kinds of modulation signals.Deep residual network is widely used in the field of computer vision.Through the method of cross-layer connection,the residual network(RN)can avoid the problem of feature graph loss[10-11].RNs for time series radio classification were proposed in Ref.[12]to reduce the training time,but they did not significantly improve the recognition accuracy.

In signal modulation recognition,compressive sensing(CS)is introduced to preprocess the signal to retain the most useful information before inputting the signal into the neural network.The possibility of using measurement signals without reconstruction for detection was provided in Ref.[13].Spectrum detection based on eigenvalues for compressed unreconstructed signals was proposed in Ref.[14],which greatly reduced the computational complexity,but did not affect the probability of signal detection.

Residual neural network can solve the problems that the model training time is too long and the similar features of the sample are hardly to distinguish,which finally lead to a certain enhancement for the recognition accuracy of samples.The image is two-dimensional convolution while the signal is one-dimensional one.The use of neural networks to classify signal modulation is based on the idea of image classification.In this work,the received signal is regarded as a single channel“picture”,and a method based on the improved RN to identify the signal modulation is proposed.Firstly,the received signal is compressed in the preprocessing to extract most of the useful information.Furthermore,the scaled exponential linear units(SELU)activation function is introduced to construct the RN structure.Global residual is introduced to solve the problem of network convergence.Simulation results show that the proposed method has higher recognition efficiency and accuracy.

1 Signal Model

Combined with the very high frequency(VHF)communications in the airport terminal area,it is assumed that the received signal isr(t),which can be written as

wheres(t)is a modulated signal,crepresents the path loss,andn(t)the white Gaussian noise.The measurement value is obtained by CS as follows

whererisN×1 vector of Nyquist sampling denoting signalr(t),ytheM×1 measurement signal,andΦ∈CM×N(M?N)the measurement matrix.During one cycle,the atom with the largest inner product for the measurement matrix and residual is found and incorporated into the estimation support set.The single atomic selection process can be represented as

whereφiis theith line vector of measurement matrixΦ.

Discrete cosine transformation(DCT)matrix is selected as the sensing matrix in the absence of prior information about the signal.Most of the energy information of the time-domain signal can be compressed into a small number of DCT domains[15],as represented by

wherek∈(0,M-1),i∈(0,N-1),andc(k)is the matrix coefficient.Hence,we arrive at the measurement signal notation of

To prevent over-fitting,three parts data are divided into train set,test set and cross-validation set,y=[y1y2y3]T,wherey1is a training model of training set,y2the validation set,andy3a test set.

2 Improved Residual Network

The proposed residual network and global skip connection are introduced in this section.

2.1 The proposed residual network

Instead of the original rectified linear unit(RELU),the residual unit based on SELU activation function(R-SELU)is constructed to solve the problem of“neuronal death”at the negative gradient,as shown in Fig.1.

Fig.1 Residual block structure

In the residual unit,the input of the first residual unit is set tox l,then the structure of the residual unit is

whereσis the activation function andF(x l,w l)the function of residual unit.

For any input of unit,the relationship between multiple residual units can be obtained by

Assuming thatJis loss function,then the calculation expression for the error back-propagation is followed by

A 6-layer residual neural network is constructed based on the above residual blocks.The structure diagram is shown in Fig.2.

Fig.2 Structure of ResNet network

The expression of each layer of the above network structure is followed by

whereyis the low dimensional measurements of compressed signals,bthe basis,Wconvolution kernel,the numbernand the sizek×k×c.cis the number of channels.Whenn=1,W1=3×1×1.“*”is convolution operation.

wherey'is the reconstructed measurements of the last convolutional layer.

2.2 Activation function

The activation function used for the full-connection layer is SELU

As can be seen from Eq.(11),when the gradient is greater than 0,the positive half-axis gradient is greater than 1,which is well solved in the RELU activation function.It is not simple to set 0 at the negative half-axis,which solves the case of neuronal death.The introduction of the SELU activation function solves the problem of indistinguishable similar features in the sample.The activation function used by the output layer is

Using training setTto train classifiers is to find appropriate parameters to minimize some loss functions of classifiers.The cross entropy loss function is generally used as

where 1{y i=j}means that wheny=j,the value is 1,otherwise the value is 0.The smaller the value of the loss function,the more accurate the result of the classification training set.

However,Eq.(14)is not a strict convex function,which does not guarantee that it has a unique solution.Increasing the weighted attenuation termpenalizes excessive attenuation value,then,the new loss function is

Using the iterative optimization method to solve Eq.(15),the gradient equation can be obtained as

2.3 Global skip connection

In the residual block,the relationship between the input and the output can also be written in a simple way as

wherex llandx ll+1are the input and the output of the residual block,respectively.g(·)is the joint operation between the convolution layers andW llrepresents all contained parameters.Sinceσ(·)is nonlinear,g(x l;W l)+x lis not always greater than zero.

Residual block does not accurately learn the residual between the input and the outputx ll+1-x ll,so the overall residualx L-x llis also inaccurate.

Since the input and output of this work are similar,zero padding is used to keep the data dimension consistent.The overall model of the training network is shown in Fig.3.Received signal is compressed using the signal model method to reduce the redundancy and divided into dataset and label.The dataset is based on the constructed residual network and the global residual connection is introduced.Through the full connection layer with SELU as the activation function,the regression output is carried out,the loss function is calculated,and the training network is evaluated.By introducing global skip connection,the residual neural network detects the residuals from the overall network existing between input and output,which leads to a faster convergence rate.

Fig.3 Structure diagram of overall training network

3 Experiments and Analysis

3.1 Experiment settings

To verify the validity of the proposed method,the open dataset named RML 2018.01a in Ref.[9]was used.74 million signals were extracted,in which 60%data were the training set,20%data were the cross-verification set and the final 20%data were the test set.These extracted data set were divided into 24 parts and preprocessed,including compressing and denoising.The dataset parameters used are listed in Table 1.

Experiments were conducted on Google’s open cloud platform to overcome the inconvenience caused by the hardware limitation.For the sake ofsimplicity,the simulation environment configuration is shown in Table 2.

Table 1 ResNet network layout

Table 2 Simulation environment configuration

After many experiments,the first training time of the whole model is longer by using the setting of hardware conditions.The ResNet network layout is shown in Table 3.

Table 3 Resnet network layout

Dropuout was selected to enhance the generalization ability and robustness of the network.Adam optimizer,an extended algorithm of stochastic gradient descent,was selected in the process of parameter updating,which has excellent performance in practice.Hyper-parameter settings of the training network are shown in Table 4.

Table 4 Hyper-parameter settings of the training network

3.2 Results and disussions

In this work,the conventional CNN,the ResNet+SELU method used in Ref.[9],the SResNet+RELU method and S-ResNet+SELU method constructed based on the RN proposed in this work are selected for comparative simulations.

Network loss trend of training dataset of four methods is shown in Fig.4(a).Considering the way of batch training,the curve will show ups and downs.At the same iteration number,it can be shown that the RN performs well.It can be shown that the loss function reduction rate of S-ResNet+SELU method is the fastest.At the time of epoch about 13,the loss function begins to converge and finally converges to 1.14.The loss function of SResNet+RELU method finally converges to 1.20,and the loss function of ResNet+SELU method tends to be flat in epoch 30 and finally converges to 1.28.The loss function trend of these three methods confirms the loss function change of the activation function in Fig.4(b).The loss function of CNN shows a“hook back”when epoch is 55,indicating that the network reaches a certain depth and the CNN appeares a gradient disappearance.The convergence trend of using RELU and SELU loss function in the improved network is basically the same,but it is easier to converge than using RELU activation function in the residual function.

Fig.4 Network loss trend of training dataset

The recognition accuracy is the ratio between the correct number of samples predicted and the total number of samples,as

The four models are tested many times,the best results are recorded,and the accuracy is shown in Table 5.

Table 5 Accuracy comparison between the proposed methods and traditional methods on the dataset

Under the condition that the improved residual block uses the SELU,the accuracy of the model using the RELU in the full connection layer is only 1.8%higher.However,SELU converges much faster than RELU,as can be seen from Fig.4(b).

The accuracy in the range of SNR of the four methods is shown in Fig.5.

Fig.5 Accuracy trend of test dataset

When SNR is 10 d B,the three methods based on the RN achieve the best effect.At low SNR,the accuracy of the proposed method has been on top of the other methods.The higher-order modulation of the signal is less easily distinguished at low SNR,however,the recognition rate of the traditional residual module is low when it has similar samples,which explains the poor recognition effect of the ResNet+SELU in the low SNR.The S-ResNet+RESU and the S-ResNet+SELU can still achieve a recognition rate of 50%when SNR is 0,and the SResNet+SELU has reached 90.2%of the recognition accuracy at 8 dB.

The parameters of CNN,ResNet+SELU and the S-ResNet+SELU proposed in this work are compared,as shown in Table 6.

Table 6 Number of parameters

CNN uses 257 099 parameters,while ResNet+SELU uses 238 840 parameters.Furthermore,the number of parameters used in the proposed SResNet+SELU is 139 192.It is found that the number of used parameters is reduced by 99 648.In the process of model training,the method in this paper trains epoch that requires a mean detection time of 175 s,and the average time used by ResNet+SELU is 477 s.In hardware limitations,the detection time used in this paper is longer,and the training of the whole model takes about 2.95 h.

The confusion matrix is compared under different SNRs illustrated in Fig.6.When the three kinds of SNR are-2,6 and 10 dB,the modulation mode can still be identified accurately.It is shown that the modulation modes of 32PSK,AM-DSB-WC,FM and OOK can be accurately identified when SNR is-2 dB.AM-DSB-WC is the modulation mode used in VHF communication system,and 256QAM and QPSK cannot be easily distinguished under low SNR because of their similar characteristics.When SNR is 6 dB,there are 16 modulation modes withthe recognition rate of above 80%.When SNR is 10 dB,the confusion matrix can basically get pure diagonal lines,and most modulation modes can be recognized.However,128QAM and AM-DSB-SC have 20%recognition error rate,128APSK and 128QAM have 20% recognition error rate,256QAM and QPSK modulation mode have about 10%recognition error rate.When SNR are 20 dB and 30 dB,there will still be shadows when the high-order PSK modulation is used.However,it has a good recognition effect for the analog modulation used in the traditional VHF communication,and this work will also provide a reference for the choice of digital modulation in the civil aviation VHF communication system in the future.

Fig.6 Confusion matrix of test set

4 Conclusions

Aiming at the signal modulation pattern recognition,this work proposes an improved residual network based on SELU activation function combined with CS.It is shown to solve the problem of high computational complexity,complicated steps of the artificial feature extraction and slow convergence speed featured by the conventional residual network.The proposed algorithm firstly discards the redundant components when extracting information from the input signal using the measurement matrix.Then,when training the classification network,a residual unit and residual block structure based on SELU as the activation function is constructed.Furthermore,to solve the inaccuracy of residual learning and difficulty at network convergence,the concept of global skip connection is introduced.According to a series competing simulations,the proposed method is demonstrated to have the short training time,the faster convergence speed and the better recognition accuracy of the modulation mode when comparing with other four popular methods.

In view of the complex electromagnetic environment in the airport terminal area,the prospect offurther research in this work is to design a recognition method for the performance detection of real signal sources based on deep learning spectrum sensing technology.

AcknowledgementsThis work was supported by the National Natural Science Foundation of China(No.71874081),Special Financial Grant from China Postdoctoral Science Foundation(No.2017T 100366)and Open Fund of Hebei Province Key laboratory of Research on data analysis method under dynamic electro-magnetic spectrum situation.

AuthorProf.SHEN Zhiyuan received his B.S.,M.S,and Ph.D.degrees all from Harbin Institute of Technology in 2008,2010 and 2014,respectively.During 2010 to 2012,he was a visiting scholar at Georgia Institute of Technology.His research interests include air traffic management,airport operation and optimization,digital signal processing and deep learning.

Author contributionsProf.SHEN Zhiyuan designed the study,complied the models,conducted the analysis,interpreted the results and wrote the background of the study.Dr.LI Jia contributed to complying the models and data selection for analysis and simulation design.Ms.WANG Qianqian contributed to complying the models,conducting the simulation and discussion,and writing the manuscript.Ms.HU Yingying contributed to adding supplement for revised version and adjusting the paper format.All authors commented on the manuscript draft and approved the submission.

Competing interestsThe authors declare no competing interests.

主站蜘蛛池模板: 精品国产毛片| 久久精品国产免费观看频道| 国产丰满成熟女性性满足视频| 久久这里只精品国产99热8| 欧美三级视频网站| www.youjizz.com久久| 免费av一区二区三区在线| 日韩一区二区三免费高清| 久久这里只有精品2| 国产亚洲精品在天天在线麻豆 | 午夜免费小视频| 亚洲中文字幕无码mv| 欧美一级高清片欧美国产欧美| 日韩精品成人在线| 国产精品无码翘臀在线看纯欲| 久久影院一区二区h| 午夜综合网| 国产第一福利影院| 麻豆国产原创视频在线播放 | 无码一区二区三区视频在线播放| 无码精油按摩潮喷在线播放 | 欧美中文字幕一区二区三区| 99久久精品国产综合婷婷| 色精品视频| 57pao国产成视频免费播放| 波多野结衣一区二区三区88| 黄色网在线| 欧美国产综合视频| 国产成人综合网| 波多野结衣爽到高潮漏水大喷| 欧美一级特黄aaaaaa在线看片| 亚洲精品手机在线| 国产美女主播一级成人毛片| 欧洲高清无码在线| 四虎国产永久在线观看| 国产主播喷水| 嫩草影院在线观看精品视频| 亚洲综合一区国产精品| 青青青视频免费一区二区| 亚洲免费毛片| 在线免费不卡视频| 青草视频久久| 国产黄网永久免费| 激情综合网激情综合| а∨天堂一区中文字幕| 国产成人亚洲无码淙合青草| 亚洲人成网站18禁动漫无码| 欧美另类视频一区二区三区| 老司机午夜精品视频你懂的| 亚洲欧洲日产国产无码AV| 欧美成a人片在线观看| 伊人久久大线影院首页| 免费av一区二区三区在线| 一级毛片免费观看不卡视频| 亚洲精品无码成人片在线观看| 国产人成乱码视频免费观看| 2021国产精品自拍| 国产精品999在线| 久久亚洲黄色视频| 日韩欧美综合在线制服| 婷婷色一二三区波多野衣| 在线不卡免费视频| 曰韩人妻一区二区三区| 色妺妺在线视频喷水| 欧美日韩国产在线人成app| 亚洲最大在线观看| 色噜噜中文网| 青青青视频免费一区二区| 国产精品成人一区二区不卡| 在线免费不卡视频| 亚洲综合18p| 成人在线观看不卡| 国产网站一区二区三区| 亚洲欧洲日韩久久狠狠爱| 午夜视频日本| 亚洲中文字幕手机在线第一页| 国产精品女人呻吟在线观看| 国产精品免费露脸视频| 无码中文AⅤ在线观看| 欧美成人综合视频| 综合色在线| 亚洲无码熟妇人妻AV在线|