999精品在线视频,手机成人午夜在线视频,久久不卡国产精品无码,中日无码在线观看,成人av手机在线观看,日韩精品亚洲一区中文字幕,亚洲av无码人妻,四虎国产在线观看 ?

A Signal Recognition Algorithm Based on Compressive Sensing and Improved Residual Network at Airport Terminal Area

2021-09-26 02:40:16,,,

,,,

1.College of Civil Aviation,Nanjing University of Aeronautics and Astronautics,Nanjing 211106,P.R.China;

2.The 54th Research Institute of China Electronics Technology Group Corporation,Shijiazhuang 050081,P.R.China;

3.Transport Development Research Center,Zhejiang Scientific Research Institute of Transport,Hangzhou 310023,China

Abstract:It is particular important to identify the pattern of communication signal quickly and accurately at the airport terminal area with the increasing number of radio equipments.A signal modulation pattern recognition method based on compressive sensing and improved residual network is proposed in this work.Firstly,the compressive sensing method is introduced in the signal preprocessing process to discard the redundant components for sampled signals.And the compressed measurement signals are taken as the input of the network.Furthermore,based on a scaled exponential linear units activation function,the residual unit and the residual network are constructed in this work to solve the problem of long training time and indistinguishable sample similar characteristics.Finally,the global residual is introduced into the training network to guarantee the convergence of the network.Simulation results show that the proposed method has higher recognition efficiency and accuracy compared with the state-of-the-art deep learning methods.

Key words:compressed sensing;deep learning;residual network;modulation recognition

0 Introduction

The continuous increase of various civil aviation radio stations at the airport terminal area makes the electromagnetic environment more and more complex,and also leads to more serious frequency conflict interference[1].The United States(Next-Gen),Europe(SESAR)and International Civil Aviation Organization(ICAO)Doc9854 have proposed a vision for a new generation of navigation systems,including utilizing the digital communications to reduce the impact of interference in airspace[2].And the modulation recognition method is widely used in the field of digital communications.How to capture and interpret the communication signal effectively plays a major role in ground-to-air communication.

Traditionally,the decision theory is based on hypothesis likelihood ratio and the methods are based on higher-order statistics and feature extraction[3-4].However,the feature extraction is time-consuming and these methods have not been able to meet the demands of high precision and efficiency in the era of data explosion.So that state-of-the-art methods based on deep learning have emerged.The application in the field of communication brings brand-new opportunities and changes.A modulation recognition method based on deep belief network(DBN)was proposed in Ref[.5]to classify the twodimensional images of universal correlation functions of common digital modulation signals.Converting signal into constellation method and recog-nizing by convolution neural network(CNN)were proposed in Ref.[6]which obtained more accurate results than the support vector machine(SVM)and also avoided the manual selection of features.Instead of converting signals into an images,the inphase component and orthogonal component sampling point data proposed in Ref.[7]were directly used as the input of CNN and the stability of the deep learning method was proved at low signal-tonoise ratio(SNR).The reference data set of modulation recognition learning was given in Ref.[8].A simulation data set simulated in a more realistic radio environment was proposed in Ref[.9],which includes 24 kinds of modulation signals.Deep residual network is widely used in the field of computer vision.Through the method of cross-layer connection,the residual network(RN)can avoid the problem of feature graph loss[10-11].RNs for time series radio classification were proposed in Ref.[12]to reduce the training time,but they did not significantly improve the recognition accuracy.

In signal modulation recognition,compressive sensing(CS)is introduced to preprocess the signal to retain the most useful information before inputting the signal into the neural network.The possibility of using measurement signals without reconstruction for detection was provided in Ref.[13].Spectrum detection based on eigenvalues for compressed unreconstructed signals was proposed in Ref.[14],which greatly reduced the computational complexity,but did not affect the probability of signal detection.

Residual neural network can solve the problems that the model training time is too long and the similar features of the sample are hardly to distinguish,which finally lead to a certain enhancement for the recognition accuracy of samples.The image is two-dimensional convolution while the signal is one-dimensional one.The use of neural networks to classify signal modulation is based on the idea of image classification.In this work,the received signal is regarded as a single channel“picture”,and a method based on the improved RN to identify the signal modulation is proposed.Firstly,the received signal is compressed in the preprocessing to extract most of the useful information.Furthermore,the scaled exponential linear units(SELU)activation function is introduced to construct the RN structure.Global residual is introduced to solve the problem of network convergence.Simulation results show that the proposed method has higher recognition efficiency and accuracy.

1 Signal Model

Combined with the very high frequency(VHF)communications in the airport terminal area,it is assumed that the received signal isr(t),which can be written as

wheres(t)is a modulated signal,crepresents the path loss,andn(t)the white Gaussian noise.The measurement value is obtained by CS as follows

whererisN×1 vector of Nyquist sampling denoting signalr(t),ytheM×1 measurement signal,andΦ∈CM×N(M?N)the measurement matrix.During one cycle,the atom with the largest inner product for the measurement matrix and residual is found and incorporated into the estimation support set.The single atomic selection process can be represented as

whereφiis theith line vector of measurement matrixΦ.

Discrete cosine transformation(DCT)matrix is selected as the sensing matrix in the absence of prior information about the signal.Most of the energy information of the time-domain signal can be compressed into a small number of DCT domains[15],as represented by

wherek∈(0,M-1),i∈(0,N-1),andc(k)is the matrix coefficient.Hence,we arrive at the measurement signal notation of

To prevent over-fitting,three parts data are divided into train set,test set and cross-validation set,y=[y1y2y3]T,wherey1is a training model of training set,y2the validation set,andy3a test set.

2 Improved Residual Network

The proposed residual network and global skip connection are introduced in this section.

2.1 The proposed residual network

Instead of the original rectified linear unit(RELU),the residual unit based on SELU activation function(R-SELU)is constructed to solve the problem of“neuronal death”at the negative gradient,as shown in Fig.1.

Fig.1 Residual block structure

In the residual unit,the input of the first residual unit is set tox l,then the structure of the residual unit is

whereσis the activation function andF(x l,w l)the function of residual unit.

For any input of unit,the relationship between multiple residual units can be obtained by

Assuming thatJis loss function,then the calculation expression for the error back-propagation is followed by

A 6-layer residual neural network is constructed based on the above residual blocks.The structure diagram is shown in Fig.2.

Fig.2 Structure of ResNet network

The expression of each layer of the above network structure is followed by

whereyis the low dimensional measurements of compressed signals,bthe basis,Wconvolution kernel,the numbernand the sizek×k×c.cis the number of channels.Whenn=1,W1=3×1×1.“*”is convolution operation.

wherey'is the reconstructed measurements of the last convolutional layer.

2.2 Activation function

The activation function used for the full-connection layer is SELU

As can be seen from Eq.(11),when the gradient is greater than 0,the positive half-axis gradient is greater than 1,which is well solved in the RELU activation function.It is not simple to set 0 at the negative half-axis,which solves the case of neuronal death.The introduction of the SELU activation function solves the problem of indistinguishable similar features in the sample.The activation function used by the output layer is

Using training setTto train classifiers is to find appropriate parameters to minimize some loss functions of classifiers.The cross entropy loss function is generally used as

where 1{y i=j}means that wheny=j,the value is 1,otherwise the value is 0.The smaller the value of the loss function,the more accurate the result of the classification training set.

However,Eq.(14)is not a strict convex function,which does not guarantee that it has a unique solution.Increasing the weighted attenuation termpenalizes excessive attenuation value,then,the new loss function is

Using the iterative optimization method to solve Eq.(15),the gradient equation can be obtained as

2.3 Global skip connection

In the residual block,the relationship between the input and the output can also be written in a simple way as

wherex llandx ll+1are the input and the output of the residual block,respectively.g(·)is the joint operation between the convolution layers andW llrepresents all contained parameters.Sinceσ(·)is nonlinear,g(x l;W l)+x lis not always greater than zero.

Residual block does not accurately learn the residual between the input and the outputx ll+1-x ll,so the overall residualx L-x llis also inaccurate.

Since the input and output of this work are similar,zero padding is used to keep the data dimension consistent.The overall model of the training network is shown in Fig.3.Received signal is compressed using the signal model method to reduce the redundancy and divided into dataset and label.The dataset is based on the constructed residual network and the global residual connection is introduced.Through the full connection layer with SELU as the activation function,the regression output is carried out,the loss function is calculated,and the training network is evaluated.By introducing global skip connection,the residual neural network detects the residuals from the overall network existing between input and output,which leads to a faster convergence rate.

Fig.3 Structure diagram of overall training network

3 Experiments and Analysis

3.1 Experiment settings

To verify the validity of the proposed method,the open dataset named RML 2018.01a in Ref.[9]was used.74 million signals were extracted,in which 60%data were the training set,20%data were the cross-verification set and the final 20%data were the test set.These extracted data set were divided into 24 parts and preprocessed,including compressing and denoising.The dataset parameters used are listed in Table 1.

Experiments were conducted on Google’s open cloud platform to overcome the inconvenience caused by the hardware limitation.For the sake ofsimplicity,the simulation environment configuration is shown in Table 2.

Table 1 ResNet network layout

Table 2 Simulation environment configuration

After many experiments,the first training time of the whole model is longer by using the setting of hardware conditions.The ResNet network layout is shown in Table 3.

Table 3 Resnet network layout

Dropuout was selected to enhance the generalization ability and robustness of the network.Adam optimizer,an extended algorithm of stochastic gradient descent,was selected in the process of parameter updating,which has excellent performance in practice.Hyper-parameter settings of the training network are shown in Table 4.

Table 4 Hyper-parameter settings of the training network

3.2 Results and disussions

In this work,the conventional CNN,the ResNet+SELU method used in Ref.[9],the SResNet+RELU method and S-ResNet+SELU method constructed based on the RN proposed in this work are selected for comparative simulations.

Network loss trend of training dataset of four methods is shown in Fig.4(a).Considering the way of batch training,the curve will show ups and downs.At the same iteration number,it can be shown that the RN performs well.It can be shown that the loss function reduction rate of S-ResNet+SELU method is the fastest.At the time of epoch about 13,the loss function begins to converge and finally converges to 1.14.The loss function of SResNet+RELU method finally converges to 1.20,and the loss function of ResNet+SELU method tends to be flat in epoch 30 and finally converges to 1.28.The loss function trend of these three methods confirms the loss function change of the activation function in Fig.4(b).The loss function of CNN shows a“hook back”when epoch is 55,indicating that the network reaches a certain depth and the CNN appeares a gradient disappearance.The convergence trend of using RELU and SELU loss function in the improved network is basically the same,but it is easier to converge than using RELU activation function in the residual function.

Fig.4 Network loss trend of training dataset

The recognition accuracy is the ratio between the correct number of samples predicted and the total number of samples,as

The four models are tested many times,the best results are recorded,and the accuracy is shown in Table 5.

Table 5 Accuracy comparison between the proposed methods and traditional methods on the dataset

Under the condition that the improved residual block uses the SELU,the accuracy of the model using the RELU in the full connection layer is only 1.8%higher.However,SELU converges much faster than RELU,as can be seen from Fig.4(b).

The accuracy in the range of SNR of the four methods is shown in Fig.5.

Fig.5 Accuracy trend of test dataset

When SNR is 10 d B,the three methods based on the RN achieve the best effect.At low SNR,the accuracy of the proposed method has been on top of the other methods.The higher-order modulation of the signal is less easily distinguished at low SNR,however,the recognition rate of the traditional residual module is low when it has similar samples,which explains the poor recognition effect of the ResNet+SELU in the low SNR.The S-ResNet+RESU and the S-ResNet+SELU can still achieve a recognition rate of 50%when SNR is 0,and the SResNet+SELU has reached 90.2%of the recognition accuracy at 8 dB.

The parameters of CNN,ResNet+SELU and the S-ResNet+SELU proposed in this work are compared,as shown in Table 6.

Table 6 Number of parameters

CNN uses 257 099 parameters,while ResNet+SELU uses 238 840 parameters.Furthermore,the number of parameters used in the proposed SResNet+SELU is 139 192.It is found that the number of used parameters is reduced by 99 648.In the process of model training,the method in this paper trains epoch that requires a mean detection time of 175 s,and the average time used by ResNet+SELU is 477 s.In hardware limitations,the detection time used in this paper is longer,and the training of the whole model takes about 2.95 h.

The confusion matrix is compared under different SNRs illustrated in Fig.6.When the three kinds of SNR are-2,6 and 10 dB,the modulation mode can still be identified accurately.It is shown that the modulation modes of 32PSK,AM-DSB-WC,FM and OOK can be accurately identified when SNR is-2 dB.AM-DSB-WC is the modulation mode used in VHF communication system,and 256QAM and QPSK cannot be easily distinguished under low SNR because of their similar characteristics.When SNR is 6 dB,there are 16 modulation modes withthe recognition rate of above 80%.When SNR is 10 dB,the confusion matrix can basically get pure diagonal lines,and most modulation modes can be recognized.However,128QAM and AM-DSB-SC have 20%recognition error rate,128APSK and 128QAM have 20% recognition error rate,256QAM and QPSK modulation mode have about 10%recognition error rate.When SNR are 20 dB and 30 dB,there will still be shadows when the high-order PSK modulation is used.However,it has a good recognition effect for the analog modulation used in the traditional VHF communication,and this work will also provide a reference for the choice of digital modulation in the civil aviation VHF communication system in the future.

Fig.6 Confusion matrix of test set

4 Conclusions

Aiming at the signal modulation pattern recognition,this work proposes an improved residual network based on SELU activation function combined with CS.It is shown to solve the problem of high computational complexity,complicated steps of the artificial feature extraction and slow convergence speed featured by the conventional residual network.The proposed algorithm firstly discards the redundant components when extracting information from the input signal using the measurement matrix.Then,when training the classification network,a residual unit and residual block structure based on SELU as the activation function is constructed.Furthermore,to solve the inaccuracy of residual learning and difficulty at network convergence,the concept of global skip connection is introduced.According to a series competing simulations,the proposed method is demonstrated to have the short training time,the faster convergence speed and the better recognition accuracy of the modulation mode when comparing with other four popular methods.

In view of the complex electromagnetic environment in the airport terminal area,the prospect offurther research in this work is to design a recognition method for the performance detection of real signal sources based on deep learning spectrum sensing technology.

AcknowledgementsThis work was supported by the National Natural Science Foundation of China(No.71874081),Special Financial Grant from China Postdoctoral Science Foundation(No.2017T 100366)and Open Fund of Hebei Province Key laboratory of Research on data analysis method under dynamic electro-magnetic spectrum situation.

AuthorProf.SHEN Zhiyuan received his B.S.,M.S,and Ph.D.degrees all from Harbin Institute of Technology in 2008,2010 and 2014,respectively.During 2010 to 2012,he was a visiting scholar at Georgia Institute of Technology.His research interests include air traffic management,airport operation and optimization,digital signal processing and deep learning.

Author contributionsProf.SHEN Zhiyuan designed the study,complied the models,conducted the analysis,interpreted the results and wrote the background of the study.Dr.LI Jia contributed to complying the models and data selection for analysis and simulation design.Ms.WANG Qianqian contributed to complying the models,conducting the simulation and discussion,and writing the manuscript.Ms.HU Yingying contributed to adding supplement for revised version and adjusting the paper format.All authors commented on the manuscript draft and approved the submission.

Competing interestsThe authors declare no competing interests.

主站蜘蛛池模板: 91在线一9|永久视频在线| 最新国语自产精品视频在| 四虎综合网| 欧美午夜精品| 99ri国产在线| 久久综合丝袜日本网| 在线看国产精品| 四虎成人免费毛片| 久久综合九色综合97婷婷| 国产精品性| 亚洲天堂成人| 欧美成人国产| 午夜视频免费一区二区在线看| 日本91在线| 国产精品制服| 五月婷婷中文字幕| 国产色网站| 思思热在线视频精品| 99色亚洲国产精品11p| 欧美亚洲日韩不卡在线在线观看| 国精品91人妻无码一区二区三区| 国产精品污污在线观看网站| 国产网站免费| 中文字幕永久在线观看| 日韩一区二区三免费高清| 日韩精品少妇无码受不了| 国产91全国探花系列在线播放| 日本91视频| 亚洲日本中文字幕乱码中文 | 国产精品久久久久鬼色| 日韩av在线直播| 国产精品第一区| 日韩欧美视频第一区在线观看| 国产免费a级片| 国产一级精品毛片基地| 久久成人18免费| 欧美成人午夜视频| 四虎永久免费地址在线网站| 国产小视频在线高清播放| 激情乱人伦| 最新国产你懂的在线网址| 欧美yw精品日本国产精品| 91成人免费观看| 国产自视频| 亚洲综合精品香蕉久久网| 极品国产一区二区三区| 亚洲午夜福利在线| 欧美乱妇高清无乱码免费| 扒开粉嫩的小缝隙喷白浆视频| 一本久道久久综合多人| 手机永久AV在线播放| 中国精品自拍| 亚洲天堂视频网站| 88国产经典欧美一区二区三区| 国产精品99一区不卡| 免费人成视网站在线不卡| 欧美a网站| 国产精品专区第1页| 在线不卡免费视频| 国产精品嫩草影院av| 成人日韩视频| 亚洲永久免费网站| 中日无码在线观看| 欧美一区二区福利视频| 91在线视频福利| 成人欧美在线观看| 最新加勒比隔壁人妻| 免费黄色国产视频| 欧美一道本| 免费观看亚洲人成网站| 91免费国产在线观看尤物| 九九热在线视频| 国产在线第二页| 欧美日韩国产成人在线观看| 无码福利视频| 亚洲人成网址| 国产无码精品在线| 四虎国产精品永久一区| 在线中文字幕日韩| 成人在线天堂| 97精品久久久大香线焦| 中文无码影院|