Skip to main content

METHODS article

Front. Comput. Neurosci., 06 December 2023

A combination network of CNN and transformer for interference identification

  • School of Aerospace Science and Technology, Xidian University, Xi'an, China

Communication interference identification is critical in electronic countermeasures. However, existed methods based on deep learning, such as convolutional neural networks (CNNs) and transformer, seldom take both local characteristics and global feature information of the signal into account. Motivated by the local convolution property of CNNs and the attention mechanism of transformer, we designed a novel network that combines both architectures, which make better use of both local and global characteristics of the signals. Additionally, recognizing the challenge of distinguishing contextual semantics within the one-dimensional signal data used in this study, we advocate the use of CNNs in place of word embedding, aligning more closely with the intrinsic features of the signal data. Furthermore, to capture the time-frequency characteristics of the signals, we integrate the proposed network with a cross-attention mechanism, facilitating the fusion of temporal and spectral domain feature information through multiple cross-attention computational layers. This innovation obviates the need for specialized time-frequency analysis. Experimental results demonstrate that our approach significantly improves recognition accuracy compared to existing methods, highlighting its efficacy in addressing the challenge of communication interference identification in electronic warfare.

1 Introduction

Interference identification has received increasing attention in military and civilian applications (Zhang et al., 2013). Interference identification aims at recognizing the category of interference without any prior information, which is of great importance for anti-interference communications.

Interference identification methods are commonly classified into two categories: feature-based and learning-based methods. Feature-based techniques utilize parameters such as amplitude, phase, and wavelet transform as extracted features for classifiers (Ibrahim et al., 2019; Nishio et al., 2019). In their work, Zhang and Cao (2018) introduced a waveform classification approach based on Support Vector Machines (SVM) tailored for automotive radar interference.

Subsequently, the integration of machine learning and swarm intelligence techniques has shown significant promise in yielding exemplary outcomes across diverse fields (Malakar et al., 2020; Bacanin et al., 2021, 2022; Ramanan et al., 2022; Tuba et al., 2022).

Recently, the widespread adoption of deep learning has garnered significant attention in various fields, including the analysis of clustered weather patterns (Chattopadhyay et al., 2020), as well as image detection (Qian et al., 2021; Lin et al., 2022; Zhou et al., 2022; Lin et al., 2023) and processing (Gawande et al., 2022; Zivkovic et al., 2022; Zhang et al., 2023). Benefitting from the powerful feature extraction capability of deep learning, learning-based methods also have achieved good performance in identification of communication signals (Kattenborn et al., 2021; Sun et al., 2021). O’Shea et al. (2016) used convolutional neural networks (CNNs) to classify wireless modulated signals, and the effectiveness of the method was experimentally demonstrated. After that, Schmidt et al. (2021) used CNNs to study the automatic recognition of interference signals. Due to the simple structure of the network, the recognition accuracy could also be improved.

In Li et al. (2019), carried out radio signal recognition method based on gated recurrent unit (GRU). Compared to CNNs, GRU has more advantages in feature extraction of one-dimensional signals. However, it is difficult to make GRU into a multi-layer structure and that limits its feature extraction capability for long sequences. Residual network (ResNet) was employed for modulation mode identification in West and O’Shea (2017). The method alleviates the problem of gradient decay in deeper networks. However, excessive use of the residual structure can also lead to a larger amount of model parameters and waste of computational resources. In Zhang et al. (2018), a combination of CNNs and long short-term memory (LSTM) was proposed and experimental results showed that it has better recognition performance than either CNNs or LSTM. It was shown that the effective combination of composite networks can improve recognition results. Zhang et al. (2019) constructed four classical neural network models to identify three types of wireless interference signals, which demonstrate the generality of the effectiveness of deep learning at the considered task. Wang et al. (2020) achieved satisfactory results in modulation mode classification by using two CNNs for weight sharing and designing a new loss function. Influenced by the development of transformer (Vaswani et al., 2017; Dosovitskiy et al., 2021; Liu et al., 2021), the utilizations of transformer in signal recognition field (Huang et al., 2022; Wang et al., 2022a) have achieved better performance than CNNs. In Wang et al. (2022b), short-time Fourier transform (STFT) was used for time-frequency analysis, and this method exploits the multi-domain information of the signal. However, signals in different domains need to be processed with different branched networks, while the dedicated time-frequency analysis step adds to the process of interference identification.

Inspired by the above study, we explore the application of transformer in interference identification. Moreover, considering that the disadvantage of transformer in local feature capture capability, this paper designs a novel network architecture, which combines CNNs and transformer (CNNTF). This fusion is not only unique, but also enables more comprehensive signal analysis. In summary, this paper makes the following contributions:

• Firstly, we introduce a CNNTF network. In contrast to the conventional practice of employing simple network combinations, this paper introduces a novel approach by utilizing CNNs in lieu of word embedding. This decision stems from the recognition of the inherent complexity associated with contextual semantics in signal data, which poses challenges for comprehension using word embedding techniques. This modification significantly enhances the network’s applicability in extracting features from signal data, which equip it with both local and global extraction capabilities.

• In addition, we integrated CNNTF with a cross-attention mechanism (CNNTF-CA) to exploit the correlations between different features. This integration allows the network to extract multiple domain features simultaneously, without requiring any special time-frequency analysis. As a result, the network can associate time-domain and frequency-domain features effectively. Our approach represents an innovative way to enhance the capabilities of neural networks for feature extraction.

• The experimental results validate the effectiveness of the proposed method.

2 Signal model

In this section, five types of single interference signals, which consists of single-tone (ST), multi-tone (MT), linear sweep (LS), partial band noise (PBN) and noise frequency modulation (NFM), are used. The signal model can be denoted as

R t = S t e j 2 π f s t + φ s t + J t e j 2 π f J t + φ J t + W t     (1)

where R t represents the received signal. S t is communication signal, f s and φ s t are separately carrier frequency and initial phase of S t . J t is jamming signal, f J t and φ J t are carrier frequency and initial phase of J t , respectively. W t is additive white Gaussian noise (AWGN).

Additionally, the interference signals can be expressed in both time-domain and frequency-domain. Frequency domain data can be obtained from time domain data by fast Fourier transform (FFT), which can be written as

x k = n = 0 N 1 x n e j 2 π k n / N     (2)

where e j 2 π / N denotes the rotation factor. n and k denote the discrete points in the time and frequency domains, respectively. j is the imaginary part.

After that, take the amplitude and phase of the FFT data to obtain the amplitude spectrum and phase spectrum data.

3 Methods

In this paper, we propose a CNNTF method which combines CNNs and transformer. Based on CNNTF, we introduce a cross-attention mechanism to design the CNNTF-CA model, which can effectively fuse features from different domains to achieve the purpose of time-frequency analysis.

3.1 CNNTF

The CNNTF is designed to combine CNN and the encoding of transformer, discarding the word embedding layer of transformer. The utilization of this module has two main advantages. Firstly, for communication interference signals, the local correlation between adjacent sampling points affects the training effect of the model and should not be ignored. CNNs has the advantage of local connectivity in learning features specifically for features between adjacent samples of the signal sequence. Secondly, considering the complexity in extracting contextual semantics from 1D signal data, CNNs are deemed more appropriate than word coding for effectively addressing the practical challenges in this task.

The structure of the CNNs module is as follows. The dimensional convolution kernel scans the interfering data sequence first. In order to avoid gradient dissipation, batch normalization (BN) and rectified linear unit (ReLU) activation function processing are performed after the convolutional operation.

The mathematical expressions below can model the operations of the local 1-D convolution module:

O c = F c I θ O ReLU = ReLU F B N W O c     (3)

where F c means the convolution function, I is the input signal and θ is the parameter in CNNs. F B N denotes the BN processing, and W stands for the weight of convolutional layer. In addition, O c and O ReLU are the output of CNNs layer and the ReLU activation layer, respectively.

The transformer module consists of an attention layer (AL) and a feedforward network (FFN). The attention function can be described as

Attention Q K V = softmax Q K T d V     (4)

where Q , K and V are the query, key and value matrices separately.

The FFN is used after AL, which is composed of two linear translation layers. After the first linear layer, a ReLU activation function is employed, and the whole process can be described as

F F N x = ReLU W 2 W 1 x + b 1 + b 2     (5)

where W 1 R C × C and W 2 R C × C can be used to describe the weights of different layers, separately; b 1 and b 2 denote the offset quantity of different layers, respectively.

There is an interlayer between the attention and FFN layers, which consists of residual connection (RC) and layer normalization (LN). The reason for using the residual connection is to prevent gradient dissipation with the network depth increasing, which can be formulated as follows:

x l + 1 = H x l + F x l W l     (6)

where x l and x l + 1 are the input and output vectors of the lth layer, respectively. H x l means the direct mapping; F x l W l represents the residual mapping. All the layers use residual connections to each other. LN follows RC in the interlayer, which provides better performance for the processing of batches with small size.

To ensure that the dimension of the output is consistent with that of the previous layer, a one-dimensional deconvolution layer is needed to reduce the dimension before the output. Then, after the linear layer and normalization, realized by SoftMax function, the output result is obtained.

3.2 Cross-attention mechanism

The time domain and frequency domain are the basic properties of the communication signals. In the field of signal processing, there are usually special time-frequency analysis steps to combine the time-frequency domain data, which will also make the interference identification process more complex. Therefore, this paper introduces the cross-attention mechanism to combine the characteristics of time domain and frequency domain to play the role of time-frequency analysis. In this paper, in order to reduce the time-frequency analysis process, a cross-attention mechanism is used to correlate the data from two different domains. The overall structure of CNNTF-CA is shown in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1. The structure diagram of CNNTF-CA. The CNNTF-CA contains structure of CNNTF, and PE is the positional encoding. A, T, and P represent amplitude spectrum, time domain and phase spectrum data.

The detailed cross-attention calculation of layer1 process is shown in Figure 2.

FIGURE 2
www.frontiersin.org

Figure 2. Cross-attention calculation detail diagram. www.frontiersin.org represents the dot product.

The cross-attention operation of layer1 can be formulated by

O 1 = softmax Q t K a T d k V a O 2 = softmax Q t K p T d k V p     (7)

where Q t = q t 1 , q t 2 , , q t n is the query vector composed of time-domain feature sequences. K a = k a 1 , k a 2 , , k a n and K p = k p 1 , k p 2 , , k p n represent the key vectors composed of feature sequence in the frequency domain after linear mapping. Besides, V a = v a 1 , v a 2 , , v a n and V p = v p 1 , v p 2 , , v p n are value vectors. means the dot product of the matrix. O 1 and O 2 represent the output of the first layer of two cross-attention modules.

The cross-attention operation of next layer can be described as follows:

O r = softmax Q O 1 K O 2 T d k V O 2     (8)

where the query vector Q O 1 is constructed by linear transformation of O 1 . O 2 is linearly transformed to obtain the key vector K O 2 and the value vector V O 2 . After that, Q O 1 , K O 2 and V O 2 are fed into the next layer of the cross-attention module for deep feature fusion.

The result obtained after the cross-attention mechanism is the input of the FFN, which can be described as

O F F N = F F N O r     (9)

where O r is the result of a two-level cross-attention module.

The output of the previous layer is subjected to an inverse convolution operation, which can be formulated as

O final = softmax ( ConvTranspose 1 D O F F N     (10)

where O F F N is the output of FFN, and O final is the identification result of CNNTF-CA. ConvTranspose 1 D represents the deconvolution operation, which performs up sampling on data to ensure that the output dimensions match the input dimensions.

4 Experiments and results analysis

4.1 Datasets

We select two signals, Binary phase Shift Keying (BPSK) and Quadrature Phase Shift Keying (QPSK), as the communication signal S t . The carrier frequency is set to 2 MHz for signal S t . In addition, the signal-noise-ratio (SNR) is set to [−20 dB, 18 dB] with an interval of 2 dB for the experiments in this paper.

For the interference data set, this paper firstly simulates five single interference signals, generates 1,000 samples under each SNR, each sample is sampled 1,024 times in the time domain. The parameters such as the center frequency, period and bandwidth of each type of interference signal are randomly distributed to simulate the real environment. Then the time domain data is changed by FFT to obtain the amplitude spectrum and phase spectrum data.

Under each SNR, the time domain, amplitude spectrum and phase spectrum are used as the three characteristics of the signal to splice and construct the data sets. The main simulation parameters for each type of interference signal are shown in Table 1. The interference signals are generated in MATLAB and model training and testing using python.

TABLE 1
www.frontiersin.org

Table 1. Interference signal simulation parameters.

4.2 Performance of the proposed CNNTF

To evaluate the performance of our proposed method, the CNNTF are compared with the state-of-the-art methods including CNN (O’Shea et al., 2016), LSTM (Rajendrans et al., 2018), ResNet (West and O’Shea, 2017), CLDNN (Zhang et al., 2018), and GRU (Hong et al., 2017) in this paper.

Table 2 shows the overall recognition accuracy of each model on different sources. The overall accuracy represents the average recognition accuracy of each model for various types of interference under each SNR.

TABLE 2
www.frontiersin.org

Table 2. Overall accuracy (%) of different models.

It can be observed that the method we proposed is higher in recognition accuracy than current mainstream methods. The average recognition accuracy of the six models for various types of recognition accuracy with SNR for six models is shown in Figure 3.

FIGURE 3
www.frontiersin.org

Figure 3. Accuracy diagram of interference recognition.

4.3 Performance of the proposed CNNTF-CA

Our proposed CNNTF demonstrates certain advantages over similar methods, owing to its capacity in extracting both global and local features, which brings in a high degree of information concentration. CNN, LSTM and GRU could not extract both global and local features. Compared with ResNet and CLDNN, which consider both global and local feature information, the advantages of the proposed CNNTF is slightly better. To further improve the performance, we introduced a cross-attention mechanism.

Figure 4 shows comparison chart of overall recognition accuracy between CNNTF and CNNTF-CA. From the figure, it can be observed that the recognition performance of CNNTF-CA has significantly improved under low SNR. The results are due to the use of the cross-attention mechanism, the time-frequency features are deeply correlated and the features are more differentiated between each type of modulated signal. CNNTF only performs simple feature splicing, so its performance is slightly worse than CNNTF-CA.

FIGURE 4
www.frontiersin.org

Figure 4. Overall recognition accuracy CNNTF and CNNTF-CA.

Table 3 presents the recognition performances of CNNTF-CA for each type of interference.

TABLE 3
www.frontiersin.org

Table 3. Average accuracy (%) of CNNTF-CA.

It can be seen from Table 3 that ST has the highest probability of being accurately identified among the five types of interference signals. In addition, the recognition effect of interference on QPSK is better than that on BPSK, which also shows that QPSK contains more information than BPSK. Simultaneously, PBN and NFM are the two types of interference that are most difficult to identify, whether under BPSK or QPSK. We display the recognition accuracy of CNNTF-CA for each interference in Figure 5.

FIGURE 5
www.frontiersin.org

Figure 5. Identification accuracy of CNNTF-CA for each interference under BPSK and QPSK. Among them, ST is single-tone interference, MT is multi-tone interference, LS is linear scan interference, PBN and NFM represent partial band noise interference and noise frequency modulation interference respectively.

The recognition accuracy of the CNNTF-CA approach for various interferences under BPSK and QPSK is depicted in Figure 5, as shown in this scientific figure.

It can be seen from the figure that the recognition accuracy curve of CNNTF-CA for different interferences has a similar trend, which also reflects the versatility and mobility of CNNTF-CA. We find that the recognition accuracy of different interference signals varies greatly, especially when the SNR is low.

In order to present the results more intuitively, we use histograms in Figure 6 to depict the two signals with the best and worst effects in BPSK and QPSK at -20 dB, respectively. This approach aims to provide a more intuitive description of the results.

FIGURE 6
www.frontiersin.org

Figure 6. Identification accuracy of CNNTF-CA for each interference under BPSK and QPSK.

It is apparent that the model favors the identification of ST signals; however, its performance in recognizing NFM interference signals remains inadequate.

5 Conclusion

The performance of the proposed CNNTF-CA model is evaluated through the confusion matrices presented in Figures 7A,B for BPSK and QPSK, respectively, at a signal-to-noise ratio of -10 dB.

FIGURE 7
www.frontiersin.org

Figure 7. Confusion matrix of CNNTF-CA at -10 dB.

According to the confusion matrix illustrated in Figure 7, which represents the accuracy of identifying various interference signals under an SNR of -10 dB, it is apparent that the NFM and PBN signals exhibit relatively higher rates of misidentification when compared to the other signals present in the single interference data set. Specifically, the network demonstrates significant recognition errors in identifying NFM and PBN signals, highlighting a limitation that requires further attention in future research endeavors.

In addition, more precise assessment metrics can be derived based on Table 4. It can be seen that ST and LS are more likely to be correctly identified whether under BPSK or QPSK.

Furthermore, it is evident that regardless of the type of interference signal, the accurate recognition rate for QPSK is higher than that for BPSK, indicating the richer signal information contained within QPSK. These findings help the proposed model identify different interference signals faster and more accurately, playing a more important role in actual confrontation scenarios.

TABLE 4
www.frontiersin.org

Table 4. Performance evaluation of CNNTF-CA under –10 dB.

In this paper, we propose a novel method that combines these CNN and transformer (CNNTF), to address the problem of identifying five single interferences. Given the challenge of extracting contextual semantics from one-dimensional signals using word encoding, this paper introduces a pioneering approach that exploits CNN instead. This novel combination, tailored to the unique data characteristics of one-dimensional signals, represents a significant contribution to the field. To further enhance the performance of the CNNTF model, we also incorporate a cross-attention mechanism that facilitates the correlation of the time and frequency domains of the input signals. This mechanism replaces the traditional approach of separate time-frequency analysis, leading to improved accuracy and efficiency in the identification and classification of different interference types. The effectiveness of the proposed approach is evaluated through extensive experiments and comparisons with other state-of-the-art methods. The experimental results demonstrate that the proposed CNNTF model with cross-attention mechanism achieves better performance in identifying and classifying different types of interferences.

Despite the promising results, it is important to acknowledge certain limitations and directions for future research. Current research is mainly limited to the evaluation of the CNNTF-CA model in simple scenarios. Further research on its performance under complex interference scenarios would be beneficial. To bridge the gap between theory and practical implementation, future research efforts will focus on optimizing the model’s robustness to changes in real-world signal conditions and extending its applicability to different signal interference environments.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

HZ: Conceptualization, Methodology, Writing – original draft. MeZ: Conceptualization, Methodology, Writing – original draft. MiZ: Funding acquisition, Writing – review & editing. SL: Writing – original draft, Writing – review & editing. YD: Writing – review & editing. HW: Supervision, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This research was funded by the National Natural Science Foundation of China Under Grant 12003018.

Acknowledgments

The authors would like to thank the reviewers for their valuable and detailed comments that are crucial in improving the quality of this manuscript.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Bacanin, N., Stoean, R., Zivkovic, M., Petrovic, A., Rashid, T., and Bezdan, T. (2021). Performance of a novel chaotic firefly algorithm with enhanced exploration for tackling global optimization problems. Applic. Dropout Regular. 9, 2–33. doi: 10.3390/math9212705

CrossRef Full Text | Google Scholar

Bacanin, N., Zivkovic, M., Ai-Turjman, F., Pavel, V., Strumberger, I., and Bezdan, T. (2022). Hybridized sine cosine algorithm with convolutional neural networks dropout regularization application. Sci. Rep. 12:6302. doi: 10.1038/s41598-022-09744-2

CrossRef Full Text | Google Scholar

Chattopadhyay, A., Hassanzadeh, P., and Pasha, S. (2020). Predicting clustered weather patterns: a test case for applications of convolutional neural networks to spatio-temporal climate data. Sci. Rep. 10:1317. doi: 10.1038/s41598-020-57897-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., et al. An image is worth 16x16 words: transformers for image recognition at scale. In Proceedings of the international conference on learning representations (ICLR), Vienna, Austria (2021)

Google Scholar

Gawande, U., Hajari, K., and Golhar, Y. (2022). Robust pedestrian detection using scale and illumination invariant mask r-cnn. Int. J. Comput. Sci. Eng. 25, 607–618. doi: 10.1504/IJCSE.2022.127190

CrossRef Full Text | Google Scholar

Hong, D., Zhang, Z., and Xu, X. (2017). Automatic modulation classification using recurrent neural networks. In Proceedings of the IEEE 3 international conference on computer and communications (ICCC), Chengdu, China, 695–700.

Google Scholar

Huang, J., Li, X., Wu, B., Wu, X., and Li, P. (2022). Few-shot radar emitter signal recognition based on attention-balanced prototypical network. Remote Sens. 14:6101. doi: 10.3390/rs14236101

CrossRef Full Text | Google Scholar

Ibrahim, M., Parrish, D. J., Brown, T. W. C., and McDonald, P. J. (2019). Decision tree pattern recognition model for radio frequency interference suppression in NQR experiments. Sensors 19, 3153–3156. doi: 10.3390/s19143153

PubMed Abstract | CrossRef Full Text | Google Scholar

Kattenborn, T., Leitloff, J., Schiefer, F., and Hinz, S. (2021). Review on convolutional neural networks (CNN) in vegetation remote sensing. ISPRS J. Photogramm. Remote Sens. 173, 24–49. doi: 10.1016/j.isprsjprs.2020.12.010

CrossRef Full Text | Google Scholar

Li, R., Hu, J., and Yang, S. (2019). Deep gated recurrent unit convolution network for radio signal recognition. In Proceeding of the 2019 IEEE 19th International Conference on Communication Technology (ICCT), Xi’an, China, 159–163.

Google Scholar

Lin, S., Zhang, M., and Cheng, X. (2022). Dual collaborative constraints regularized low rank and sparse representation via robust dirtionaries construction for hyperspectral anomaly detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 16, 2009–2024. doi: 10.1109/JSTARS.2022.3214508

CrossRef Full Text | Google Scholar

Lin, S., Zhang, M., and Cheng, X. (2023). Dynamic low-rank and sparse priors constrained deep autoencoders for hyperspectral anomaly detection. IEEE Trans. Instrum. Meas. doi: 10.1109/TIM.2023.3323997

CrossRef Full Text | Google Scholar

Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., et al. (2021). Swin transformer: hierarchical vision transformer using shifted windows. In Proceedings of the IEEE international conference on computer vision (ICCV), Montreal, Canada

Google Scholar

Malakar, S., Ghosh, M., Bhowmik, S., Sarkar, R., and Nasipuri, M. (2020). A GA based hierarchical feature selection approach for handwritten word recognition. Neural Comput. Applic. 32, 2533–2552. doi: 10.1007/s00521-018-3937-8

CrossRef Full Text | Google Scholar

Nishio, T., Okamoto, H., Nakashima, K., Koda, Y., Yamamoto, K., Morikura, M., et al. (2019). Proactive received power prediction using machine learning and depth images for mmWave networks. IEEE J Select Areas Commun 37, 2413–2427. doi: 10.1109/JSAC.2019.2933763

CrossRef Full Text | Google Scholar

O’Shea, T., Timothy, J., Corgan, J., and Clancy, T.C. (2016). Convolutional radio modulation recognition networks. In Proceedings of the engineering applications of neural networks (EANN), Aberdeen, UK, 213–226.

Google Scholar

Qian, X., Cheng, X., and Cheng, G. (2021). Two-stream encoder GAN with progressive training for co-saliency detection. IEEE Signal Process. Lett. 28, 180–184. doi: 10.1109/LSP.2021.3049997

CrossRef Full Text | Google Scholar

Rajendrans, S., Meert, W., Giustiniano, D., Lenders, V., and Pollin, S. (2018). Deep learning models for wireless signal classification with distributed low-cost Spectrum sensors. IEEE Trans. on Cog. Commun. and Net. 4, 433–445. doi: 10.1109/TCCN.2018.2835460

CrossRef Full Text | Google Scholar

Ramanan, M., Singh, L., Suresh Kumar, A., Suresh, A., Sampathkumar, A., Jain, V., et al. (2022). Secure blockchain enabled cyber-physical health systems using ensemble convolution neural network classification. Computes Electr Engin. 101:108058. doi: 10.1016/j.compeleceng.2022.108058

CrossRef Full Text | Google Scholar

Schmidt, M., Block, D., and Meier, U. (2021). Wireless interference identification with convolutional neural networks. In Proceedings of the IEEE 15th international conference on industrial informatics (INDIN), Palam de Mallorca, Spain, 180–185

Google Scholar

Sun, Y., Xue, B., Zhang, M., Yen, G. G., and Lv, J. (2021). Automatically designing CNN architectures using the genetic algorithm for image classification. IEEE Trans. Cybern. 50, 3840–3854. doi: 10.1109/TCYB.2020.2983860

CrossRef Full Text | Google Scholar

Tuba, E., Strumberger, I., Tuba, I., Bacanin, N., and Tuba, M. (2022). Acute lymphoblastic leukemia detection by tuned convolutional neural network. In 2022 32nd international conference RADIOELEKTRONIKA (RADIOELEKTRONIKA).

Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., et al. (2017). Attention is all you need. In Proceedings of the conference workshop neural information processing systems (NIPS)

Google Scholar

Wang, P., Cheng, Y., Dong, B., and Peng, Q. (2022a). Bring Globality into convolutional neural networks for wireless interference classification. IEEE Wireless Commun. Lett. 11, 538–542. doi: 10.1109/LWC.2021.3135901

CrossRef Full Text | Google Scholar

Wang, P., Cheng, Y., Dong, B., Peng, Q., and Li, S. (2022b). Multi-domain networks for wireless interference recognition. IEEE Trans. Veh. Technol. 71, 6534–6547. doi: 10.1109/TVT.2022.3164908

CrossRef Full Text | Google Scholar

Wang, P., Cheng, Y., Dong, B., and Xun, H. (2020). Convolutional neural network-based interference recognition. In Proceedings of the IEEE 20th international conference on communication technology (ICCT), Nanning, China 1–5.

Google Scholar

West, N.E., and O’Shea, T.J. (2017). Deep architectures for modulation recognition. In Proceedings of the IEEE 2017th international symposium on dynamic Spectrum access networks (DySPAN), Baltimore, USA 1–6

Google Scholar

Zhang, R., and Cao, S. (2018). Support vector machines for classification of automotive radar interference. In Proceedings of the 2018 IEEE radar conference, Oklahoma, USA, 366–371.

Google Scholar

Zhang, D., Ding, W., Zhang, B., Xie, C., Li, H., Liu, C., et al. (2018). Automatic modulation classification based on deep learning for unmanned aerial vehicles. Sensors 18:924. doi: 10.3390/s18030924

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, X., Seyfi, T., Ju, S., Ramjee, S., Gamal, A.E., and Eldar, Y.C. (2019). Deep learning for interference identification: band, training SNR, and sample selection. In Proceedings of the IEEE 20th international workshop on signal processing advances in wireless communications (SPAWC), Cannes, France 1–5

Google Scholar

Zhang, L., Wang, H., and Li, T. (2013). Anti-jamming message-driven frequency hopping-part I: System design. IEEE Trans. Wirel. Commun. 12, 70–79. doi: 10.1109/TWC.2012.120312.111706

CrossRef Full Text | Google Scholar

Zhang, Q., Xiao, J., and Tian, C. (2023). A robust deformed convolutional neural network (CNN) for image denoising. CAAI Trans. Intell. Technol. 8, 331–342. doi: 10.1049/cit2.12110

CrossRef Full Text | Google Scholar

Zhou, K., Zhang, M., Wang, H., and Tan, J. (2022). Ship detection in SAR images based on multi-scale feature extraction and adaptive feature fusion. Remote Sens. 14:755. doi: 10.3390/rs14030755

CrossRef Full Text | Google Scholar

Zivkovic, M., Bacanin, N., Antonijevic, M., Nikolic, B., Kvascev, G., Marjanovic, M., et al. (2022). Hybrid CNN and XGBoost model tuned by modified arithmetic optimization algorithm for COVID-19 early diagnostics from X-ray images. Electronics 3798, 1–30. doi: 10.3390/electronics11223798

CrossRef Full Text | Google Scholar

Keywords: communication interference identification, electronic countermeasures, convolutional neural network, transformer, cross-attention mechanism

Citation: Zhang H, Zhao M, Zhang M, Lin S, Dong Y and Wang H (2023) A combination network of CNN and transformer for interference identification. Front. Comput. Neurosci. 17:1309694. doi: 10.3389/fncom.2023.1309694

Received: 08 October 2023; Accepted: 14 November 2023;
Published: 06 December 2023.

Edited by:

Miodrag Zivkovic, Singidunum University, Serbia

Reviewed by:

Nebojsa Bacanin, Singidunum University, Serbia
Milos Antonijevic, Singidunum University, Serbia

Copyright © 2023 Zhang, Zhao, Zhang, Lin, Dong and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Min Zhang, minzhang@xidian.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.