Next Article in Journal
Forest Aboveground Biomass Estimation Based on Unmanned Aerial Vehicle–Light Detection and Ranging and Machine Learning
Previous Article in Journal
Regulation and Liquid Sensing of Electromagnetically Induced Transparency-like Phenomena Implemented in a SNAP Microresonator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification and Identification of Frequency-Hopping Signals Based on Jacobi Salient Map for Adversarial Sample Attack Approach

1
School of Electronics and Information Engineering, Nanjing University of Information Science and Technology, Nanjing 210044, China
2
Sixty-Third Research Institute, National University of Defense Technology, Nanjing 210007, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(21), 7070; https://doi.org/10.3390/s24217070
Submission received: 11 September 2024 / Revised: 29 October 2024 / Accepted: 31 October 2024 / Published: 2 November 2024
(This article belongs to the Section Communications)

Abstract

:
Frequency-hopping (FH) communication adversarial research is a key area in modern electronic countermeasures. To address the challenge posed by interfering parties that use deep neural networks (DNNs) to classify and identify multiple intercepted FH signals—enabling _targeted interference and degrading communication performance—this paper presents a batch feature point _targetless adversarial sample generation method based on the Jacobi saliency map (BPNT-JSMA). This method builds on the traditional JSMA to generate feature saliency maps, selects the top 8% of salient feature points in batches for perturbation, and increases the perturbation limit to restrict the extreme values of single-point perturbations. Experimental results in a white-box environment show that, compared with the traditional JSMA method, BPNT-JSMA not only maintains a high attack success rate but also enhances attack efficiency and improves the stealthiness of the adversarial samples.

1. Introduction

In the increasingly complex electromagnetic environment, frequency-hopping communication systems, known for their robust anti-interference, anti-fading, and multiple access networking capabilities, play a crucial role not only in civilian communication but also in the defense and military sectors, where they have achieved remarkable success [1]. Consequently, the effective identification and classification of frequency-hopping communication signals has become a key research focus within the field of signal processing. With the rapid advancement of electronic countermeasure technologies, the demands for greater reliability and stability in frequency-hopping communication have intensified, and the inherent limitations of traditional systems are becoming increasingly apparent [2]. In communication countermeasures, reconnaissance equipment can frequently detect and capture multiple frequency-hopping signals originating from various independent radio stations. To gain a tactical advantage on the battlefield, it is essential to classify and identify these mixed signals within a limited reconnaissance time-frame.
Traditional methods for classifying and recognizing frequency-hopping (FH) signals typically rely on manually extracted or self-learned features from the time and frequency domains. In [3], a method is proposed for classifying and identifying single FH stations by extracting the transient characteristics of FH signals. This approach achieves a higher classification and recognition rate for a single FH station compared to classical template matching methods, with an average recognition accuracy of 92.6%. In [4], a method is introduced that combines the K-means clustering algorithm with hop periods to sort non-blind signals across different networks. Building on this, additional parameters such as the direction of arrival, power, and hop time are employed to further sort stations within a network. In [5], the author extracts fingerprint features representing emitter signals and uses a decision tree model within an integrated learning framework to effectively classify and recognize these signals.
In recent years, the successful application of deep learning in fields such as image [6] and voice recognition [7] has prompted researchers to explore its use in wireless communication, particularly for FH signal classification and recognition. For instance, in [8], the author innovatively applies principles from _target detection in image recognition to address the problem of radio signal recognition, significantly enhancing both recognition performance and the intelligence level of signal processing in complex electromagnetic environments. In [9], a three-layer sparse autoencoder is used to extract key features from modulated signals, followed by a backpropagation (BP) neural network to identify communication signal modulation modes. The method achieves a classification accuracy of 98% across a signal-to-noise ratio (SNR) range of 0–20 dB. Lastly, in [10], the author conducts time-frequency analysis of FH signals, extracts contour information to construct a contour map, and uses a convolutional neural network (CNN) to classify and recognize FH signals.
Deep learning technology has demonstrated outstanding performance in the classification and recognition of frequency-hopping signals, significantly enhancing both accuracy and efficiency [10]. In electronic countermeasure environments, adversaries can leverage deep neural networks to classify and identify intercepted frequency-hopping signals, enabling them to implement _targeted jamming strategies. Such strategies pose a serious threat to the information security of communicators. Since 2014, Szegedy [11] and other researchers have highlighted the existence of adversarial samples, which have been shown to induce deep learning models to make incorrect predictions. Given that frequency-hopping signals often contain higher levels of noise and uncertainty, they are particularly susceptible to the influence of adversarial samples when deep learning is applied directly for classification and recognition. As a result, research into the generation of adversarial samples capable of deceiving enemy classifiers holds significant practical value.
Currently, adversarial sample generation methods [12,13,14,15,16,17,18] in the field of electromagnetic signals predominantly adopt global attack strategies. These approaches often fail to adequately account for the local characteristics of the signals when implementing adversarial perturbations, thereby limiting the effectiveness of the attacks. To address these limitations and to generate higher-quality adversarial samples for communication signal attacks, this paper introduces a method for generating adversarial samples for frequency-hopping signal classification, inspired by the Jacobian-based saliency map attack (NT-JSMA) [12], originally developed for image recognition. The proposed method significantly enhances the speed of adversarial sample generation by selectively _targeting key feature points on the frequency-hopping signal feature map for perturbation. Moreover, the method restricts the selection of feature points to ensure that the generated adversarial samples remain more concealed. The main contributions of this paper can be summarized as follows:
  • This experiment demonstrates the limitations of both the conventional gradient attack method and the JSMA method in attacking classification and recognition models for frequency-hopping signals. In response to the non-stationary characteristics of frequency-hopping signals, a new method—BPNT-JSMA—has been proposed to generate adversarial samples specifically tailored for the classification and recognition of such signals.
  • The BPNT-JSMA method generates a feature saliency map of the frequency-hopping signal by computing the Jacobian matrix. It then selects, in batches, the feature points with the highest saliency to introduce perturbations, thereby producing adversarial samples. This approach significantly accelerates the generation of adversarial samples compared to the conventional _targetless JSMA method.
  • The BPNT-JSMA method introduces a clipping function, which is not available in the NT-JSMA method, and adds  L  restriction to ensure that the perturbation values added to the samples do not exceed a certain range, thus enhancing the stealthiness of the generated adversarial samples.
Section 2 of this paper provides a systematic overview of the fundamental concepts of adversarial samples and the underlying principles of adversarial sample generation methods. Section 3 presents a detailed description of the BPNT-JSMA adversarial sample generation method, which is based on salient graphs for batch feature points. This section also offers an in-depth theoretical analysis of the proposed method. In Section 4, experimental validation is carried out. The experimental setup is first described, followed by the introduction of evaluation metrics for the algorithm. Subsequently, a series of experiments are conducted, and the results are thoroughly analyzed. Section 5 concludes the paper by summarizing the findings and discussing potential directions for future research.

2. Related Literature Review

2.1. Basic Concepts of Adversarial Samples

Adversarial samples are crafted by introducing small, often imperceptible perturbations to the original data set. While these modifications may go unnoticed by a human observer, they can significantly alter the model’s output, leading to incorrect predictions. Such samples are highly stealthy and are specifically designed to exploit vulnerabilities in the model, causing it to confidently misclassify the input. The process of generating adversarial samples can be described as follows:
x a d v = x + arg min | | δ | | p
F ( x a d v ) F ( x )
where  x a d v  is the adversarial samples,  δ  is the adversarial perturbation,  p  is the paradigm constraint of the adversarial perturbation  δ F  is the deep learning model, and  x  is the original sample.

2.2. Adversarial Sample Generation Methods

2.2.1. FGSM

Fast Gradient Sign Method (FGSM) [13] and its variants [14,15,16,17] are currently the most efficient and widely used methods for generating adversarial samples, which maximizes the loss function by adding a perturbation in the opposite direction of the gradient to generate adversarial samples that induce the model to produce erroneous outputs. With its implementation, the disadvantage is that its attack mode is global attack, and the generated adversarial samples are insufficient in terms of concealment.
In 2015, Goodfellow et al. [13] proposed the most classical gradient attack algorithm, the Fast Gradient Sign Method (FGSM), based on the assumption of linearity of neural networks in high-dimensional space. The adversarial sample generation algorithm of FGSM is shown in Equation (3).
x a d v = x + ε s i g n ( x L ( x , y ; θ ) )
where  x  is the original sample,  x a d v  is the generated adversarial sample,  y  denotes the correct label of the original sample  x θ  is the parameter of the deep learning model,  ε  denotes the size of the added perturbation,  L ( )  is the loss function,  x  is the derivation of the loss function (the gradient information of the original sample), and  s i g n ( )  denotes the sign function. The most important feature of the FGSM is its high efficiency, which is often used in scenarios that require a large number of generated adversarial samples such as adversarial training. The most notable feature of the FGSM is its efficient operational speed, making it widely used in scenarios that require a large number of generated adversarial samples, such as adversarial training. However, a notable drawback is that the performance of the generated adversarial samples is generally inferior.

2.2.2. I-FGSM

In order to overcome the limitations of FGSM, Kurakin et al. [14] proposed the Basic Iterative Method (BIM) in a follow-up study. BIM is also known as Iterative Fast Gradient Sign Method (I-FGSM), and its core strategy is to change the single-step addition of perturbations in FGSM to a small number of times, i.e., to add small perturbations along the direction of the gradient increase in each iteration, to observe the change in the model’s decision-making, and to return to the adversarial samples once there is an error; the equations are shown in Equations (4) and (5).
x 0 a d v = x
x i + 1 a d v = C l i p x , ε { x i a d v + α s i g n ( x L ( x i a d v , y ; θ ) ) }
where  c l i p { · }  is the cropping function, which ensures that the pixel values of the adversarial samples are in the appropriate range;  x i a d v  is the adversarial samples generated in the  i  iteration, and  α  is the single-step iteration size. Compared with FGSM, I-FGSM can construct more accurate perturbations and improve the quality of the adversarial samples, but the computation amount increases.

2.2.3. MI-FGFSM

Based on the FGSM framework, Dong et al. [15] cleverly integrated the concept of momentum in physics, and then proposed the momentum iterative fast gradient sign method (MI-FGSM). While retaining some gradients in the previous step, they introduced a small amount of gradients generated in the current step to stabilize the update direction and avoid falling into local extreme points. The improvement of this method is to accumulate the velocity vector in the gradient direction by using momentum. Equations are shown in (6) and (7):
g i + 1 = μ · g i + x i a d v J ( x i a d v , y ) | | x i a d v J ( x i a d v , y ) | | 1
x i + 1 a d v = x i a d v + α · sign ( g i + 1 )
First, input  x a d v i  to the classifier  f  to obtain the gradient  x J ( x a d v i , y ) . Use Formula (6) to accumulate the velocity vector in the gradient direction to update  g i + 1 , and then use Formula (7) to update  x a d v i + 1 , and finally generate disturbance. Compared to FGSM and I-FGSM, MI-FGSM generates adversarial samples that are more aggressive and migratable.

2.2.4. PGD

Mardy et al. [18] proposed Projected Gradient Descent (PGD), which, unlike the cropping operation of I-FGSM, limits the size of the perturbation by projecting the results of each iteration onto the domain of pure inputs. Compared with the one-step countermeasure of FGSM, PGD also adopts a ‘small steps, many steps’ strategy to attack by accumulating small perturbations through multiple iterations. PGD uses a uniform random noise initialization to project the gradient instead of limiting the cropping. The attack process is shown in Formula (8).
x t + 1 a d v = p r o j x , ε ( x t a d v + α s i g n ( x J ( x t a d v , y , θ ) ) )
where  p r o j x , ε ( · )  is the projection operation.

2.2.5. JSMA

In order to minimize the number of alterations to the original data points, researchers have introduced the Jacobian-based saliency map attack (JSMA) method [19] and its variants [12]. The methods for salient _target detection only require inputting the original image and iteratively adding perturbations to the salient feature regions to generate adversarial samples without exploiting the gradient information of the model [20], which belongs to the  L 0  (percentage of the number of perturbed data points) limiting attack, i.e., altering as few data points as possible.
JSMA serves as an effective adversarial attack method whose main objective is to mislead or deceive classification models such as deep neural networks for image recognition. Compared to other widely used white-box attack methods, JSMA is characterized by the innovative introduction of saliency graphs. Originally conceived to visualize the prediction process of classification models [21], in reference [19], Papernot et al. first proposed a saliency map for adversarial attacks, from which the information needed by a neural network classifier to classify a _target on the sample data can be obtained. If an attacker wants to misclassify the original datum  x  so that it classifies  t l a b e l ( x ) , then it needs to strengthen all the features of  t l a b e l ( x )  while shrinking the features of  t = l a b e l ( x )  until it reaches the condition that it can fulfill the attack goal. The salient graph is shown in Formula (9).
S ( x ( i ) , t ) = 0 , F t ( X ) X i or j t F j ( X ) X i ( F t ( X ) X i ) j t F j ( X ) X i ,   else
where  i  is the input feature. First, the negative derivative classification of the specified _target and the positive derivative classification of other _targets are excluded.  F t ( X ) X i  is positive at this time, so that when the feature pixel  x i  increases, the probability value of  F t ( X )  also increases, and  j t F j ( X ) X i  needs to be negative or unchanged, and then  S ( x ( t ) , c )  is obtained by using all the remaining derivative components. Through  S ( x ( t ) , c ) , all the input features can be easily compared, and the _target can be attacked by continuously increasing these features.
JSMA enables the generation of adversarial samples that are classified as a specified _target category by altering the pixel characteristics of the attack _target to achieve an extreme value. In reference [12], the author introduced NT-JSMA, a variant of JSMA that does not require a predefined _target class. NT-JSMA eliminates the dependency between the _target and the category, allowing for un_targeted attacks. However, due to the complexity of calculating the Jacobian matrix and the significant resource consumption involved, this method results in slow generation of adversarial samples. Currently, research on the JSMA algorithm primarily focuses on significantly accelerating the generation of adversarial samples while minimizing the number of pixel alterations in the attack _target to achieve the desired attack effect.

3. BPNT-JSMA Method for Batch Feature Point Non-_target Attack Based on Jacobi Saliency Map

3.1. Basic Idea

To address the challenges associated with the JSMA method in generating adversarial samples for frequency-hopping signals, this paper proposes the BPNT-JSMA method. First, obtain the parameter information of the trained DNN model  F , input the frequency-hopping signal sample  x  to the model  F , and calculate the Jacobian matrix  J  for all categories. Subsequently, generate the feature saliency map according to the  c  of other categories, combined with the Jacobian matrix. Finally, select a subset of feature points exhibiting the strongest saliency in batches, according to the feature saliency map of the other categories  c , to introduce perturbations and obtain  x a d v . If the perturbations result in the model  F  recognizing the sample  x a d v  as belonging to category  c , the generation of adversarial samples is considered successful, and the adversarial samples are returned. If not, apply iterative perturbations to the feature points and evaluate whether the perturbation at each point reaches the predefined upper limit  ε . If the upper limit is reached, select the next most significant feature point and repeat the process until an adversarial sample is successfully generated. If the available feature points in the saliency map allow the model  F  to recognize them as adversarial samples  x a d v  of other categories before they are exhausted, the generation of adversarial samples will be successful. Conversely, if the feature points are depleted without successful generation, the attempt will be deemed a failure. The specific attack process is shown in Figure 1.

3.2. Description of the Attack Method

3.2.1. Jacobian Matrix Calculation

In deep learning, the Jacobian matrix is commonly utilized to compute gradients within the backpropagation algorithm and to update model parameters. By calculating the Jacobian matrix of the output variable with respect to the input variable, we can accurately assess the impact of the input variable on the output variable. A larger value of the Jacobian matrix indicates a greater influence of the input on the output. Let the model  F  be an  n  classification model, and the dimension of the input variable  x  is  a × b ; then, obtain the score of the model  F  with respect to the input  x  for each category as the starting point, and calculate the Jacobian matrix of  x  for  n  categories according to Formula (10).
J ( F r ( x ) ) = F r ( x ) x = F r ( x ) x i i [ 1 , a × b ]
where the result of  J ( F r ( x ) )  is the contribution of the location data point to  F r ( x ) , F r ( x )  represents the score of input  x  on the category  r x i  represents the  i  feature point in  x , and  x  has  a × b  feature points in total.
A deep neural network is a model composed of multiple layers of neural networks. Each layer aims to extract useful features or information from the input data and subsequently passes this information to the next layer for higher-level abstraction or processing. Most of these layers, particularly the intermediate layers, are referred to as “hidden layers”. Their inputs and outputs are not directly visible or interpretable; instead, they are automatically adjusted and optimized through the network’s learning process. Consequently, it is essential to utilize the chain rule of functions to calculate the Jacobian matrix, as demonstrated in Formulas (11) and (12).
F r ( x ) x = F r ( x ) H ( x ) H ( x ) x
H k ( x ) x i = [ f k ( W k H k 1 + b k ) x i ]
where  H k ( x )  represents the  k  hidden layer,  w  represents the weight of the  k 1  layer output as the  k  layer input,  f k  represents the activation function, and  b k  represents the offset.

3.2.2. Saliency Map Generation

According to the aforementioned formula, the Jacobian matrix  J  for all categories can be derived. Subsequently, it is necessary to compute and generate the saliency map  S  for all categories except the true category. The generation process of the saliency map is divided into two directions: perturbation addition and perturbation reduction. In the direction of perturbation addition, larger values of significant feature points correspond to higher classification scores for the _target class, while the scores for the relative non-_target class decrease accordingly. In this scenario, the influence of the feature points on the classification result of the _target class is positive. Conversely, in the direction of perturbation reduction, larger values of significant feature points result in lower classification scores for the _target class and higher scores for the relative non-_target class. Here, the influence of the feature points on the classification result of the _target class is negative. Considering the effects of these two directions, Formula (13) can be derived. The key feature points that enable  x  to be classified as other class  c  can be obtained through Formula (13), which generates the feature salient graph  S ( x , c )  of  x  for  c .
S ( x , c ) [ i ] = J i c ( x ) | r c J i r ( x ) | , J i c ( x ) > 0 r c J i r ( x ) < 0 0 , J i c ( x ) = 0 r c J i r ( x ) = 0 J i c ( x ) r c J i r ( x ) , J i c ( x ) < 0 r c J i r ( x ) > 0
where  i  represents the  i -th feature point of input  x , and  J i c ( x )  represents the value of the Jacobian matrix of  x ’s score in other  c  categories to the  i  feature point.  J i c ( x ) > 0  means that the contribution of feature point  i  to the classification of  x  as  c  is positive, that is, increasing the value of  i  will increase the score of  x  in the category  c r c J i r ( x ) < 0 , which means that the total contribution of  i  to the correct category is negative; that is, increasing the value of  i  will reduce the score of the correct category.  J i c ( x ) > 0  and  r c J i r ( x ) < 0  indicate that the contribution of the modified point to the classification of  c  is positive;  J i c ( x ) < 0  and  r c J i r ( x ) > 0  show that the increase in the  i  value would lead to a decrease in the  c  score.
In order to increase the probability of classifying  x  as  c , for the characteristic points of  J i c ( x ) > 0  and  r c J i r ( x ) < 0 , the direction of adding disturbance value is positive. For the characteristic points with  J i c ( x ) < 0  and  r c J i r ( x ) > 0 , the direction of adding disturbance value is negative. By performing the above calculation for each feature point in  x , the feature saliency map  S ( x , c )  can be obtained.

3.2.3. Adversarial Sample Generation

After obtaining the saliency map for category  c , the traditional JSMA method iteratively selects the pair of points with the highest saliency  ( p 1 , p 2 )  to introduce perturbations and generate adversarial samples. To optimize this process and enhance the generation rate of adversarial samples, this paper proposes an improved method for selecting feature points within the JSMA framework.
The proposed method begins by sorting all feature points in the saliency map based on their saliency values. It then selects the top 8% (which is an empirical value) of feature points, which represent the highest saliency values as the _target set to add a one-step disturbance  θ . The selection of feature points is shown in Formula (14).
( p 1 , , p n ) = [ s o r t ( S ( x , t ) [ i ] ) ] × Top 8 %
Add the disturbed feature points to the original sample  x  to obtain the adversarial sample  x a d v . In addition, considering that JSMA is limited by  L 0 , it will be easy to be recognized by human eyes if the disturbance added to the feature points is too large. Therefore, in order to enhance the concealment, this paper introduces  L  to limit  ε , that is, the maximum disturbance added by each significant point does not exceed  ε , as shown in Formula (15). Then, the adversarial sample is generated iteratively through Formula (16).
| | x i a d v x i | | ε
x n a d v = C l i p ε [ ( x i , n 1 a d v ) i { p 1 , p n } + θ ]
where  x n a d v  is the countermeasure sample generated through  n  iterations, and  C l i p ( · )  is the clipping function to limit the disturbance size. The overall process of the BPNT-JSMA algorithm is shown in Algorithm 1.
Algorithm 1 BPNT-JSMA adversarial sample attacks
Input: Normal signal sample  x , DNN model  F , Other types of  c , Single step disturbance size  θ , Single data point disturbance limit  ε , Total disturbance limit  λ , Iteration number  n .
Output: Adversarial samples  x a d v .
1:   Input the sample  x  into the model  F , and return the score  r  of each category  F r ( x ) ( r [ 1 , n ] )  
2:   Calculate the Jacobian matrix  J ( F r ( x ) )  of each category according to  F r ( x )  and Formula (9)
3:   Calculate the characteristic saliency diagram  S ( x , c )  of other categories of  c  according to Formula (12) and  J ( F r ( x ) )  
4:   While saliency graph  S ( x , c )  is not empty
5:     For  ( i = 0 , i < m , i + + )  
6:       Select the significant feature point  { p 1 , p n }  from  S ( x , c )  according to Formula (13) to add the disturbance  θ  
7:       Generate adversarial samples according to Equation (15)
8:       If  F ( x a d v ) = c  
9:         Return adversarial sample  x a d v  
10:     Else
11:       If single point disturbance  θ < ε  and total disturbance  < λ  
12:        Continue
13:     Else
14:        Failure to generate adversarial samples
15:        Break for
16:   End for
17:  End while

4. Experimental Results and Analysis

4.1. Experimental Setup

4.1.1. Data Set

In this study, four frequency-hopping signals are generated as experimental data samples through Matlab R2020a software simulation. In order to better verify the classification recognition accuracy under multiple frequency-hopping signals, the modulation mode of the frequency-hopping signal samples is BPSK, the number of frequency-hopping points is 64, the bandwidth of the frequency-hopping band is 1.6 MHz, the signal sampling rate is fs = 12.8 MHz, the Gaussian white noise is added at the same time to simulate the actual channel conditions, and the range of the signal-to-noise ratio is from −20 dB to 18 dB, with a step of 2 dB. The data set generates 1600 samples at each signal-to-noise ratio, that is, 400 samples for each frequency-hopping signal, and a total of 32,000 samples are generated. In order to ensure the effectiveness of model training and validation, the data set is processed by a random partition strategy: 70% as the training set, 20% as the validation set, and the remaining 10% as the test set. The parameters of frequency-hopping signals in the data set are shown in Table 1.

4.1.2. DNN Model

The performance of the _target model in adversarial attacks is crucial in reflecting the effectiveness of the attack. If the _target model itself has low accuracy in recognition, then the attack loses its significance. Given that VTCNN2 is a convolutional neural network specifically designed for processing time-series data, it can better capture temporal variations compared to a traditional CNN, making it more suitable for handling frequency-hopping signals. The CLDNN, as the only network model that combines CNNs and Long Short-Term Memory (LSTM) networks, is capable of simultaneously considering both temporal and spatial information in signals. It retains information from previous inputs when processing sequential data, thereby maintaining long-range dependencies and avoiding issues such as vanishing or exploding gradients, which are commonly encountered by traditional network architectures such as DNNs and CNNs when dealing with long sequences. This gives CLDNN a significant advantage in the recognition and adversarial generation of frequency-hopping signals. Meanwhile, ResNet employs residual connections, enabling more efficient training of deeper networks and addressing the vanishing gradient problem that can arise during the training of deep networks. Considering the differences in features and parameters between frequency-hopping signals and images, this paper adopts three classification models: VTCNN2 [22], CLDNN [23], and ResNet [24]. Each model is trained for 500 epochs with an initial learning rate of 0.001, and the learning rate is halved if the validation loss does not decrease for five consecutive evaluations.
To ensure that the data set of multi-frequency-hopping signals is compatible with the DNN model, this paper optimizes and adjusts the network structure of the VTCNN2 model for the specified data set. The model consists of one input layer, four convolutional layers, four pooling layers, four dropout layers, two fully connected layers, and one output layer. By incorporating multiple convolutional layers to extract features at different levels, the model is better equipped to understand and recognize complex data content. The addition of pooling layers reduces computational complexity and accelerates the training speed of the model. Furthermore, the inclusion of dropout layers effectively prevents overfitting and enhances the model’s generalization ability. The specific architecture of the VTCNN2 model is presented in Table 2.

4.1.3. Evaluation Index

To effectively evaluate the algorithm proposed in this paper, the evaluation method outlined in reference [25] is enhanced and examined from three dimensions: the concealment of adversarial samples, the effectiveness of the attack, and the efficiency of the attack.
  • Attack effectiveness
This paper divides attack effectiveness into three dimensions: attack success rate (ASR), Average Confidence of Adversarial Class (ACAC) and Average Confidence of True Class (ACTC).
From the perspective of countermeasure samples, the ASR is defined as the ratio of the number of countermeasure samples that successfully induce the _target model to make incorrect predictions to the total number of samples. This metric directly reflects the effectiveness of the attack method, with a value range between 0 and 1. A value closer to 1 indicates a higher attack success rate. If the total number of test samples is  M  and the number of samples of the successful attack model is  N , the formula for calculating the ASR is
ASR = N M × 100 %
From the perspective of the decision-making mechanism, the average confidence of the adversarial category refers to the mean confidence level of the _target model across all potential categories, excluding the true category, in relation to a batch of adversarial samples that have successfully executed an attack. This value ranges from 0 to 1; a higher value indicates greater confidence of the model in classifying the sample into incorrect categories, thereby reflecting a more effective attack. Let the confidence of the attack category for a single sample be denoted as  a , and let the total number of samples be  M . Then, the average confidence of the adversarial category, denoted as ACAC, is defined as
ACAC = ( 1 M 1 M a ) × 100 %
Similarly, the average confidence of the real category refers to the mean confidence of the model regarding the true category of a batch of successfully attacked adversarial samples, with a value range between 0 and 1. A higher value indicates that the model has greater confidence in correctly identifying the sample as belonging to the real category, suggesting a weaker attack effect of the adversarial samples. Let  b  represent the confidence of the model in the correct category for a single sample, and let  M  denote the total number of samples. Then, the ACTC can be expressed as follows:
ACTC = ( 1 M 1 M b ) × 100 %
2.
Attack efficiency
When evaluating attack efficiency, the ATC for generating each adversarial sample serves as the key indicator. This metric directly reflects the efficiency of the attack; a smaller ATC value indicates that more adversarial samples can be generated within the same time-frame. If the number of successful attack samples is denoted as  N  and the total time taken is  T , then the ATC can be calculated as follows:
ATC = T N
3.
Adversarial sample concealment
The concealment of adversarial samples is quantified by the difference between adversarial samples and original samples and is measured using structural similarity (SSIM). The main parameters of SSIM include the brightness  l ( x , x a d v ) , contrast  c ( x , x a d v ) , and structure  s ( x , x a d v )  of two kinds of pictures. The frequency-hopping signal data used in the experiment can be regarded as a single channel image, and the brightness  l  and contrast  c  results are both 1, so only the structural similarity index SSIM needs to be calculated. The calculation formula is as follows:
SSIM ( x , x a d v ) = [ l ( x , x a d v ) ] α [ c ( x , x a d v ) ] β [ s ( x , x a d v ) ] γ
s ( x , x a d v ) = σ x 1 x 2 + C σ x 1 σ x 2 + C
where  σ x 1  and  σ x 2  are the standard deviations of  x 1  and  x 2 , respectively,  σ x 1 x 2  is the covariance of  x 1  and  x 2 , and  C  is a constant to avoid the denominator approaching zero. The SSIM ranges from 0 to 1, with values closer to 1 indicating greater similarity between the adversarial sample and the original sample.

4.2. Experimental Results and Analysis

4.2.1. _target Model Training Settings

Considering the characteristics of the signal samples, as well as the model parameters and recognition performance under normal conditions, we selected three DNN models—VTCNN2, CLDNN, and ResNet—as the neural network models for evaluating the performance of the comparison algorithms. In the experimental results presented in Figure 2, we assessed the classification recognition accuracy of these three models under non-attack conditions. It was observed that as the SNR gradually increased, the recognition accuracy of all three models exhibited a positive upward trend and eventually stabilized. Among them, the VTCNN2 and CLDNN models demonstrated superior recognition performance, significantly outperforming the ResNet model. In scenarios with a low signal-to-noise ratio, the noise power is considerably greater than the power of the signal itself, leading to severe distortion of the signal waveform. This distortion adversely affects the model’s recognition capability, resulting in generally low accuracy in recognition.
When the recognition rate of the model for the signal under low-SNR conditions is insufficient, it becomes challenging to effectively evaluate the attack’s impact and verify the effectiveness of the attack method. In contrast, when the SNR exceeds 2 dB, the recognition accuracy of both the VTCNN2 and CLDNN models utilized in this experiment remains stable above 95%. Additionally, the ResNet model achieves a recognition accuracy of approximately 80%, which adequately meets the _target model’s requirements for resilience against attacks. The recognition accuracy of the three models under normal conditions is shown in Figure 2.

4.2.2. Analysis of Experimental Results

In order to better compare and analyze the methods in this paper, FGSM [9], I-FGSM [10], MI-FGSM [11], PGD [14] and NT-JSMA [8] methods are selected, and these methods are set as non-_target-attack mode. The experimental results of the three models are shown in Table 3, Table 4 and Table 5.

Attack Success Rate (ASR)

The attack success rate can be quantified by the recognition accuracy of the models at different signal-to-noise ratios; a higher attack success rate corresponds to a lower accuracy of the model. The recognition accuracy graphs of the three models at different signal-to-noise ratios are shown in Figure 3.
As shown in Table 3, Table 4 and Table 5, regardless of the attack method employed, the success rate of the three models exceeds 59%. Among these methods, JSMA demonstrates the highest attack efficacy, with a success rate surpassing 83%. Based on Figure 3, it is evident that from −20 dB to 18 dB, as the SNR increases, the model’s recognition accuracy tends to improve, stabilizing around 8 dB. The comparison of the algorithms reveals the following: (1) The effectiveness of the same attack varies across different models, with FGSM performing the worst. A possible reason is that FGSM, as a single-step attack, may produce incorrect gradient directions for generating adversarial perturbations in highly nonlinear deep models. In Figure 3a, at 8 dB, FGSM only reduces the model’s recognition rate by 35%. (2) I-FGSM performs better than FGSM, reducing the VTCNN2 model’s recognition accuracy by 50% at 10 dB. Although MI-FGSM theoretically improves the attack by introducing a momentum term to refine the gradients, its effectiveness is nearly identical to that of I-FGSM, as seen in the figure. In some cases, such as with the CLDNN model, MI-FGSM even underperforms I-FGSM, possibly due to structural differences between models affecting the output. (3) The PGD algorithm combines gradient information with projection operations, primarily utilizing gradient information to iteratively adjust the samples while ensuring that the adversarial samples remain within a certain allowable range after each step. This effectively increases the intensity of the attack. As shown in Figure 3a,b, the model’s recognition rate under PGD attack decreases to below 40%. (4) The algorithm proposed in this paper generates adversarial samples by adding perturbations to key feature points based on the salient feature maps of the samples. To enhance efficiency, a method of batch selecting the most salient points is employed, addressing the issue of weak attack capabilities on certain signals due to the limited information content of the signal samples. This approach reduces ineffective perturbations, resulting in more refined perturbations and improved attack effectiveness. At 10 dB, the recognition rate of the VTCNN2 model can be reduced by approximately 75%, which represents an improvement of 25% over FGSM and 15% over MI-FGSM.

Average Confidence of True Class (ACTC) and Average Confidence of Adversarial Class (ACAC)

Figure 4 is a comparison diagram of ACTC and ACAC drawn according to Table 3, Table 4 and Table 5. It can be seen from Figure 4a that the average confidence of any attack method in the correct category is less than 5%, and the values of NT-JSMA and BPNT-JSMA are less than 1%, which shows that in terms of attack effect, the attack based on a saliency graph is stronger than the attack based on gradient. It can be seen from Figure 4b that the average confidence of the confrontation category is above 45%, while the values of the NT-JSMA and BPNT-JSMA attack methods can reach 79%, indicating that the BPNT-JSMA method proposed in this paper will not reduce the attack ability of the original JSMA method. Compared with the other four attack methods, this method still has strong inducement and makes the model decision wrong.

Average Time Consumption (ATC)

The consumption time required for generating a single countermeasure sample is illustrated in Figure 5. As shown in the figure, despite the earlier analysis indicating that the FGSM attack exhibits the poorest performance, it generates samples the fastest due to its nature as a non-iterative, one-step attack. In contrast, FGSM, I-FGSM, MI-FGSM, and PGD are all iterative attacks with the same number of iterations, resulting in longer running times; however, the average time required to generate a single sample among these methods is not significantly different. The JSMA method requires the calculation of the Jacobian matrix for each pixel of the input sample and necessitates multiple selections of the most resistant pixel for perturbation, making it the most time-consuming attack method. Although the BPNT-JSMA method proposed in this paper has a longer ATC compared to FGSM, I-FGSM, MI-FGSM, and PGD, it is significantly faster than the original JSMA method for selecting feature points to attack, as it selects feature points in batches after calculating the saliency map.

Structural Similarity (SSIM)

When adversarial samples exhibit a high similarity to the original data in the feature space, the model is often more easily misled. Additionally, the similarity between samples reflects the model’s vulnerability in specific regions; a high similarity indicates that even minor perturbations can lead to significant output changes, thereby increasing the success rate of the attack. It can be seen from Table 3, Table 4 and Table 5 that the structural similarities between the adversarial samples generated by the FGSM, I-FGSM, MI-FGSM, and PGD attack methods and the original samples are approximately 50%, 60%, 70%, and 75%, respectively. In contrast, the structural similarity between the adversarial samples generated by the NT-JSMA method and the original samples is about 83%; it is positively correlated with the variation in attack success rate. This discrepancy can be attributed to the fact that gradient-based attacks focus more on global gradient information, applying perturbations in the direction of the gradient. In contrast, the JSMA emphasizes local pixel information, allowing for precise localization of key pixels. Consequently, the attack _target can be achieved with smaller perturbations, resulting in adversarial samples that exhibit higher similarity to the original samples. Notably, the proposed BPNT-JSMA method achieves a maximum structural similarity of 92% when generating adversarial samples, which is significantly better than that of other methods. This demonstrates that the BPNT-JSMA method can produce adversarial samples that are both highly similar to and less detectable than the original samples while effectively maintaining the attack’s efficacy. See Figure 6.

5. Conclusions

To address the security issue where deep learning models for frequency-hopping signal classification and recognition are vulnerable to non-_targeted attacks, this paper proposes a novel BPNT-JSMA adversarial sample generation method. In contrast to the traditional NT-JSMA anti-attack method, the proposed approach batch-selects the key feature points that most significantly affect the model’s recognition results, based on the saliency map of frequency-hopping signals, to introduce perturbations. Additionally, it incorporates a perturbation limit, which is absent in the NT-JSMA method, to constrain the magnitude of the perturbations. Experimental results demonstrate that the proposed method significantly improves the speed of adversarial sample generation while enhancing the concealment of the adversarial samples, all without compromising the success rate of the non-_targeted JSMA on DNN models. Future research can explore the following directions: (1) The experiments in this paper were conducted in a white-box environment without integrating black-box attacks. In real-world battlefield scenarios, information related to the _target model is often unknown to communicators. Therefore, future research could focus on attack methods in black-box environments. (2) While the proposed method shows significant improvements in generation efficiency compared to the original JSMA method, there is still a performance gap compared to gradient-based attack methods. Further efforts to enhance the generation efficiency of adversarial samples while maintaining attack success rates could be a promising avenue for future research.

Author Contributions

Conceptualization, Y.Z. and Y.L.; methodology, Y.Z., Y.L. and T.W.; software, Y.Z. and Y.L.; validation, Y.Z., Y.L. and T.W.; formal analysis, Y.Z., Y.L. and T.W.; investigation, Y.Z. and Y.L.; resources, Y.Z.; data curation, Y.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, Y.Z. and Y.L.; visualization, Y.Z.; supervision, Y.Z.; project administration, Y.Z.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by The National Natural Science Foundation of China grant number 2022-SKJJ-B-112.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are not available due to privacy.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, J.; Yu, L.; Hou, C.B.; Zhang, J.; Lin, J.X. Frequency hopping signal modulation recognition based on time-frequency characteristics. J. Terahertz Sci. Electron. Inf. 2022, 20, 40–46. [Google Scholar]
  2. Ji, L.S. Analysis of anti-jamming technology for ultra short wave radio communication. Electron. Commun. Comput. Sci. 2023, 5, 74–76. [Google Scholar]
  3. Gu, C.M.; Wang, L.W. Individual identification of frequency hopping radio stations based on instantaneous features in the frequency domain. Comput. Eng. Appl. 2013, 49, 223–226. [Google Scholar]
  4. Yue, F.W. Research and Simulation Implementation of Sorting Method for Frequency Hopping Communication Network Stations. Ph.D. Thesis, University of Electronic Science and Technology of China, Chengdu, China, 2010. [Google Scholar]
  5. Zhang, X.Q. A Study on the Application of Machine Learning in the Fingerprinting of Radiation Source Signals. Master’s Thesis, University of Electronic Science and Technology of China, Chengdu, China, 2018. [Google Scholar]
  6. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  7. GUI, T.; Xi, Z.H.; Zheng, R. Review of research on robustness of natural language processing based on deep learning. Acta Comput. Sci. 2024, 47, 90–112. [Google Scholar]
  8. Lin, X.T.; Zhang, L.; Wu, Z.Q. Modulation recognition method based on convolutional neural network with cyclic spectrograms. J. Terahertz Sci. Electron. Inf. Technol. 2021, 19, 617–622. [Google Scholar]
  9. Yao, Y.C.; Peng, H. Deep Learning Based Automatic Modulation Recognition of Communication Signals. Appl. Electron. Tech. 2019, 45, 12–15. [Google Scholar]
  10. Lv, G.P.; Xie, Y.L. Deep learning based frequency hopping signal recognition. Telecommun. Eng. 2020, 60, 1142–1147. [Google Scholar]
  11. Szegedy, C.; Zaremba, W.; Sutskever, I. Intriguing properties of neural networks. In Proceedings of the 2nd International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2013; pp. 1–10. [Google Scholar]
  12. Wiyatno, R.; Xu, A. Maximal jacobian-based saliency map attack. arXiv 2018, arXiv:1808.07945. [Google Scholar]
  13. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  14. Kurakin, A.; Goodfellow, I.J.; Bengio, S. Adversarial examples in the physical world. In Artificial Intelligence Safety and Security; Chapman and Hall/CRC: London, UK, 2018; pp. 99–112. [Google Scholar]
  15. Dong, Y.; Liao, F.; Pang, T.; Su, H.; Zhu, J.; Hu, X.; Li, J. Boosting adversarial attacks with momentum. In Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 9185–9193. [Google Scholar]
  16. Zhao, H.; Lin, Y.; Gao, S.; Yu, S. Evaluating and improving adversarial attacks on DNN-based modulation recognition. In Proceedings of the GLOBECOM 2020—2020 IEEE Global Communications Conference, Taipei, Taiwan, 7–11 December 2020; pp. 1–5. [Google Scholar]
  17. Wang, C.; Wei, X.L.; Tian, Q. Feature Gradient Based Deep Network Against Attacks for Modulation Recognition. Comput. Sci. 2021, 48, 25–32. [Google Scholar]
  18. Madry, A.; Makelov, A.; Schmidt, L. Towards Deep Learning Models Resistant to Adversarial Attacks. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018; Available online: https://openreview.net/forum?id=rJzIBfZAb (accessed on 7 December 2023).
  19. Papernot, N.; McDaniel, P.; Jha, S.; Fredrikson, M.; Celik, Z.B.; Swami, A. The limitations of deep learning in adversarial settings. In Proceedings of the 2016 IEEE European symposium on security and privacy (EuroS&P), Saarbrücken, Germany, 21–24 March 2016; pp. 372–387. [Google Scholar]
  20. Li, Z.M.; Wang, J.D.; Hou, J.Z. A significant region optimisation based approach to adversarial samples attacks. Comput. Eng. 2023, 49, 246–255. [Google Scholar]
  21. Christian, S.; Wojciech, Z.; Ilya, S. Intriguing Properties of Neural Networks. arXiv 2021, arXiv:1312.6199. [Google Scholar]
  22. Sadeghi, M.; Larsson, E.G. Adversarial attacks on deep-learning based radio signal classification. IEEE Wirel. Commun. Lett. 2019, 8, 213–216. [Google Scholar] [CrossRef]
  23. Sainath, T.N.; Vinyals, O.; Senior, A.; Sak, H. Convolutional, long short-term memory, fully connected deep neural networks. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane, Australia, 19–24 April 2015; pp. 4580–4584. [Google Scholar]
  24. O’Shea, T.J.; Roy, T.; Clancy, T.C. Over-the-air deep learning based radio signal classification. IEEE J. Sel. Top. Signal Process. 2018, 12, 168–179. [Google Scholar] [CrossRef]
  25. Zhou, X.; Zhang, J.; Li, N.A. A salient graph based method for electromagnetic signal adversarial samples generation. Acta Electron. Sin. 2023, 51, 1917–1928. [Google Scholar]
Figure 1. BPNT-JSMA attack flow chart.
Figure 1. BPNT-JSMA attack flow chart.
Sensors 24 07070 g001
Figure 2. Recognition accuracy of the three models on clean data.
Figure 2. Recognition accuracy of the three models on clean data.
Sensors 24 07070 g002
Figure 3. Recognition accuracy of the three models.
Figure 3. Recognition accuracy of the three models.
Sensors 24 07070 g003aSensors 24 07070 g003b
Figure 4. Comparison of ACTC and ACAC in three models.
Figure 4. Comparison of ACTC and ACAC in three models.
Sensors 24 07070 g004
Figure 5. Generation time of single countermeasure sample.
Figure 5. Generation time of single countermeasure sample.
Sensors 24 07070 g005
Figure 6. Structural similarity of countermeasure samples.
Figure 6. Structural similarity of countermeasure samples.
Sensors 24 07070 g006
Table 1. Simulation parameters of frequency-hopping signals.
Table 1. Simulation parameters of frequency-hopping signals.
Signal TypeHopping SpeedCarrier FrequencyLabel
Signal 15000–1.60
Signal 25001.6–3.21
Signal 310000–1.62
Signal 410001.6–3.23
Table 2. VTCNN2 model.
Table 2. VTCNN2 model.
LayersOutput Shape
Conv2D(512,512)
Maxpooling2D(256,512)
Dropout(256,512)
Conv2D(256,512)
Maxpooling2D(128,512)
Dropout(128,512)
Conv2D(128,128)
Maxpooling2D(64,128)
Dropout(64,128)
Conv2D(64,128)
Maxpooling2D(32,128)
Dropout(32,128)
Flatten4096
Dense256
Dense4
Table 3. VTCNN2 model results.
Table 3. VTCNN2 model results.
ASR/%ACTC/%ACAC/%ATC/sSSIM/%
FGSM60.094.1348.300.001757.87
I-FGSM69.312.5658.260.011664.49
MI-FGSM68.442.7057.270.012570.10
PGD74.561.8363.980.012775.03
NT-JSMA83.330.8579.370.106181.34
BPNT-JSMA84.760.7280.140.070688.12
Table 4. CLDNN model results.
Table 4. CLDNN model results.
ASR/%ACTC/%ACAC/%ATC/sSSIM/%
FGSM59.534.1848.310.003458.63
I-FGSM72.692.0163.180.022763.62
MI-FGSM70.152.3560.590.024470.28
PGD78.441.3069.840.025276.96
NT-JSMA84.590.6981.400.145183.85
BPNT-JSMA85.490.6182.530.089392.46
Table 5. ResNet model results.
Table 5. ResNet model results.
ASR/%ACTC/%ACAC/%ATC/sSSIM/%
FGSM62.074.1745.010.001551.30
I-FGSM71.972.4356.680.010360.12
MI-FGSM71.492.4956.320.010367.06
PGD77.901.5963.790.011075.72
NT-JSMA85.690.7480.770.264483.66
BPNT-JSMA86.030.7980.090.199590.72
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, Y.; Li, Y.; Wei, T. Classification and Identification of Frequency-Hopping Signals Based on Jacobi Salient Map for Adversarial Sample Attack Approach. Sensors 2024, 24, 7070. https://doi.org/10.3390/s24217070

AMA Style

Zhu Y, Li Y, Wei T. Classification and Identification of Frequency-Hopping Signals Based on Jacobi Salient Map for Adversarial Sample Attack Approach. Sensors. 2024; 24(21):7070. https://doi.org/10.3390/s24217070

Chicago/Turabian Style

Zhu, Yanhan, Yong Li, and Tianyi Wei. 2024. "Classification and Identification of Frequency-Hopping Signals Based on Jacobi Salient Map for Adversarial Sample Attack Approach" Sensors 24, no. 21: 7070. https://doi.org/10.3390/s24217070

APA Style

Zhu, Y., Li, Y., & Wei, T. (2024). Classification and Identification of Frequency-Hopping Signals Based on Jacobi Salient Map for Adversarial Sample Attack Approach. Sensors, 24(21), 7070. https://doi.org/10.3390/s24217070

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop
  NODES
admin 3
Association 2
Idea 3
idea 3
innovation 2
INTERN 34
Note 16
Project 7
twitter 1
Verify 2