Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
A First-Out Alarm Detection Method via Association Rule Mining and Correlation Analysis
Previous Article in Journal
Novel Entropy-Based Metrics for Long-Term Atrial Fibrillation Recurrence Prediction Following Surgical Ablation: Insights from Preoperative Electrocardiographic Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fine-Grained Radio Frequency Fingerprint Recognition Network Based on Attention Mechanism

School of Electronics and Communication Engineering, Sun Yat-sen University, Shenzhen 518107, China
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(1), 29; https://doi.org/10.3390/e26010029
Submission received: 9 November 2023 / Revised: 22 December 2023 / Accepted: 26 December 2023 / Published: 27 December 2023
(This article belongs to the Section Signal and Data Analysis)

Abstract

:
With the rapid development of the internet of things (IoT), hundreds of millions of IoT devices, such as smart home appliances, intelligent-connected vehicles, and wearable devices, have been connected to the network. The open nature of IoT makes it vulnerable to cybersecurity threats. Traditional cryptography-based encryption methods are not suitable for IoT due to their complexity and high communication overhead requirements. By contrast, RF-fingerprint-based recognition is promising because it is rooted in the inherent non-reproducible hardware defects of the transmitter. However, it still faces the challenges of low inter-class variation and large intra-class variation among RF fingerprints. Inspired by fine-grained recognition in computer vision, we propose a fine-grained RF fingerprint recognition network (FGRFNet) in this article. The network consists of a top-down feature pathway hierarchy to generate pyramidal features, attention modules to locate discriminative regions, and a fusion module to adaptively integrate features from different scales. Experiments demonstrate that the proposed FGRFNet achieves recognition accuracies of 89.8% on 100 ADS-B devices, 99.5% on 54 Zigbee devices, and 83.0% on 25 LoRa devices.

1. Introduction

The Internet of things (IoT) [1] consists of many context-aware products and technologies, ranging from analog and digital sensors to global positioning systems (GPSs), radio frequency identification devices (RFIDs), near field communication (NFC) sensors, weather detectors and emergency alarms, etc. The sector has been growing rapidly and is greatly changing our daily lives and industries.
Compared with a traditional network, decentralization and heterogeneity are the biggest features of IoT. These also introduce security issues, such as passive eavesdropping, identity spoofing attacks, replay attacks, and so on. Authentication is gaining increased attention, especially in the context of wireless communication within the IoT, as the open nature of the wireless medium exposes additional security vulnerabilities [2].
Traditional authentication employs cryptography-based methods to verify whether a message is generated by a trusted source or not [3]. Generally, these methods require high communication overhead and complexity at the application layer due to complicated operations such as encryption and decryption, key sharing, and management [4]. Thus, they are not affordable for common IoT devices with limited computation and power resources. By contrast, physical-layer authentication is believed to be a kind of secure and terminal cost-free method for IoT systems.
Physical-layer identification is the process of fingerprinting a wireless device by extracting features of hardware imperfections [5,6]. These hardware imperfections are referred to as RF fingerprints, which are, similar to human fingerprints, intrinsic and unique. Moreover, they are independent of the modulation mode, remain unaffected by the environment, and introduce subtle differences in signals. Even for wireless devices produced by the same manufacturer with the same components in the same batch, differences in RF fingerprints still exist.
In the past few decades, many efforts have been made for RF-fingerprint-based recognition, and they can be classified into two categories: machine-learning-based [7,8,9,10,11,12,13] and deep-learning-based methods [14,15,16,17,18,19,20,21,22]. The former requires manually extracting the designed and specific RF fingerprint features from the RF signal and then using a classifier to identify the device. The latter employs deep neural networks to automatically extract features and identify different devices and achieves higher accuracy [23]. However, deep-learning-based methods are faced with two main challenges:
  • Because the differences between emitters are fairly subtle, especially for transmitters of the same model produced in the same batch by the same manufacturer, the inter-class variation in RF fingerprints is low.
  • RF fingerprints are unintentional and weak modulations that are incidental to the transmitted signal. They are susceptible to complex channel conditions, multi-path effects, and environmental noise. As a result, large intra-class variation exists in the RF fingerprints of the same device.
In this work, we attempt to deal with the above issues by combining the fine-grained recognition used in computer vision with RF-fingerprint-based recognition. Fine-grained recognition can divide a coarse-grained meta-category into more-detailed subcategories: distinguishing subcategories with low inter-class variation and high intra-class variation. Compared with general image classification, fine-grained image recognition can complete more difficult classification tasks.
Inspired by this, we propose a fine-grained RF fingerprint recognition network, which includes an adaptive spatial feature fusion module integrating both the high-level semantic and the low-level detailed information as well as an attention module to refine feature maps. As the structure of a CNN goes deeper, the neurons in high layers respond strongly to entire images and are rich in semantics, but they inevitably lose detailed information (e.g., color, edge junctions, texture patterns) from small discriminative regions [24]. To extract finer-grained features, we use ResNet [25] as the backbone and generate the pyramidal features by establishing a top-down feature pathway hierarchy on a basic CNN. In addition, we also utilize an attention mechanism to enhance the feature representation and to accurately locate discriminative regions. The contributions of this paper are as follows:
  • We propose a fine-grained RF fingerprint recognition network that generates pyramidal features by establishing a top-down feature pathway hierarchy on the basic network. Those feature maps are further refined by the attention module and are integrated to learn fine-grained features. To the best of our knowledge, we are the first to combine fine-grained recognition with RF fingerprint identification tasks.
  • We combine spatial attention and channel attention and propose a novel second-order attention module. It sequentially infers the attention map along two independent dimensions: channel and space. Subsequently, the attention map is multiplied by the input feature map to localize significant regions, enabling the learning of fine-grained and discriminative features.
  • In contrast to the conventional approach of element-wise summation or concatenation used in previous studies to fuse multi-scale features, we propose the adaptive spatial feature fusion module (ASFFM) to adaptively integrate features from different scales. This allows for the fusion of high-level semantic features and low-level detailed features: ensuring that the fine-grained details are not lost and thus facilitating more accurate fine-grained recognition.
  • We conduct extensive experiments on three challenging datasets including 100 ADS-B devices, 54 Zigbee devices, and 25 LoRa devices and achieve superior performance over the state-of-the-art approaches. A visualization and comprehensive experiments are further conducted to draw insights about our method.
The remainder of this paper is organized as follows. A literature review of RF-fingerprint-based identification and fine-grained recognition is presented in Section 2. A signal model is described in Section 3, and the proposed method is elaborated upon in Section 4. Experiments and the relevant discussion are presented in Section 5, while conclusions are drawn in Section 6.

2. Related Works

In this section, we provide a brief review of RF-fingerprint-based recognition in IoT and fine-grained recognition in deep learning.

2.1. RF Fingerprint Recognition

Prior to 2018, research on RF fingerprints focused on machine-learning-based algorithms, such as support vector machine (SVM), to identify the identity of each mobile device. After 2018, research using deep learning gradually emerged.

2.1.1. Machine Learning

The machine-learning-based approach mainly consists of two stages: extraction of manually designed RF fingerprint features and classification. Manually designed RF fingerprints can be divided into transient-state and steady-state features according to the signal state. Transient-based RF fingerprinting techniques use the transition from the turn-off to the turn-on of a transmitter before the transmission of the actual signal [5]. Hall et al. [7] used signals captured from Bluetooth transceivers and extracted instantaneous amplitude, phase, in-phase, quadrature, instantaneous power, and discrete wavelet transform (DWT) coefficients as RF fingerprints to classify 10 different Bluetooth devices. Yuan et al. [8] employed the transient signal’s time–frequency–energy distribution obtained by the Hilbert–Huang transform (HHT) to construct feature vectors and used SVM to classify the devices. These approaches require accurate transient extraction (start point and duration) before feature extraction and identification [9], which limits their real application.
Steady-state features are extracted from a signal when the transmitter is in a stable working state, making these signals more enduring and easier to obtain. The authors of [10] proposed a radiometric-signature-based device identifier named the passive radiometric device identification system (PARADIS). It takes into account the influence of various non-ideal characteristics of the device on the modulated signal, including I/Q channel offset, frequency offset, amplitude limit, etc., and successfully realized the identification of 138 wireless network cards with QPSK modulation by SVM. Although the overall recognition accuracy of PARADIS is high, it is easily affected by ambient noise. The work in [11] proposed a method for identifying different devices by extracting the RF fingerprint from the modulation shape and spectral features acquired from RFID transponders. This method is computationally efficient and can form compact fingerprints. But the fingerprints extracted from the same device vary when acquired from a long distance (e.g., more than 1 m). Peng et al. [12] employed the DCTF to capture the time-varying modulation error of Zigbee devices and developed a low-overhead classifier to identify 54 Zigbee devices. DCTF can extract features (e.g., I/Q channel deviations and non-linear variations) caused by hardware defects and eliminate the influence of the carrier frequency offset. Williams et al. [13] proposed a physical-layer authentication (PLA) scheme based on the features of the amplitude, phase, and frequency, and used a Fisher-based classifier to perform an authentication decision. The lowest recognition accuracy for three devices of the same type was only 83.25 % , which demonstrates that the differences between the extracted RF fingerprints are a little subtle.

2.1.2. Deep Learning

Sankhe et al. in [14], designed a convolutional neural network (CNN) with two convolutional layers and two fully connected layers to train on RF fingerprint data from 16 bit-similar USRP X310 SDRs. The network structure is relatively simple, which limits further improvement to its accuracy. Some works used general architectures in computer vision classification to classify RF fingerprints, such as 1D modified versions of AlexNet (AlexNet1D) and ResNet-50 (ResNet1D) [15,16], VGG-16 [17], attention mechanism [18], and so on. The researchers modified the convolution kernel to be one-dimensional to make it suitable for the 1D RF signals without considering a network architecture specifically applicable to the existing challenges of RF fingerprint recognition. Zhang et al. [19] proposed an adaptive RF fingerprint fusion network to extract and adaptively fuse multiple RF fingerprints in a data-driven manner that is robust to channel and SNR variations. Generative adversarial nets (GANs) [20] have also been applied to implement adversarial learning for identifying RF transmitters. Most of the neural-network-based approaches mentioned above primarily utilize monomodal information from the time or frequency domain. However, they overlook the potential benefits of integrating multimodal information from multiple transformation domains, which can provide complementary processing gains. Consequently, some works have adopted multimodal methods that leverage neural networks to learn features from different modalities with the aim of enhancing recognition accuracy. An et al. [21] proposed a novel approach called the series constellation multimodal feature network (SC-MFNet) for identifying the modulation types of MIMO-OFDM subcarriers. Qi et al. [22] employed the waveform–spectrum multimodal fusion (WSMF) method combined with a deep residual network (ResNet) to implement an automatic modulation classification (AMC) algorithm. By extracting features from multimodal information using ResNet and employing a feature fusion strategy, the multimodal features of the signals are merged to obtain more discriminative characteristics.

2.2. Fine-Grained Recognition

Fine-grained recognition is designed to deal with objects that belong to multiple subordinate categories of the same meta-category (such as the Alaskan and Husky subcategories in the dog category) and to discriminate objects that are highly similar in overall appearance but differ in fine-grained features. The majority of the fine-grained community focuses on two streams: (1) part localization methods and (2) feature encoding methods. Most part-based methods focus on learning explicit part localizations by detection/segmentation [24,26,27] or attention learning. The methods for detection or segmentation are typically two-stage: first, region proposals are generated, and then, the proposed regions are classified. Two-stage networks are usually time-consuming and slow.
Attention-based methods are normally one-stage and end-to-end and are widely applied in fine-grained recognition. TASN [28] is a trilinear attention sampling network that learns fine-grained details from hundreds of part proposals and efficiently distills the learned features into a single CNN. API-Net [29] learns unique features by simultaneously training the differences between a pair of similar images through the attention mechanism. ACNet [30] presents an attention-based convolutional binary neural tree to facilitate coarse-to-fine hierarchical fine-grained feature learning. The main innovation of this method is the combination of a binary tree and an attention mechanism as well as the application of an attention transformer. In this way, the network can learn the coarse-to-fine hierarchical discriminative features.
For feature encoding methods, MC-Loss [31] introduced a single loss function to learn subtle details without the requirement of an overly complicated network design. MC-Loss can easily be integrated into any backbone network with low complexity and stable performance. However, when tackling complex datasets, the improved accuracy is limited. Bilinear-CNN [32] encodes the second-order statistical information of convolutional activations as image features by multiplying the feature maps extracted by two CNNs and then performing pooling. It significantly improved fine-grained recognition, but the relevant covariance leads to a large increase in the number of associated parameters.

3. Signal Model

We take the DCTF representation in [33] as the input for FGRFNet. This first requires obtaining a differential form of the transmitted RF signals and then converting the time-domain complex signals into a 2D image.

3.1. RF Signal

For simplicity, we consider the ideal RF signal as follows:
x t = x I t + j x Q t
and the transmitted RF signal can be written as
x t = β I x I t + α I + j β Q x Q t + α Q · e j 2 π f c t t
where x I t and x Q t represent the real and imaginary parts of the RF signal x t : namely, the I-channel part and the Q-channel part, respectively; β I and β Q are the respective I/Q gain imbalances; α I and α Q are the respective DC offsets on the I/Q channel; and f c t is the carrier frequency of the transmitter.
The I/Q imbalance and DC offset are typical RF fingerprints and are difficult to extract directly in the time domain because they are mixed with modulation and encoding. Assuming that the channel and receiver are ideal, the received signal can be demodulated as:
y t = x t · e j 2 π f c r t + n t = y I t + j y Q t
where y I t and y Q t are, respectively, the real and imaginary parts of the I/Q channel of the received signal y t , f c r is the carrier frequency of the receiver, and n t is the noise.
Generally, there exists a carrier frequency offset (CFO) θ = f c r f c t , which is mainly caused by the crystal oscillator mismatch between the receiver and the transmitter or by the Doppler frequency shift.

3.2. DCTF

As pointed out in [33], DCTF is generated by converting the complex time-domain differential RF signal into a 2D constellation figure.

3.2.1. Differential Signal

The complex time-domain differential RF signal is obtained by:
d t = y I t + j y Q t + ε · y I t + λ + j y Q t + λ + ε * = β I x I t + α I + j β Q x Q t + ε + α Q · e j 2 π θ t · β I x I t + λ + α I j β Q x Q t + λ + ε + α Q · e j 2 π θ t + λ = β I x I t + α I · β I x I t + λ + α I j β I x I t + α I · β Q x Q t + λ + ε + α Q + j β Q x Q t + ε + α Q · β I x I t + λ + α I + β Q x Q t + ε + α Q · β Q x Q t + λ + ε + α Q · e j 2 π θ λ = d I t + j d Q t
where d I t and d Q t are the real and imaginary parts of the differential signal, respectively; λ is the differential time interval; ε is the introduced I/Q phase mismatch, which is used to amplify the fingerprint features; and · * represents the conjugation operation.
Note that the CFO introduces a constant phase rotation factor e j 2 π θ λ in d t and does not change with the position of the sample point.

3.2.2. Constellation Figure

In digital communication, digital signals are often represented by points on the complex plane to visually represent the relationship between different signals, which is precisely the constellation figure. In order to display the characteristics of RF fingerprints more intuitively, DCTF counts the distribution density of d t over the I/Q channels on the constellation figure.
We divide the entire constellation figure into N × N sub-regions. A density matrix Φ is built to count the number of distribution points in each sub-region. Φ i , j represents the number of points in the sub-region i , j on the constellation figure. Finally, each element of Φ is normalized in the range from 0 to 255, and the normalized Φ is just the DCTF.

3.2.3. Zigbee Examples

Some DCTF examples of Zigbee devices are shown in Figure 1, where each row represents the DCTFs from different devices, and each column represents the DCTFs of different signals emitted by the same device.
It can be seen that the differences in DCTFs between Device 1 and 2 are subtle, and this indicates the low inter-class variation. But the two DCTFs from Device 1 in the first column can be distinguished by the naked eye, and this indicates the large intra-class variation of the same emitter.
Since the Zigbee signal is modulated by offset quadrature phase-shift keying (OQPSK), x I n = ± 1 and x Q n = ± 1 . By substituting x I and x Q into (4), it can be concluded that d t has the following possibilities:
d t = β I 2 m 1 + β Q 2 m 2 + α I β I n 1 + α Q β Q n 2 + α I 2 + α Q 2 + j β I β Q n 3 + α I β Q n 4 + α Q β I n 5
where m j = ± 1 , j = 1 , 2 , n i = 0 , ± 2 , i = 1 , 2 , , 5 . As can be deduced from Equation (5), the locations of the DCTF gathering center can effectively reflect the inherent defects of the transmitter brought by the I/Q modulator: namely, the RF fingerprint. In addition, DCTF can also suggest the inherent non-linear characteristics of the power amplifier. The input signal may be distorted after the power amplifier; this is mainly manifested as constellation point deviation and dispersion.
In general, DCTF can extract I/Q channel deviations, non-linear variations, etc., and does not require any synchronization information. Its aggregation points reflect the statistical mean of d t , and the distribution reflects the variance caused by noise and random wireless channels [33].

4. The Proposed Method

Most of the previous works simply migrated networks commonly used in deep learning to the task of RF fingerprint recognition, and their performance may not be that outstanding. In our experiments, we observed that devices exhibit small inter-class differences and large intra-class differences in RF fingerprints. This is a challenge that was not pointed out and addressed by previous methods. Therefore, targeting this challenge in RF fingerprint recognition, we specifically propose FGRFNet. Unlike previous research, we propose a second-order attention module. In contrast to earlier attention-based approaches, our method captures long-range dependencies and finely models complex relationships between elements, enhancing the model’s expressive power. Furthermore, we propose an adaptive spatial feature fusion module. Previous works typically used element-wise summation or concatenation to fuse multi-scale features, implying equal weighting for different features, which is evidently unreasonable. The adaptive fusion module we propose autonomously learns fusion weights for each pixel in feature maps of different scales, fusing high-level semantic features with low-level details. This enables the network to extract discriminative fine-grained features for distinguishing subtly different devices.

4.1. Data Sample

We found that if the original complex I/Q signal is directly fed into the neural network, the neural network cannot learn the true RF fingerprint. For example, in ADS-B data, it will learn the international civil aviation organization (ICAO) address code that represents the aircraft’s identity.
Therefore, we directly generate the DCTF representations of the signals and feed the DCTFs into the fine-grained network. Notably, each transmitter has a unique label. Different devices of the same batch and type from the same manufacturer have different labels, and this conforms to the scope of the fine-grained classification. In the subsequent experiments, we visually present the learned RF fingerprints in a heat map.

4.2. FGRFNet

An overview of the proposed method for fine-grained RF fingerprint recognition is shown in Figure 2. It consists of four modules: namely, the backbone network, the channel and positional attention module (CPAM), the adaptive spatial feature fusion module (ASFFM), and the classifier.
Firstly, the network takes as input a DCTF, which is fed into convolutional layers to extract pyramidal feature maps. Then, these feature maps are further transformed into pyramidal enhanced discriminative feature representations by the CPAM, following a bottom-up pathway. Once the attention feature pyramid has been obtained from the raw input, an ASFFM integrates the low-level detailed information with high-level semantic feature representations. Finally, a group of probability scores over the fused features to fine-grained categories are predicted by fully connected and softmax layers.
The proposed FGRFNet is optimized for convergence by learning a cross-entropy loss. Note that it can be trained end-to-end, and the framework is flexible on CNN backbone structures.

4.3. CPAM

Inspired by [34], we combine spatial attention and channel attention and propose a novel second-order attention module. As illustrated in Figure 3, given a feature map F as input, CPAM sequentially infers a channel attention matrix M c R C × C and a spatial attention matrix M p R H × W × H × W , where M c and M p respectively model the channel and spatial relationships between any two pixels of the features.
Channel Attention Module: As pointed out in [28,35], a convolutional feature channel often corresponds to a certain type of visual pattern. By clustering spatially correlated channels for which the peak responses appear in neighboring locations, the fine-grained part-feature representation can be improved. Therefore, we build a channel attention module (CAM) to explicitly model the channel dependency.
Considering an input image, we first extract the feature maps by a base network that includes a series of convolution and pooling operations. Assume that the dimension of the feature map F is C × H × W , where C, H, and W denote the channel number, height, and width, respectively. We first reshape F into a matrix with a shape of C × N , N = H × W , and denote it as F r = f 1 , , f c T R C × N . Then, we correlate F r and its transpose as an attention map: namely,
M c = F r F r T
We further normalize M c using a softmax layer, i.e.,
M c i , j = e M c i , j j = 1 C e M c i , j = e f i · f j j = 1 C e f i · f j
Each element of M c i , j measures the similarity between the ith and jth channels and indicates the impact of channel j on channel i.
In order to guide the network to focus on discriminative regions, we further multiply M c and F r and reshape the result. The final output F R C × H × W is acquired by multiplying the refined output by a learnable parameter γ and performing an element-wise summation operation with F :
F = γ · reshape M c · F r + F
where γ is initialized to be 0 and gradually learns to assign more weight [36]. Equation (8) indicates that the refined feature map F of each channel is a weighted summation of all channel features with the original features. Through the CAM, semantically significant parts of the feature map can be highlighted by integrating spatially correlated channels.
Position Attention Module: Discriminant feature representations are essential for fine-grained recognition; they can be obtained by capturing long-range contextual information. However, convolution mainly deals with local neighborhoods and must be stacked with many layers to (capture long-range dependencies to) associate different parts of the whole image. This introduces computational inefficiency and high complexity [37]. We use a position attention module (PAM) to mine locally discriminative regions. It can capture long-range dependencies directly by computing interactions between any two positions regardless of their positional distances.
The PAM takes the refined output F of the CAM as input. Different from the CAM, the PAM first feeds F into a convolutional layer with a kernel size of one to reduce parameter overhead. This generates two new feature maps: K and Q . The subsequent operations are similar to those for the CAM. The two feature maps are further reshaped to K , Q R C × N , where N = H × W is the number of pixels.
Then, we multiply the transpose of K = k 1 , k 2 , , k N R C × N and Q = q 1 , q 2 , , q N R C × N to obtain the attention map M p R N × N :
M p = K T Q
We further normalize M p by using a softmax layer:
M p i , j = e M p i , j i = 1 C e M p i , j = e k i · q j i = 1 C e k i · q j
where M p i , j measures the jth position’s impact on the ith position. Equation (10) indicates that any two pixels in the image can interact with each other, and space dependencies can directly be captured in a feedforward fashion by this method.
Meanwhile, in the same way that K , Q is obtained, we feed the feature map F into a convolution layer with a kernel size of one to generate a new feature map V R C × H × W and reshape it to V R C × N . Then, we multiply V and M p and reshape the result to R C × H × W . The final output F R C × H × W is achieved by multiplying the refined result by a learnable parameter δ and performing an element-wise summation with the original input feature map F :
F = δ · reshape V · M p + F
where δ also starts from 0 and gradually learns a weight. An initialization of 0 means that the PAM is not used when the training starts; this is because the network is expected to learn the local information first. As training proceeds, the network will slowly try to use the PAM to learn the long-range dependencies between distant pixels.
Equation (11) indicates that each position of F is a weighted summation of features across all positions and the original features. PAM can selectively aggregate semantically similar parts in space, and discriminative regions achieve higher gains: thus facilitating fine-grained identification.

4.4. ASFFM

Previous works [38,39,40] have typically exploited element-wise summation or concatenation to fuse multi-scale features, which means that semantically strong parts are weighted the same as semantically weak parts. Inspired by [41], an ASFFM is proposed to adaptively learn the spatial weight of fusion for feature maps at each scale.
As illustrated in Figure 4, let X 1 , X 2 , and X 3 denote the three input feature maps, respectively, among which, the resolution of the former is twice that of the latter, while the channel number of the former is 1 / 2 that of the latter. Thus, we first resize the feature maps to the same shape of X 3 . Concretely, X 2 is fed into a 3 × 3 convolution layer with a stride of two to simultaneously modify the channel number and downsample the resolution to one-half of the original. For X 1 , to achieve downsampling with a 1 / 4 ratio, we add a max-pooling layer with a stride of two before the two-stride convolution.
Let F 1 , F 2 , and F 3 denote the resized feature maps with the same size. Then, the three feature maps at different scales are integrated as follows:
F i j f = ξ i j · F i j 1 + μ i j · F i j 2 + ν i j · F i j 3
where F i j f and F i j n , n = 1 , 2 , 3 , denote the descriptor at position i , j on the feature maps; and ξ i j , μ i j , and ν i j are weights that indicate the importance of different positions on the feature map. These weights are adaptively learned by the network and are broadcast along the channel dimension during multiplication. Specifically, ξ i j , μ i j , ν i j 0 , 1 are formulated by a softmax layer:
ξ i j = e η ξ i j e η ξ i j + e η μ i j + e η ν i j μ i j = e η μ i j e η ξ i j + e η μ i j + e η ν i j ν i j = e η ν i j e η ξ i j + e η μ i j + e η ν i j
where η ξ , η μ , and η ν are generated by applying a 1 × 1 convolution layer on F 1 , F 2 , and F 3 , respectively, and thus, they can be learned through standard back-propagation.

5. Experiments

5.1. Dataset

To assess the performance of our proposed FGRFNet for learning discriminative parts and fine-grained image recognition, we conduct experiments on ADS-B, Zigbee, and LoRa datasets. Instead of the original time-domain RF signals, we utilized their DCTF representations with the parameters λ , ε , and N, which are described in Section 3.2 to be 10, 2, and 65, respectively.
ADS-B: An ADS-B signal contains geo-coordinates, velocities, altitudes, headings, as well as the unique identifier information of the commercial aircraft. They are easy to receive and decode but are subject to identity spoofing attacks. We used the public dataset found in [42], for which the raw I/Q data of the signals were obtained at a sampling rate of 8 MHz and with a 1090 MHz center frequency; we generated 40,000 DCTFs in 100 subordinate categories and included 32,000 images for training and 8000 images for testing.
Zigbee: A Zigbee signal is modulated with O-QPSK modulation according to the IEEE 802.15.4 standard [43]. In the public dataset [33], the receiver works at 10 MSPS (million samples per second), and for each Zigbee device, 5 measurements with 9 segments of 40 K samples are recorded. To fully evaluate the proposed method, we also added additive Gaussian white noise (AWGN) with different signal-to-noise ratios (SNRs) ranging from 0 to 25 dB to the received signals. We converted the received signals to DCTFs and obtained 24,300 DCTFs of 54 classes in total; 60% (14,580) of the images are used for training, and the remaining 40% (9720) are used for testing.
LoRa: LoRa technology is a wireless communication technology widely utilized by IoT applications. The publicly available LoRa dataset [44] was collected from 25 distinct IoT devices with LoRa functionality using a USRP B210 receiver. The dataset offers a comprehensive collection of fundamental experimental scenarios, including indoor and outdoor environments as well as various network deployments and configurations. For each class, we partitioned 320 samples for training and 80 samples for testing.

5.2. Implementation

In the training stage, we first normalize the input DCFTs. Then, we use either ResNet-50 or ResNet-101 as the backbone and choose the last output features of the residual blocks from Stage 3, Stage 4, and Stage 5 in ResNet to establish a pyramidal hierarchy. We take a tensor of size [3, 64, 64] as an example, where the first number in the tensor represents the channels, and the subsequent two numbers respectively represent the height and width of the image. As the features pass through each stage, the number of channels doubles, and the size reduces by half. Consequently, the feature sizes of the outputs for Stages 3, 4, and 5 are [512, 8, 8], [1024, 4, 4], and [2048, 2, 2], respectively. We take the output of Stage 3 as an example to analyze how the CPAM and ASFFM operate.
Suppose the tensor size of F is [512, 8, 8]. First, channel attention is performed by reshaping F into [512, 16], denoted as F r . After the covariance operation, the channel attention matrix M c has a size of [512, 512]. Normalizing it with a softmax layer yields attention scores M c of size [512, 512]. Multiplying the attention scores by F r and reshaping the result into [512, 8, 8] followed by multiplication with a learnable parameter and addition to the original input feature map gives the refined feature map after channel attention mechanism correction, denoted as F and with a size of [512, 8, 8]. Spatial attention calculation follows a similar process. Using the output F of channel attention as input, three 1 × 1 convolutions are applied to obtain K , Q , and V with a size of [512, 8, 8]. After reshaping Q and K into [512, 16], multiplication is performed, and normalization yields spatial attention scores M p with a size of [16, 16]. Multiplying the attention scores by V , reshaping the result into [512, 8, 8], and calculating based on Equation (11) gives the feature map F after the attention mechanism.
From the above analysis, it is evident that the size of the feature map remains unchanged after the CPAM. Therefore, the feature-enhanced outputs X 1 , X 2 , and X 3 of Stages 3, 4, and 5 remain [512, 8, 8], [1024, 4, 4], and [2048, 2, 2], respectively. Subsequently, the ASFFM integrates these features. After applying max-pooling layers or two-stride convolution layers to resize X 1 and X 2 to [512, 8, 8], denoted as F 1 , F 2 , respectively, 1 × 1 convolutions are applied to F 1 , F 2 , and F 3 to obtain fusion weights. Multiplying the feature maps by their respective fusion weights results in the size of the adaptively spatially fused feature being [512, 8, 8]. This is then fed into the classifier for experimental radio frequency identification.
During training, the learning rate is initially set to 0.005 with a series of decays to change the value to one-tenth of the original value per 10 epochs. The weight decay is set to 0.0005, the stochastic gradient descent (SGD) algorithm is used as the optimizer, and the batch size is set to 32. On the ADS-B dataset, 100 epochs are trained, and 80 epochs are trained on the Zigbee and LoRa datasets.
During testing, we evaluate the performance of RF fingerprint recognition using metrics such as accuracy, F1-score, and AUC (Area Under the ROC Curve). F1-score combines precision and recall into a single value, providing a balanced measure of a classifier’s accuracy. The higher the three metrics mentioned above, the better the classification performance.
All experiments are carried out on a single Nvidia GeForce RTX 3090 GPU card, and the PyTorch toolbox is used as the main implementation substrate.
Figure 5 shows the training and test loss curves of the FGRFNet on the three datasets. As can be seen, the training loss reaches a significantly low value, and this indicates that the model fits the data well.

5.3. Comparison and Discussion

To further validate the proposed method, we compare it with other state-of-the-art methods for both fine-grained recognition and specific emitter identification fields in Table 1 and Table 2. In Table 1, our FGRFNet achieves a significant improvement over the corresponding backbones on the ADS-B dataset, with clear margins of 4.21% and 3.66% in accuracy on ResNet-50 and ResNet-101, respectively. In terms of F1-score and AUC, our network achieved the highest values of 0.8982 and 0.9989, respectively, surpassing the other methods. This suggests strong discriminatory ability and reliability of the network. Compared with ICAM [45], which integrates a plug-and-play attention block into the backbone to enhance the exploitation of spatial and channel information by the convolutional layers, we get a better result, with a relative accuracy improvement of 1.83%. In comparison to methods that solely utilize a few simple convolutional layers, such as AlexNet [46] and CNN [33], our approach yields superior results, with a notable relative accuracy improvement of over 6.5%. This shows the superior attention learning ability of the proposed approach. In addition, compared with the leading result achieved by GoogLenet, our method has an improved accuracy of 1.04% with ResNet-50.
Table 2 lists several representative approaches with high accuracy on the Zigbee and LoRa datasets. On both datasets, our network achieves AUC values of 0.9999 and 0.9915, respectively, indicating that our network is very close to being a “perfect classifier”. Using ResNet-50 as the baseline, our method still achieves a competitive accuracy of 83.00%, which is 2.55% and 4.2% better than MobileNetV3 [49] and GoogLenet [46], respectively, on the LoRa dataset. ICAM [45] and ARFNet [19] also utilize spatial and channel attention mechanisms. However, our approach outperforms them on the Lora dataset with improvements of over 2% in accuracy and F1-score. In contrast to the attention mechanisms employed in the aforementioned two methods, our CPAM facilitates interactions between every pixel, rather than confining interactions solely to local regions as in ICAM and ARFNet, providing a more robust modeling capability. This empowers the model to better grasp long-range dependencies between elements. Furthermore, our method ensures precise attention scores at each pixel for positional attention, facilitating the extraction of finer features and capturing subtle differences between different devices, thereby enabling effective discrimination between them. Additionally, FGRFNet incorporates ASFFM to fuse features from different scales, merging high-level semantic features with low-level details. This allows the model to extract more discriminative fine-grained features. Consequently, our approach outperforms other state-of-the-art methods and achieves the highest scores across all three metrics.

5.4. Ablation Study

To analyze the contributions of different components in the proposed framework, we conduct various ablation experiments on the three datasets.
Impact of CPAM: In Table 3, compared with the backbone-only model, the CPAM-based model enhances the accuracy by 2.81% for ADS-B, 0.77% for Zigbee, and 3.7% for LoRa. This shows that CPAM can localize the discriminative parts to learn the fine-grained part-feature representation, which is beneficial for fine-grained RF fingerprint recognition.
Table 3 also presents the contributions of CAM and PAM to the entire attention module. CAM calculates the importance of each channel of the input image and thus makes the input image more meaningful. That is, a channel with key information will be given a higher weight, and one without important information will be given a lower weight. In this way, more informative discriminative regions can be located, and background noise can be excluded. Meanwhile, PAM can capture long-distance spatially related information and make up for the limited receptive field of CNN (owing to the CNN’s limited convolution kernel size). The combination of CAM and PAM enables CPAM to locate key areas more accurately, which is helpful for learning region-based fine-grained features.
Impact of ASFFM: In Table 4, with the ASFFM, the backbone model and the backbone–CPAM model achieve accuracy improvements of 2.43% and 1.78%, respectively, on the ADS-B dataset. This indicates that the ASFFM can integrate low-level detailed information with enhanced high-level semantic information features and thus learn part-feature representations for fine-grained classification by fusing pyramidal features.

5.5. Impact of DCTF Generation Parameters

DCTF image generation is influenced by various factors, including differential interval λ , I/Q phase distortion ε , DCTF sizes, and sample lengths. DCTF exhibits significant variations under different differential intervals and I/Q phase mismatches. Therefore, it is essential to search for optimal parameters to maximize the distinction between different devices. At the same time, image sizes and the number of samples also affect the quality of DCTF. Larger image sizes decrease the likelihood of different samples coinciding within the same pixel; this necessitates a higher number of samples to preserve graphic details. Conversely, smaller image sizes result in blurred features between devices. Therefore, in this section, we investigate the effect of the DCTF generation parameters on the performance of the model using the Zigbee dataset.
Figure 6 illustrates the accuracy of FGRFNet under various DCTF generation parameters when the SNR is 25 dB and the DCTF size is 64 × 64 . As shown in Figure 6, the left plot demonstrates the impact of varying ε on the accuracy of the model when λ is fixed at 10. It can be observed that setting ε to 0 leads to a significant decrease in accuracy, while for ε greater than 1, the overall accuracy consistently remains above 98%. Particularly, the best performance is achieved when ε is set to 2. The right plot depicts the influence of different λ values on the model’s accuracy when ε is held constant at 2. It is evident that setting extremely short differential intervals leads to a significant drop in accuracy. However, when λ exceeds 3, the overall accuracy consistently surpasses 98.5%. Notably, the optimal performance is observed when λ is set to 10.
The accuracy when using different DCTF image qualities is illustrated in Figure 7 for an SNR of 25 dB, λ = 10 , and ε = 2 . The left graph presents the relationship between model accuracy and DCTF sizes under a fixed sample length. It is evident that when the image size is too small, such as below 16 × 16 , the RF fingerprint features become indistinct, resulting in a decline in model accuracy. However, as the image size decreases, the computational complexity of the network also decreases. Experimental results indicate that an image size of 64 × 64 provides the best performance of FGRFNet with relatively low complexity. With a fixed DCTF size of 64 × 64 , the right graph clearly illustrates that as the number of samples used for generating DCTF increases, the quality of the DCTF images improves, resulting in enhanced recognition accuracy.

5.6. Impact of Distance

Table 5 describes the impact of the distance between the transmitter and receiver on recognition accuracy. The study examines two specific distances: 5 m and 15 m. For each distance, data from 25 different LoRa devices were collected and analyzed. It is evident that the recognition accuracy of the model decreases as the distance between the transmitter and receiver increases, owing to the impact of varying channel conditions. Notably, methods like CNN [33] and AlexNet [46] exhibit relatively lower precision. Thus, we can infer that varying channel conditions have a significant impact on the accuracy of CNN and AlexNet methods. The experimental results indicate that our proposed FGRFNet exhibits a certain level of resilience against channel effects, showcasing more stable feature extraction compared to other methods. As the distance increases and the SNR decreases, the accuracy of this approach gradually declines.

5.7. Impact of SNR

In order to evaluate the influence of SNR on RF fingerprinting recognition, we also carried out experiments under different SNR levels in the range of [0, 25] dB at a step of 5 dB using the Zigbee dataset. Figure 8 draws the accuracy curves using CNN [33], SCNN [51], AlexNet [46], GoogLenet [46], ResNet-50 [25], ICAM [45], MobileNetV3 [49], and our method for different SNR levels. It can be seen that as the SNR decreases, all the accuracies degrade. The performance deteriorates substantially, especially when the SNR is lower than 15 dB. But FGRFNet achieves an overall higher accuracy than other methods. In particular, when the SNR is lower than 15 dB, FGRFNet exhibits a performance advantage of approximately 10% over CNN. This indicates that FGRFNet is capable of effectively learning discriminative features and exhibiting greater robustness to noise. Figure 9 further displays the recognition confusion matrix of FGRFNet for different SNRs. With increasing SNR, the accuracy of RF fingerprint recognition is improved as well.

5.8. Visualization

We adopt gradient-weighted class activation mapping (Grad-CAM) [52] to visualize and analyze the learned RF fingerprints to make our FGRFNet more interpretable. Grad-CAM generates heat maps that show the contribution distribution of the predicted output for input images. A higher score indicates a stronger response and greater contribution from the corresponding areas of the original image. This highlights the importance of each location to the class.
We selected five DCTFs from five different Zigbee devices at 25 dB for visualization. Figure 10 displays the original DCTFs along with the visualization results of the backbone and FGRFNet. Device 4 and Device 5 are misclassified by the backbone ResNet but are correctly classified by FGRFNet. Through careful observation, we can see that FGRFNet pays more attention to the global structure, while the backbone ResNet just focuses on the gathering centers. As pointed out in Section 3.2, the RF fingerprint is related to the locations of gathering centers as well as the deviation and dispersion of constellation points. Locations of gathering centers effectively reflect the inherent defects in the transmitter brought about by the I/Q modulator, such as I/Q gain imbalances, DC offsets, and CFO. In addition, constellation point deviation and dispersion suggest the inherent non-linear characteristics of the power amplifier.
Figure 10 demonstrates that the backbone network selectively focuses on prominent gathering centers, indicating that it solely learns the inherent defects of the transmitter caused by the I/Q modulator, such as I/Q gain imbalances, DC offsets, and carrier frequency offsets. However, FGRFNet captures additional subtle and fine-grained features, including constellation point deviation and dispersion, indicating that FGRFNet simultaneously learns the non-linearities of the amplifier and the inherent defects of the transmitter caused by the I/Q modulator: namely, the RF fingerprints. The proposed FGRFNet successfully learns more implicit and fine-grained features, such as the non-linearities of the amplifier, thereby achieving higher performance in Figure 10.

6. Conclusions

In this article, we developed a fine-grained RF-fingerprint-based recognition network to identify different wireless devices by fusing multi-level features that are enhanced by CPAMs to deliver both the high-level semantic and the low-level detailed information. The attention module can effectively locate discriminative regions and reduce the effect of background noise, and the fusion module can enhance the recognition robustness. Experimental results on the ADS-B, Zigbee, and LoRa datasets verified the effectiveness of the proposed FGRFNet.
While our method achieved excellent performance, it comes with some drawbacks. On the one hand, there is relatively high computational complexity, primarily due to covariance calculations in the attention module, leading to a substantial increase in the number of parameters. On the other hand, this method is only suitable for closed-set classification: meaning that the device categories in the training set and the test set must be the same.
However, in real-world scenarios, the electromagnetic environment is open and dynamic, and the types and quantities of wireless devices can vary and remain uncertain. The closed-set authentication approach may not be suitable for handling such situations. Therefore, in addition to advancing fine-grained recognition techniques, we also aim to explore open-set recognition methods in the future. By dedicating our efforts to open-set recognition, we aim to address the challenges posed by the dynamic and diverse nature of the IoT environment. Our ultimate goal is to facilitate the widespread and practical application of RF fingerprint recognition in IoT settings.

Author Contributions

Conceptualization, Methodology, and Software, Y.Z.; Formal Analysis and Investigation, Y.Z., R.J., and Z.L.; Writing—Original Draft Preparation, Y.Z.; Writing—Review and Editing, J.H. and Y.Z.; Supervision, Project Administration, and Funding Acquisition, J.H. and Z.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work waspartly supported by the Shenzhen Fundamental Research Program under grant No. JCYJ20180307151430655 and the Shenzhen Science and Technology Program under grant No. KQTD20190929172704911.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Our source codes have been released at https://github.com/zhangyu1an/FGRFNet (accessed on 22 December 2023).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
FGRFNetFine-grained RF Fingerprint Recognition Network
DCTFDifferential Constellation Trace Figure
CFOCarrier Frequency Offset
ICAOInternational Civil Aviation Organization
CPAMChannel and Positional Attention Module
ASFFMAdaptive Spatial Feature Fusion Module

References

  1. Al-Garadi, M.A.; Mohamed, A.; Al-Ali, A.K.; Du, X.; Ali, I.; Guizani, M. A survey of machine and deep learning methods for internet of things (IoT) security. IEEE Commun. Surv. Tutor. 2020, 22, 1646–1685. [Google Scholar] [CrossRef]
  2. Xie, N.; Li, Z.; Tan, H. A Survey of Physical-Layer Authentication in Wireless Communications. IEEE Commun. Surv. Tutor. 2021, 23, 282–310. [Google Scholar] [CrossRef]
  3. Liu, Y.; Wang, J.; Li, J.; Song, H.; Yang, T.; Niu, S.; Ming, Z. Zero-bias deep learning for accurate identification of Internet-of-Things (IoT) devices. IEEE Internet Things J. 2020, 8, 2627–2634. [Google Scholar] [CrossRef]
  4. Zhang, J.; Woods, R.; Sandell, M.; Valkama, M.; Marshall, A.; Cavallaro, J. Radio Frequency Fingerprint Identification for Narrowband Systems, Modelling and Classification. IEEE Trans. Inf. Forensics Secur. 2021, 16, 3974–3987. [Google Scholar] [CrossRef]
  5. Soltanieh, N.; Norouzi, Y.; Yang, Y.; Karmakar, N.C. A review of radio frequency fingerprinting techniques. IEEE J. Radio Freq. Identif. 2020, 4, 222–233. [Google Scholar] [CrossRef]
  6. Rajendran, S.; Sun, Z. RF Impairment Model-Based IoT Physical-Layer Identification for Enhanced Domain Generalization. IEEE Trans. Inf. Forensics Secur. 2022, 17, 1285–1299. [Google Scholar] [CrossRef]
  7. Barbeau, M.; Hall, J.; Kranakis, E. Detection of rogue devices in bluetooth networks using radio frequency fingerprinting. In Proceedings of the 3rd IASTED International Conference on Communications and Computer Networks, CCN, Lima, Peru, 4–6 October 2006; pp. 4–6. [Google Scholar]
  8. Yuan, Y.; Huang, Z.; Wu, H.; Wang, X. Specific emitter identification based on Hilbert–Huang transform-based time–frequency–energy distribution features. IET Commun. 2014, 8, 2404–2412. [Google Scholar] [CrossRef]
  9. Satija, U.; Trivedi, N.; Biswal, G.; Ramkumar, B. Specific emitter identification based on variational mode decomposition and spectral features in single hop and relaying scenarios. IEEE Trans. Inf. Forensics Secur. 2018, 14, 581–591. [Google Scholar] [CrossRef]
  10. Brik, V.; Banerjee, S.; Gruteser, M.; Oh, S. Wireless device identification with radiometric signatures. In Proceedings of the 14th ACM International Conference on Mobile Computing and Networking, San Francisco, CA, USA, 8–12 September 2008; pp. 116–127. [Google Scholar]
  11. Danev, B.; Heydt-Benjamin, T.S.; Capkun, S. Physical-layer Identification of RFID Devices. In Proceedings of the USENIX Security Symposium, Montreal, QC, Canada, 10–14 August 2009; pp. 199–214. [Google Scholar]
  12. Peng, L.; Hu, A.; Zhang, J.; Jiang, Y.; Yu, J.; Yan, Y. Design of a hybrid RF fingerprint extraction and device classification scheme. IEEE Internet Things J. 2018, 6, 349–360. [Google Scholar] [CrossRef]
  13. Williams, M.D.; Temple, M.A.; Reising, D.R. Augmenting bit-level network security using physical layer RF-DNA fingerprinting. In Proceedings of the 2010 IEEE Global Telecommunications Conference GLOBECOM 2010, Miami, FL, USA, 6–10 December 2010; pp. 1–6. [Google Scholar]
  14. Sankhe, K.; Belgiovine, M.; Zhou, F.; Riyaz, S.; Ioannidis, S.; Chowdhury, K. ORACLE: Optimized radio classification through convolutional neural networks. In Proceedings of the IEEE INFOCOM 2019-IEEE Conference on Computer Communications, Paris, France, 29 April–2 May 2019; pp. 370–378. [Google Scholar]
  15. Soltani, N.; Reus-Muns, G.; Salehi, B.; Dy, J.; Ioannidis, S.; Chowdhury, K. RF fingerprinting unmanned aerial vehicles with non-standard transmitter waveforms. IEEE Trans. Veh. Technol. 2020, 69, 15518–15531. [Google Scholar] [CrossRef]
  16. Jian, T.; Rendon, B.C.; Ojuba, E.; Soltani, N.; Wang, Z.; Sankhe, K.; Gritsenko, A.; Dy, J.; Chowdhury, K.; Ioannidis, S. Deep learning for RF fingerprinting: A massive experimental study. IEEE Internet Things Mag. 2020, 3, 50–57. [Google Scholar] [CrossRef]
  17. Zong, L.; Xu, C.; Yuan, H. A RF fingerprint recognition method based on deeply convolutional neural network. In Proceedings of the 2020 IEEE 5th Information Technology and Mechatronics Engineering Conference (ITOEC), Chongqing, China, 12–14 June 2020; pp. 1778–1781. [Google Scholar]
  18. Jagannath, A.; Kane, Z.; Jagannath, J. RF Fingerprinting Needs Attention: Multi-task Approach for Real-World WiFi and Bluetooth. arXiv 2022, arXiv:2209.03142. [Google Scholar]
  19. Zhang, W.; Zhao, W.; Tan, X.; Shao, L.; Ran, C. Adaptive RF Fingerprints Fusion via Dual Attention Convolutions. IEEE Internet Things J. 2022, 9, 25181–25195. [Google Scholar] [CrossRef]
  20. Roy, D.; Mukherjee, T.; Chatterjee, M.; Pasiliao, E. Detection of rogue RF transmitters using generative adversarial nets. In Proceedings of the 2019 IEEE Wireless Communications and Networking Conference (WCNC), Marrakesh, Morocco, 15–18 April 2019; pp. 1–7. [Google Scholar]
  21. An, Z.; Zhang, T.; Shen, M.; De Carvalho, E.; Ma, B.; Yi, C.; Song, T. Series-constellation feature based blind modulation recognition for beyond 5G MIMO-OFDM systems with channel fading. IEEE Trans. Cogn. Commun. Netw. 2022, 8, 793–811. [Google Scholar] [CrossRef]
  22. Qi, P.; Zhou, X.; Zheng, S.; Li, Z. Automatic modulation classification based on deep residual networks with multimodal information. IEEE Trans. Cogn. Commun. Netw. 2021, 7, 21–33. [Google Scholar] [CrossRef]
  23. Jagannath, A.; Jagannath, J.; Kumar, P.S.P.V. A comprehensive survey on radio frequency (RF) fingerprinting: Traditional approaches, deep learning, and open challenges. Comput. Netw. 2022, 219, 109455. [Google Scholar] [CrossRef]
  24. Ding, Y.; Ma, Z.; Wen, S.; Xie, J.; Chang, D.; Si, Z.; Wu, M.; Ling, H. AP-CNN: Weakly supervised attention pyramid convolutional neural network for fine-grained visual classification. IEEE Trans. Image Process. 2021, 30, 2826–2836. [Google Scholar] [CrossRef]
  25. Kumar, Y.; Sheoran, M.; Jajoo, G.; Yadav, S.K. Automatic Modulation Classification Based on Constellation Density Using Deep Learning. IEEE Commun. Lett. 2020, 24, 1275–1278. [Google Scholar] [CrossRef]
  26. Yang, S.; Liu, S.; Yang, C.; Wang, C. Re-rank coarse classification with local region enhanced features for fine-grained image recognition. arXiv 2021, arXiv:2102.09875. [Google Scholar]
  27. Ge, W.; Lin, X.; Yu, Y. Weakly supervised complementary parts models for fine-grained image classification from the bottom up. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3034–3043. [Google Scholar]
  28. Zheng, H.; Fu, J.; Zha, Z.J.; Luo, J. Looking for the devil in the details: Learning trilinear attention sampling network for fine-grained image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5012–5021. [Google Scholar]
  29. Zhuang, P.; Wang, Y.; Qiao, Y. Learning attentive pairwise interaction for fine-grained classification. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 13130–13137. [Google Scholar]
  30. Ji, R.; Wen, L.; Zhang, L.; Du, D.; Wu, Y.; Zhao, C.; Liu, X.; Huang, F. Attention convolutional binary neural tree for fine-grained visual categorization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10468–10477. [Google Scholar]
  31. Chang, D.; Ding, Y.; Xie, J.; Bhunia, A.K.; Li, X.; Ma, Z.; Wu, M.; Guo, J.; Song, Y.Z. The devil is in the channels: Mutual-channel loss for fine-grained image classification. IEEE Trans. Image Process. 2020, 29, 4683–4695. [Google Scholar] [CrossRef]
  32. Lin, T.Y.; RoyChowdhury, A.; Maji, S. Bilinear convolutional neural networks for fine-grained visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 1309–1322. [Google Scholar] [CrossRef] [PubMed]
  33. Peng, L.; Zhang, J.; Liu, M.; Hu, A. Deep Learning Based RF Fingerprint Identification Using Differential Constellation Trace Figure. IEEE Trans. Veh. Technol. 2020, 69, 1091–1095. [Google Scholar] [CrossRef]
  34. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual Attention Network for Scene Segmentation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 3141–3149. [Google Scholar] [CrossRef]
  35. Zheng, H.; Fu, J.; Mei, T.; Luo, J. Learning multi-attention convolutional neural network for fine-grained image recognition. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5209–5217. [Google Scholar]
  36. Zhang, H.; Goodfellow, I.; Metaxas, D.; Odena, A. Self-attention generative adversarial networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 7354–7363. [Google Scholar]
  37. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  38. Si, W.; Luo, J.; Deng, Z. Multi-label hybrid radar signal recognition based on a feature pyramid network and class activation mapping. IET Radar Sonar Navig. 2022, 16, 786–798. [Google Scholar] [CrossRef]
  39. Chen, T.; Gao, S.; Zheng, S.; Yu, S.; Xuan, Q.; Lou, C.; Yang, X. EMD and VMD Empowered Deep Learning for Radio Modulation Recognition. IEEE Trans. Cogn. Commun. Netw. 2023, 9, 43–57. [Google Scholar] [CrossRef]
  40. Wu, W.; Hu, S.; Lin, D.; Liu, Z. DSLN: Securing Internet of Things Through RF Fingerprint Recognition in Low-SNR Settings. IEEE Internet Things J. 2022, 9, 3838–3849. [Google Scholar] [CrossRef]
  41. Liu, S.; Huang, D.; Wang, Y. Learning spatial fusion for single-shot object detection. arXiv 2019, arXiv:1911.09516. [Google Scholar]
  42. Liu, Y.; Wang, J.; Niu, S.; Song, H. ADS-B signals records for non-cryptographic identification and incremental learning. IEEE Dataport 2021. [Google Scholar] [CrossRef]
  43. IEEE Std 802.15.4-2020 (Revision of IEEE Std 802.15.4-2015); IEEE Standard for Low-Rate Wireless Networks. IEEE: Piscataway, NJ, USA, 2020; pp. 1–800. [CrossRef]
  44. Elmaghbub, A.; Hamdaoui, B. Comprehensive RF dataset collection and release: A deep learning-based device fingerprinting use case. In Proceedings of the 2021 IEEE Globecom Workshops (GC Wkshps), Madrid, Spain, 7–11 December 2021; pp. 1–7. [Google Scholar]
  45. Yin, Z.; Chen, B.; Zhen, W.; Wang, C.; Zhang, T. The Performance Analysis of Signal Recognition Using Attention Based CNN Method. IEEE Access 2020, 8, 214915–214922. [Google Scholar] [CrossRef]
  46. Peng, S.; Jiang, H.; Wang, H.; Alwageed, H.; Zhou, Y.; Sebdani, M.M.; Yao, Y.D. Modulation Classification Based on Signal Constellation Diagrams and Deep Learning. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 718–727. [Google Scholar] [CrossRef]
  47. Du, R.; Chang, D.; Bhunia, A.K.; Xie, J.; Ma, Z.; Song, Y.Z.; Guo, J. Fine-grained visual classification via progressive multi-granularity training of jigsaw patches. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 153–168. [Google Scholar]
  48. Liu, Y.; Wang, J.; Li, J.; Niu, S.; Song, H. Class-incremental learning for wireless device identification in IoT. IEEE Internet Things J. 2021, 8, 17227–17235. [Google Scholar] [CrossRef]
  49. Ran, X.; Yan, T.; Cai, T. MobileNet for Differential Constellation Trace Figure. In Proceedings of the 2021 13th International Conference on Communication Software and Networks (ICCSN), Chongqing, China, 4–7 June 2021; pp. 168–172. [Google Scholar] [CrossRef]
  50. Luo, W.; Zhang, H.; Li, J.; Wei, X.S. Learning semantically enhanced feature for fine-grained image classification. IEEE Signal Process. Lett. 2020, 27, 1545–1549. [Google Scholar] [CrossRef]
  51. Zeng, Y.; Zhang, M.; Han, F.; Gong, Y.; Zhang, J. Spectrum Analysis and Convolutional Neural Network for Automatic Modulation Recognition. IEEE Wirel. Commun. Lett. 2019, 8, 929–932. [Google Scholar] [CrossRef]
  52. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
Figure 1. Examples of the DCTFs generated by three real Zigbee devices.
Figure 1. Examples of the DCTFs generated by three real Zigbee devices.
Entropy 26 00029 g001
Figure 2. The structure of FGRFNet.
Figure 2. The structure of FGRFNet.
Entropy 26 00029 g002
Figure 3. The structure of CPAM.
Figure 3. The structure of CPAM.
Entropy 26 00029 g003
Figure 4. The structure of ASFFM.
Figure 4. The structure of ASFFM.
Entropy 26 00029 g004
Figure 5. Training loss and test loss of the FGRFNet on the ADS-B, Zigbee, and LoRa datasets.
Figure 5. Training loss and test loss of the FGRFNet on the ADS-B, Zigbee, and LoRa datasets.
Entropy 26 00029 g005
Figure 6. Accuracy with different generation parameters.
Figure 6. Accuracy with different generation parameters.
Entropy 26 00029 g006
Figure 7. Accuracy with different data sizes and sample lengths.
Figure 7. Accuracy with different data sizes and sample lengths.
Entropy 26 00029 g007
Figure 8. Accuracy of CNN, SCNN, AlexNet, GoogLenet, ResNet-50, ICAM, MobileNetV3, and FGRFNet on the Zigbee dataset for different SNRs.
Figure 8. Accuracy of CNN, SCNN, AlexNet, GoogLenet, ResNet-50, ICAM, MobileNetV3, and FGRFNet on the Zigbee dataset for different SNRs.
Entropy 26 00029 g008
Figure 9. Confusion matrix of FGRFNet on the Zigbee dataset for different SNRs.
Figure 9. Confusion matrix of FGRFNet on the Zigbee dataset for different SNRs.
Entropy 26 00029 g009
Figure 10. Visualization of different Zigbee devices using Grad-CAM.
Figure 10. Visualization of different Zigbee devices using Grad-CAM.
Entropy 26 00029 g010
Table 1. Comparison of different methods on the ADS-B dataset.
Table 1. Comparison of different methods on the ADS-B dataset.
MethodBackboneAccuracy (%)F1-ScoreAUC
PMG [47]-79.180.78910.9734
CNN [33]-79.370.79050.9866
AlexNet [46]-82.350.82250.9837
API-Net [29]ResNet-10182.730.81920.9879
CSIL [48]-83.500.83240.9882
ACNet [30]ResNet-5084.410.83260.9881
MobileNetV3-Large [49]-85.290.85160.9884
ResNet-50 [25]-85.030.84800.9896
ResNet-101 [25]-86.200.86030.9899
SEF [50]ResNet-5086.990.86820.9923
ICAM [45]ResNet-5087.410.87320.9925
GoogLenet [46]-87.760.87670.9957
oursResNet-5089.240.89130.9984
oursResNet-10189.860.89820.9989
Table 2. Comparison of different methods on the Zigbee and LoRa datasets.
Table 2. Comparison of different methods on the Zigbee and LoRa datasets.
MethodBackboneAccuracy (%)F1-ScoreAUC
ZigbeeLoRaZigbeeLoRaZigbeeLoRa
CNN [33]-97.6367.200.97590.66730.99960.9633
SCNN [51]-96.8376.450.96630.76110.99970.9845
AlexNet [46]-96.2261.650.95900.59810.99930.9569
GoogLenet [46]-97.7878.800.97580.78460.99940.9821
ResNet-50 [25]-98.1678.350.98120.78130.99960.9855
ICAM [45]ResNet-5098.9981.400.98980.81040.99970.9883
MobileNetV3 [49]-98.6480.450.98630.80510.99970.9886
ARFNet [19]ResNet-5098.3579.410.98300.99960.79290.9865
oursResNet-5099.5783.000.99570.82940.99990.9915
Table 3. Ablation study of the attention module.
Table 3. Ablation study of the attention module.
MethodAccuracy (%)
ADS-BZigbeeLoRa
Backbone85.0398.1678.35
Backbone + CAM87.1498.2681.10
Backbone + PAM86.2698.3080.75
Backbone + CPAM87.8498.9382.05
Table 4. Ablation study of our method.
Table 4. Ablation study of our method.
MethodAccuracy (%)
ADS-BZigbeeLoRa
Backbone85.0398.1678.35
Backbone + CPAM87.8498.9382.05
Backbone + ASFFM87.4698.5781.55
Backbone + CPAM + ASFFM89.2499.5783.00
Table 5. Accuracy comparison when varying the distance between the transmitters and the receiver.
Table 5. Accuracy comparison when varying the distance between the transmitters and the receiver.
MethodAccuracy (%)
5 m15 m
CNN [33]60.8033.05
SCNN [51]62.5533.60
AlexNet [46]57.2533.80
GoogLenet [46]62.0537.10
ResNet-50 [25]66.9037.70
ICAM [45]66.2033.75
MobileNetV3 [49]66.3537.50
ours68.4039.30
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Hu, J.; Jiang, R.; Lin, Z.; Chen, Z. Fine-Grained Radio Frequency Fingerprint Recognition Network Based on Attention Mechanism. Entropy 2024, 26, 29. https://doi.org/10.3390/e26010029

AMA Style

Zhang Y, Hu J, Jiang R, Lin Z, Chen Z. Fine-Grained Radio Frequency Fingerprint Recognition Network Based on Attention Mechanism. Entropy. 2024; 26(1):29. https://doi.org/10.3390/e26010029

Chicago/Turabian Style

Zhang, Yulan, Jun Hu, Rundong Jiang, Zengrong Lin, and Zengping Chen. 2024. "Fine-Grained Radio Frequency Fingerprint Recognition Network Based on Attention Mechanism" Entropy 26, no. 1: 29. https://doi.org/10.3390/e26010029

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop