Open Access
ARTICLE
Lightweight Meta-Learned RF Fingerprinting under Channel Imperfections for 6G Physical Layer Security
Department of Electronic Engineering, National Formosa University, Yunlin, Taiwan
* Corresponding Author: Chia-Hui Liu. Email:
(This article belongs to the Special Issue: Artificial Intelligence for 6G Wireless Networks)
Computer Modeling in Engineering & Sciences 2026, 146(3), 40 https://doi.org/10.32604/cmes.2026.077837
Received 17 December 2025; Accepted 11 February 2026; Issue published 30 March 2026
Abstract
Artificial Intelligence (AI)-native sixth-generation (6G) wireless networks require data-efficient and channel-resilient physical-layer modeling techniques that learn stable device-specific representations under channel variations and hardware imperfections to support secure and reliable device-level authentication under highly dynamic environments. In such networks, massive device heterogeneity and time-varying channel conditions pose significant challenges, as reliable authentication must be achieved with limited labeled data and constrained edge resources. To address this challenge, this paper proposes an Artificial Intelligence (AI)-assisted few-shot physical-layer modeling framework for channel robust device identification, formulated within the paradigm of Specific Emitter Identification (SEI) based on radio frequency (RF) fingerprinting. The proposed framework explicitly formulates few-shot SEI as a channel-resilient physical-layer modeling problem by integrating a lightweight convolutional neural network and Transformer hybrid encoder with a dual-branch feature decoupling mechanism. Device specific RF fingerprints are separated from channel-dependent factors through orthogonality-constrained learning, which effectively suppresses channel-induced prototype drift and stabilizes metric geometry under channel variations. A meta-learned prototypical inference module is further employed under episodic few-shot training, enabling rapid adaptation to new devices and unseen channel conditions using only a small number of labeled samples. Experimental results on multiple real-world RF datasets, including ORACLE Wi-Fi transmitter measurements and civil aviation ADS-B broadcasts (Keywords
With the rapid expansion of the Internet of Things (IoT) and industrial wireless communication systems, future network architectures must maintain stable and reliable terminal identification capabilities over long-term operations. Sole reliance on software-level credentials, such as cryptographic keys or serial numbers, is vulnerable to tampering and leakage. Moreover, frequent device replacement, high labeling costs, and constrained edge computing resources highlight the need for lightweight, self-adaptive, and scalable device-level authentication mechanisms. Against this backdrop, Specific Emitter Identification (SEI), also commonly referred to as radio frequency (RF) fingerprinting, exploits device-specific hardware imperfections introduced during RF manufacturing to identify emitters directly from physical-layer signals, providing a security mechanism independent of software-level information. This capability is particularly critical in AI-native sixth generation (6G) communication systems, where intelligent and self-adaptive signal management and authentication are expected to be achieved through learning-driven models. Early SEI methods based on traditional signal processing predominantly relied on handcrafted features and pipeline-based workflows. Typical approaches extracted RF impairments such as I/Q imbalance, carrier frequency offset (CFO), and phase noise from steady-state or transient signal segments, followed by classification using higher-order statistics and time frequency transformations, including Short-Time Fourier Transform (STFT), Continuous Wavelet Transform (CWT), Fractional Fourier Transform (FrFT), and bispectrum analysis. Although these methods offer strong interpretability, their performance degrades under low signal-to-noise ratio (SNR) and highly dynamic channel conditions, where noise, multipath effects, interference, and frequency drift tend to obscure device-specific characteristics. With the advancement of deep learning, end-to-end feature extraction has been increasingly adopted in SEI. Convolutional Neural Networks (CNNs) and hybrid CNN Transformer architectures have demonstrated improved identification performance by learning hierarchical and global feature representations directly from raw or transformed signals. However, these models typically depend on large-scale labeled datasets and assume distributional consistency, limiting their generalization to new devices or open-set environments. In practical deployments, data scarcity and heterogeneity are inevitable due to privacy constraints and high acquisition costs. To address these challenges, Few-Shot Learning (FSL) combined with Meta-Learning has emerged as a promising solution. Metric-based meta-learning approaches, such as prototypical networks, learn task-invariant embedding spaces by constructing stable geometric structures across episodic training tasks, enabling rapid adaptation to new devices and unseen channel conditions while maintaining robust decision boundaries.
The system models the wireless emission process through an AI-based feature decoupling and prototypical metric framework. Raw I/Q signals collected through transmitter channel receiver paths are normalized and fed into a lite CNN Transformer encoder that learns device-specific and channel-invariant representations. The model serves as a computational surrogate for hardware-level imperfections, enabling device-level authentication and physical-layer security in AI-native 6G and IoT networks. In summary, this study addresses the challenges of entangled device fingerprints and channel effects in SEI tasks, as well as the difficulty of achieving stable identification under few-shot conditions. To this end, we propose the FS-SEI framework, which integrates feature disentanglement with prototypical meta-learning, as illustrated in Fig. 1. In the proposed framework, device-related and channel-related features are explicitly disentangled during the feature extraction stage, with orthogonality constraints imposed to preserve their independence. At the decision stage, a prototypical space is constructed solely from device features to perform metric-based classification, thereby mitigating the impact of channel perturbations on the decision boundaries. Through an episodic meta-learning procedure and a channel parameter sweeping training strategy covering simulated disturbances such as Carrier Frequency Offset (CFO), Additive White Gaussian Noise (AWGN), and I/Q imbalance the model autonomously calibrates the embedding space across varying channel conditions. This enables rapid alignment with new class distributions under limited samples, achieving high stability and reproducibility. Overall, the proposed design not only provides an AI-driven approach for physical-layer modeling and authentication but also aligns with the requirements of AI-native 6G architectures, which emphasize self-adaptive physical-layer modeling and device-level identity verification. This framework offers a stable and scalable foundation for secure modeling in future intelligent communication systems and edge networks. As 6G communication architectures evolve toward AI-native and highly decentralized paradigms, future network nodes will no longer rely on traditional centralized core networks. Instead, they will be widely deployed across multi-tier infrastructures such as cell-free massive Multiple-Input Multiple-Output (MIMO), Reconfigurable Intelligent Surface (RIS) assisted communications, and integrated satellite terrestrial networks. In such low-latency, highly dynamic wireless environments, base stations, edge nodes, and unmanned aerial vehicle (UAV) relays must rapidly identify whether surrounding devices are trusted nodes, even under unpaired or extremely limited-sample conditions. Therefore, Few-Shot Specific Emitter Identification (FS-SEI), which exhibits few-shot capability and stable feature representation under unknown channel conditions and hardware perturbations, is poised to become a foundational mechanism for device authentication and physical-layer security in 6G systems directly addressing the core requirements of the model proposed in this study. Unlike existing FS-SEI methods that primarily rely on either meta-learning or feature disentanglement in isolation, the proposed framework explicitly couples orthogonality-guided feature decoupling with episodic meta-learning, enabling the learned metric geometry to remain stable under channel variations while preserving rapid few-shot adaptability.

Figure 1: Overall framework of the proposed AI-assisted FS-SEI computational model.
The main contributions of this work are summarized as follows:
• We formulate few-shot Specific Emitter Identification (FS-SEI) as a channel-resilient physical-layer modeling problem, explicitly addressing the challenges of data scarcity, channel variation, and device heterogeneity in AI-native 6G wireless environments.
• We propose a lightweight meta-learned FS-SEI framework that integrates a CNN Transformer hybrid encoder with a dual-branch feature disentanglement mechanism, enabling explicit separation of device-specific RF fingerprints from channel-dependent factors.
• An orthogonality-guided representation stabilization strategy is introduced to suppress feature entanglement and mitigate channel-induced prototype drift, ensuring stable metric geometry for prototype-based inference under dynamic channel conditions.
• A prototypical metric learning module is employed under episodic meta-learning, allowing rapid adaptation to new devices and unseen channel conditions using only a small number of labeled samples, without relying on large-scale pretraining or complex hybrid pipelines.
• Extensive experiments on multiple real-world RF datasets (
2.1 Large-Scale Deep Specific Emitter Identification
In the field of SEI, deep learning has become the dominant approach for building high-performance systems due to its strong feature extraction and nonlinear modeling capabilities. Under controlled conditions with large-scale labeled datasets, Convolutional Neural Networks (CNN) effectively suppress channel noise and signal distortion, while Complex-Valued Neural Networks operate directly on complex baseband signals to enhance RF feature representation. These models have established strong performance baselines, and early supervised studies formulated SEI as an end-to-end representation learning problem using standard CNN architectures [1,2]. To address deployment constraints on resource-limited edge devices, subsequent works combined one-dimensional CNNs (1D-CNNs) with Discrete Wavelet Transform (DWT) for dimensionality reduction and complexity mitigation [3], achieving superior accuracy over traditional machine learning methods even under low signal-to-noise ratio (SNR) conditions. Network compression and lightweight design techniques were further introduced to reduce the computational cost of Complex-Valued Neural Networks (CVNNs), enabling near-perfect accuracy and significant latency reduction in high-SNR scenarios [4]. Recent research has focused on improving robustness to domain shifts caused by channel variation and receiver diversity. Representative approaches include multi-receiver collaboration and Long Short-Term Memory (LSTM)-based temporal transfer modeling [5], which enhance cross-environment discriminability through multi-perspective signal decomposition. A domain-adversarial Transformer encoder has been proposed to mitigate performance degradation caused by limited annotations and distribution mismatch [6]. In addition, adaptive strategies combining contrastive learning with triplet loss, entropy minimization, self-supervised pseudo-labeling, and contrastive regularization have been introduced to adapt to receiver characteristics under unlabeled settings [7]. Bispectrum-based feature fusion with CNN decision modules has also been explored to improve stability in distributed sensing scenarios [8], often combined with transfer learning to reduce cold-start effects and convergence time. Despite these advances, deep learning-based SEI fundamentally depends on large-scale, high-quality labeled data. In real-world and dynamic environments, device identification and ground-truth annotation are difficult to obtain, and data collection is further constrained by privacy and regulatory issues. Moreover, distribution shifts caused by channel drift and receiver variability limit the effectiveness of conventional supervised strategies in maintaining stable performance and enabling rapid deployment. This mismatch between data-intensive deep learning models and sparse-data operational scenarios significantly restricts their applicability to few-shot and rapid adaptation tasks. These challenges highlight the growing need for data-efficient and adaptive physical-layer modeling frameworks, particularly in the context of future 6G and large-scale IoT networks. Despite their strong performance under abundant labeled data and controlled conditions, large-scale deep SEI models are inherently ill-suited for few-shot 6G scenarios, as their reliance on data-intensive training and heavyweight architectures limits adaptability under data scarcity and restricts practical deployment on resource-constrained edge devices.
Against this background, few-shot learning has emerged as a key viable solution to enhance the adaptability and efficiency of SEI systems in dynamic environments. In particular, for 6G and edge wireless application scenarios, FSL methods that can maintain identification stability under extremely low annotation and highly dynamic channel conditions are of critical importance for enabling AI-assisted device-layer modeling and communication security. The fundamental advantage of this approach lies in its ability to rapidly establish and maintain stable class decision boundaries using only a minimal number of support samples.
1. Transfer-Based Fine-Tuning: Among existing Few-Shot Supervised Environmental Inference (FS-SEI) methods, transfer-based FSL is widely adopted. These methods typically follow a two-stage paradigm consisting of self-supervised pretraining and subsequent (semi-)supervised fine-tuning. The pretraining stage aims to learn transferable and domain-invariant representations, which are then adapted to downstream FS-SEI tasks using limited labeled data. Recent studies primarily employ self-supervised learning (SSL) for pretraining. Contrastive self-supervised learning (CSSL) frameworks have been shown to improve robustness to noise and reduce sensitivity to weight initialization [9]. Masked modeling approaches, such as the Asymmetric Masked Autoencoder (AMAE) [10], further enhance annotation efficiency and feature transferability in RF applications. In addition, combining self-supervised pretraining with adversarial augmentation has been proposed to learn more flexible and cross-source discriminative radio frequency fingerprint (RFF) representations [11]. The main advantage of transfer-based methods lies in their high training parallelism and compatibility with existing pretrained models. However, their performance is limited by the domain gap between pretraining data and the target FS-SEI task.
2. Metric-Based Learning: In contrast to transfer-based approaches, metric-based FSL learns a discriminative embedding space and performs classification using distance or similarity measures. Instead of fixed classifiers, it relies on prototype-based decision structures to directly model semantic relationships, making it well suited for few-shot scenarios with limited labeled data. The performance of metric-based methods depends on the quality and stability of the embedding space. When class numbers increase or intra-class variation is large, distance reliability and classification accuracy may degrade. To improve robustness, several extensions have been proposed. Metric Adversarial Training (MAT) [12] combines adversarial regularization on unlabeled data with metric constraints on labeled samples, enhancing stability under low annotation ratios. Attention mechanisms such as the Convolutional Block Attention Module (CBAM) adaptively weight time–frequency features [13], while robust representation learning for SEI under class-irrelevant factors has also been explored to improve generalization and robustness in complex environments [14]. These approaches improve robustness to domain shifts and channel perturbations. Overall, metric-based learning provides a lightweight, interpretable, and effective framework for FS-SEI, and is well suited for device identification and physical-layer security in future 6G wireless systems.
3. Hybrid Models and Frameworks: As few-shot recognition tasks become more complex, recent studies have adopted hybrid frameworks that integrate multiple learning paradigms, including metric learning, self-supervised learning, transfer learning, and traditional signal processing, to improve FS-SEI robustness and generalization. These frameworks typically employ multi-branch architectures to balance fine-grained feature extraction with global representation stability. For example, the Asymmetric Dual-Path Masked Autoencoder (ADPMAE) framework [15] enhances discriminative capability under limited samples by jointly reconstructing local signal details and modeling global semantic structures. The Knowledge–Data–Model (KDM) driven framework [16] combines expert-crafted features, self-supervised learning, and supervised FS-SEI to achieve strong robustness and high processing efficiency. In addition, a hybrid FS-SEI framework based on feature contrastive reconstruction and virtual adversarial training (VAT) [17] demonstrates effective performance in both self-supervised and semi-supervised scenarios under complex electromagnetic environments. This model adopts a CNN–Transformer architecture to capture temporal and spectral characteristics, enabling a balanced representation of global semantics and local structures. Despite their strong performance, hybrid frameworks often suffer from increased structural complexity, high computational cost, and extended training time, which significantly limit their practicality for deployment. As future 6G and intelligent edge networks impose strict constraints on computational resources, lightweight and scalable AI modeling strategies are increasingly required. Consequently, designing hybrid FS-SEI frameworks that balance performance and computational efficiency remains a critical research challenge.
As summarized in Table 1, large-scale deep SEI methods, while effective under abundant labeled data, are fundamentally mismatched with few-shot 6G scenarios due to their data-intensive training requirements and heavyweight model architectures, which limit adaptability under data scarcity and hinder deployment on resource-constrained edge devices.

2.3 Feature Decoupling and Orthogonal De-Correlation
In SEI, device-invariant factors namely stable radio frequency fingerprints (RFFs) caused by hardware imperfections are often entangled with channel-variant factors within the same embedding space. This entanglement leads to prototype drift and unstable decision boundaries. Feature decoupling aims to explicitly separate these two types of factors within the embedding layer, making it particularly suitable for FSL scenarios. By constructing class prototypes primarily based on device-related components, feature decoupling can significantly enhance intra-class compactness and inter-class separability, thereby improving classification stability. Relevant studies have used feature disentanglement to suppress class-irrelevant factors and improve discriminability [18], ensuring that prototype distance calculations are predominantly influenced by key signal peaks, which represent device-specific characteristics, thus reinforcing feature discriminability. Other approaches employ auxiliary tasks such as phase shift prediction [19] during the pretraining stage to learn both phase-relevant and irrelevant signal components. Orthogonality constraints are then applied in downstream FS-SEI tasks to precisely disentangle device-related features from channel-induced variations. Such methods have been shown to significantly outperform various deep learning baseline models on
3 Methodology and Framework Design
Under the vision of 6G communication architectures, achieving robust physical-layer modeling and device-level identification under limited supervision is one of the key challenges in intelligent communications. In this context, we propose a Few-Shot Specific Emitter Identification framework, which formulates SEI as a metric-based meta-learning task. The objective is to learn self-adaptive device representations that remain stable under highly dynamic channel conditions.
3.1 Problem Definition: Few-Shot SEI under Multi-Domain Channel Variations
To address the challenges of data scarcity and high variability in specific emitter (SE) signals, this study formulates SEI as a few-shot metric learning problem. Furthermore, SEI is treated as an AI-driven physical-layer signal modeling task aimed at learning robust and generalizable representations under multi-domain channel conditions. Let the set of source entities be denoted as c = {1, …, N}, where only k labeled samples are available per class. Each episodic training instance follows the N-way k-shot, q-query scheme, in which N classes (i.e., devices) are randomly sampled from the training set. For each class, k support samples from the support set S, and q samples from the query set Q. The input consists of I/Q signal segments of fixed length, which are encoded by a feature extractor Q
3.2 Meta-Learned SEI Framework for AI-Native 6G Communications
This study proposes a Few-Shot FS-SEI framework based on the concept of metric-based meta-learning, as illustrated in Fig. 2. The framework can be viewed as an AI-driven physical-layer modeling pipeline designed to simulate and learn stable representations of device-level RF behavior under channel perturbations. Given input I/Q signal segments, the model first employs a dual-branch feature extractor to explicitly decouple the representations into two components: a device-specific feature

Figure 2: Few-shot SEI workflow.
3.2.1 Feature Disentanglement for Robust Physical-Layer Representation Learning
Prior to prototype generation, the feature extractor module is responsible for extracting and disentangling device-intrinsic characteristics from channel-variant factors embedded in raw radio frequency signals. The primary design objective is to suppress channel perturbations such as path loss, obstruction, multipath effects, and hardware gain discrepancies. By mitigating these disturbances, the module outputs stable and quantifiable device representations to the decision-making stage. The overall framework comprises three interconnected components: (1) a lightweight hybrid backbone, (2) a dual-branch feature decoupling module, and (3) a sequence-level integration and decorrelation unit.
(1) Lightweight Hybrid Backbone for I/Q Representation: I/Q segments from both the support and query sets are first processed by the backbone to obtain intermediate sequential representations. The backbone adopts a hybrid architecture consisting of a front-end convolutional neural network (CNN) and a back-end lightweight Transformer. This design captures both local temporal or time frequency textures and long-range dependencies while effectively suppressing redundant information. The output is a scale-consistent feature tensor that facilitates the subsequent separation of shared and device-specific features. It also serves as the common input to both the decoupling and metric learning modules.
(2) Dual-Branch Feature Decoupling for Shared and Specific Components: This module is embedded within the CNN layers of the backbone to generate the shared feature
(3) Orthogonality-Guided Representation Stabilization: A lightweight Transformer is employed to enhance long-range temporal dependencies while suppressing redundancy along the channel dimension. Through an attention mechanism, dynamic weighting is applied to both shared and specific features, enabling global context aggregation at low computational cost. The integrated representation is subsequently projected back into two complementary subspaces to obtain the final outputs
After feature disentanglement, an orthogonality constraint is imposed during the training phase to explicitly suppress residual statistical dependence between the two subspaces. This constraint is applied exclusively to the batch-normalized representations
The orthogonality constraint is applied uniformly across all experiments and datasets and is jointly optimized with the classification objective, without dataset-dependent weighting or tuning. This constraint maintains linear independence and suppresses feature entanglement. By enforcing orthogonality at the representation level, channel-related variations are prevented from perturbing the device-specific subspace, ensuring that subsequent prototype construction relies on stable and discriminative device-level attributes. In the subsequent decision-making stage, only
The lightweight nature of the proposed framework is not achieved merely by reducing model size, but is jointly constrained by architectural and training design choices. On the feature extraction side, a lite CNN Transformer hybrid encoder is adopted, explicitly avoiding global self-attention with quadratic complexity in sequence length, resulting in near-linear scaling during training and inference. On the decision side, prototype-based metric inference is employed, eliminating large learnable classifier weights and concentrating parameters on the shared backbone and lightweight disentanglement modules. Compared to hybrid FS-SEI frameworks that rely on self-supervised pretraining, multi-branch reconstruction, contrastive learning, or VAT, the proposed method requires only a single episodic supervised meta-learning procedure, substantially reducing training complexity and deployment overhead. Quantitatively, the proposed model contains approximately 1.45 × 105 trainable parameters, which is significantly lower than typical CNN/Transformer-based architectures. Under the most demanding
3.2.2 Prototype-Guided Metric Module for Few-Shot Device-Level Identification
This study adopts a metric-based inference head as the final prediction module. The core principle aligns with that of the Prototypical Network (ProtoNet), relying on class prototypes and metric distance calculations. Rather than depending on a large number of learnable classification weights, the model performs inference through a measurable geometric structure in the embedding space. This approach better conforms to the modeling philosophy of AI-driven physical-layer representation learning. In each N-way k-shot episodic task, the device-specific features
Formally, for the support subset Sc corresponding to class c, the prototype vector is defined as:
This step aggregates the device-specific features of a limited number of support samples into class prototypes via mean pooling. Under conditions of scarce data and varying channel environments, this operation helps mitigate the impact of sample-specific noise on the decision geometry, providing a stable anchor for subsequent metric evaluation.
For any query sample
where
The distance is transformed into a probability by applying the Softmax function to the negative distance, thereby obtaining the posterior probability P that a query sample belongs to class c as Eq. (4).
The prototype classification loss is defined as Eq. (5). During training, we compute the average cross-entropy loss over the query set Q as the prototype classification loss
where
The overall training objective is given by:
The parameter
3.2.3 Hybrid Meta-Learning Strategy for Cross-Channel Robustness
Under few-shot SEI episodic training, prototype estimation often exhibits high variance and unstable convergence due to task sampling variability and channel perturbations, resulting in gradient oscillation and decision boundary drift. To address these issues, this study proposes a supervised hybrid meta-learning framework to improve training stability and cross-channel generalization. The training follows a bi-level optimization architecture. At the inner level, N-way k-shot, q-query episodic tasks are used to preserve few-shot adaptation characteristics. At the outer level, losses from multiple episodic tasks are aggregated, and a single meta-update is performed to control convergence behavior. Within each episode, support and query samples are processed by a dual-branch encoder to extract shared and distinctive features. Metric-based classification is performed using only the distinctive features to construct class prototypes, ensuring that learning focuses on device-specific characteristics while remaining robust to channel variations. After aggregating losses across tasks, model parameters are updated once using a unified learning rate and a positive step size, which stabilizes prototype geometry and mitigates gradient direction drift. Training and validation both follow a fixed episodic protocol, with statistical performance monitoring used for stability assessment and early stopping. This approach enables stable few-shot training without requiring an additional pretraining stage, while maintaining strong adaptability and cross-channel robustness. Furthermore, the proposed pipeline is reproducible, scalable, and consistent with self-adaptive learning principles in sixth-generation (6G) communication systems, providing a practical foundation for physical-layer modeling and device-level authentication tasks.
3.2.4 Geometric Modeling of Device-Specific Prototypes
The core objective of the proposed prototype space design is to learn a geometrically stable and semantically consistent device representation space, rather than merely increasing classifier capacity. To this end, prototypes are constructed exclusively from device-specific features, without incorporating channel-related shared components. This design prevents prototype positions from being distorted by varying channel conditions, allowing each prototype to focus on capturing intrinsic hardware characteristics of the device, thereby enhancing stability across different channels and tasks. Compared to standard Prototypical Networks that compute prototypes using holistic embeddings, the proposed method explicitly restricts prototype construction to a device-related subspace through feature disentanglement and orthogonality constraints, conceptually redefining the semantic meaning and geometric role of prototypes. Under this setting, distance measurements in the embedding space possess clear geometric interpretation; therefore, Euclidean distance is adopted as the metric function to directly reflect the actual displacement between samples and their corresponding prototypes. The orthogonality constraint further suppresses mutual interference between device-specific and channel-related features, encouraging tighter intra-class clustering and clearer inter-class separation of prototypes within the embedding space. As a result, prototype drift under few-shot conditions and channel variations is effectively reduced, leading to improved overall decision stability.
3.3 Channel Perturbation Modeling, Simulation Design, and Evaluation Protocol
To consistently characterize multi-domain channel variation in highly dynamic wireless environments during both training and evaluation, this study applies controlled channel perturbations to the input I/Q segments during episode construction. A parameter sweep simulation is employed to systematically cover common front-end imperfections and channel conditions. The configurations and ranges of all perturbations are aligned with those used in the experimental section (see Table 1), and the same perturbation scheme is applied across the train, validation, and test subsets to ensure reproducibility and fair comparison.
(A) Unified Channel Perturbation Model and Parameter Ranges:
• Carrier Frequency Offset (CFO): Introduces a linear phase rotation in the time domain or a frequency shift in the frequency domain to simulate local oscillator drift and Doppler effects. The perturbation range is set to {−30, 50}.
• Additive White Gaussian Noise (AWGN): Gaussian white noise is added to the modulated signal to simulate various signal-to-noise ratio (SNR) conditions and evaluate robustness under weak signal scenarios. The perturbation range is set to {10, 30}.
• I/Q Imbalance: Both amplitude and phase asymmetries are introduced to emulate analog front-end imperfections, thereby reducing the model’s reliance on memorized fixed gain or phase patterns. This encourages the distinctive branch to focus on subtle distortions related to device manufacturing variations. The perturbation ranges are set to amplitude {−10, 10} and phase {−5%, 5%}.
The aforementioned perturbations are consistently applied to both the support and query sets during episode generation. A grid-based parameter sweep is employed to cover multiple combinations, thereby forming an in-silico population of channel variations. This design fulfills the AI-native communication modeling requirements emphasized in 6G and large-scale IoT systems by enabling the learning of channel-resilient representations under multi-domain variations.
(B) Task-Oriented Configuration and Data Partitioning: The dataset is partitioned into training, validation, and testing subsets with a ratio of 7:2:1. Each episode follows a fixed N-way, 10-shot, 15-query configuration, simulating a continuous frame-level monitoring scenario. For additional few-shot analysis, varying values of k ∈ {5, 10, 20, 30, 40} are also evaluated. To prevent information leakage, no consecutive time segments are allowed to span across different data subsets. The distribution of source entities (classes) and channel condition ratios is maintained consistently across the training, validation, and test sets. All segments are normalized using z-score standardization. Channel perturbations are applied uniformly to all three subsets following the principles outlined in Section (A).
(C) Evaluation Protocol and Metrics: During the testing phase, a fixed model with parameters
4.1 Dataset Preprocessing and Experimental Setup
In this study, the proposed Few-Shot FS-SEI method is evaluated on three open-source Radio Frequency Fingerprinting (RFF) datasets:
Few-shot SEI is highly susceptible to distributional shifts arising from source entity variations and environmental diversity, such as distance, occlusion, multipath effects, and mobility. When support samples are limited, prototype estimation is prone to bias, which may destabilize the metric geometry. To mitigate this issue, a fixed data preprocessing protocol tightly coupled with the FS-SEI framework is established prior to training:
• Fixed-Length Slicing: I/Q segments of fixed length (as specified in Table 1) are extracted using the packet synchronization sequence as an anchor. This ensures sample homogeneity and enhances information density.
• Z-Score Normalization: Each segment is standardized using z-score normalization to eliminate variations in energy scale and gain. This emphasizes waveform shape and phase relationships over absolute amplitude, thereby reducing interference from scale differences in orthogonal feature decoupling and prototype-based metric learning.
• Balanced Sampling: Samples are uniformly drawn across distance intervals to maintain consistent proportions of devices and channel conditions. This mitigates distributional shift and ensures consistency across training, validation, and testing phases.
After the aforementioned preprocessing, the input distribution is aligned with the training objective, enabling the formation of discriminative and generalizable class geometries within the prototypical metric space.
Channel perturbation modeling and episodic evaluation protocols strictly follow the unified design described in Section 3.3.
All experiments were conducted on a Windows platform with a 13th Gen Intel Core i5-13400 CPU, an NVIDIA GeForce RTX 3060 GPU (12 GB), and 32 GB RAM, using Python 3.10 and PyTorch 2.5.x. Model evaluation followed a fixed N-way K-shot episodic learning protocol. During testing, multiple independent episodes were randomly generated, and performance was reported as the mean ± standard deviation of query-set accuracy across episodes to reduce stochastic sampling bias and capture episode-level uncertainty. Balanced sampling and z-score normalization were applied to ensure fair comparison across devices and distance conditions by mitigating signal energy scaling and gain variations. Channel effects were simulated by synchronously applying CFO, AWGN, and I/Q imbalance to both support and query samples within each episode, avoiding intra-episode distribution mismatch. Perturbation parameters were selected from discrete ranges and combined via a grid-based sweep, with perturbation configurations randomly sampled per episode and consistently applied across training, validation, and testing to ensure fairness and reproducibility. The CFO, SNR, and I/Q imbalance ranges were designed to emulate common wireless hardware and channel non-idealities, including oscillator offsets, Doppler effects, noise variations, and RF front-end amplitude and phase mismatches, and are summarized in Table 2. We assume consistent receiver configurations within each dataset, with dataset-specific packet alignment performed when required (e.g., synchronization-based alignment for


4.2 Experimental Results and Analysis
To evaluate the robustness and modeling capability of the proposed FS-SEI framework under multi-channel and few-shot conditions, this section presents and analyzes its overall performance on three open-source radio frequency datasets:

Figure 3: Confusion matrices under the 10-shot setting during testing: (a)

Figure 4: t-SNE visualization of RFF features under the 10-shot setting: (a)
The visualization reflects the intermediate embedding space prior to prototype-based metric inference. Classification accuracy reported in Table 4 is evaluated in the meta-learned prototypical decision space rather than in this low-dimensional projection.

Fig. 4 presents the t-SNE visualization of RFF features under the 10-shot setting, reflecting the feature distributions across different datasets. Subfigure (a) illustrates the feature distribution of the
Due to its lightweight architectural design, the proposed model contains only 144,712 parameters, resulting in minimal increases in FLOPs. The computational load is significantly lower than that of typical convolutional neural networks, transformers, and other deep architectures [9,14]. This demonstrates that the model maintains both high identification performance and computational efficiency under few-shot conditions.
Table 2 summarizes the quantitative results across different values of k. When k increases from 5 to 10, there is a marked improvement in accuracy and a reduction in variance, indicating more stable prototype estimation and sharper decision boundaries. As k further increases to the range of 20 to 40, the overall accuracy begins to saturate, suggesting that under the current perturbation and intra-class variation conditions, the 10-shot setting is sufficient to capture the essential prototype features, with diminishing returns from additional samples. This trend is consistently observed across all three datasets, demonstrating that the proposed framework achieves rapid convergence and stable generalization even under extremely low-sample conditions.
Based on the results in Table 2, the proposed FS-SEI model demonstrates high accuracy and stability across all three open-source radio frequency datasets. Overall, the 10-shot setting offers the best trade-off between performance and computational cost, achieving high accuracy rates of 0.991, 0.994, and 0.998 on the

Figure 5: Comparison with recent RFF FS-SEI methods in
On the other hand, simulation results demonstrate the high value of the proposed architecture for classifying RFF device samples. The method integrates feature decoupling, lightweight temporal fusion, and a prototype-based decision module, achieving a balance between identification performance and computational efficiency. In contrast to common approaches that use short windows combined with deep stacking, which introduces additional computational cost, or rely on computationally expensive CNN Transformer architectures, the proposed method avoids the quadratic complexity of global self-attention with respect to sequence length. As a result, both training and inference exhibit computational complexity that scales more linearly. This architecture offers high information density and low resource demand on edge devices, aligning with the energy efficiency and deploy ability requirements of sixth-generation communication systems.
This study has several validity-related limitations that should be acknowledged. Regarding internal validity, the proposed model is evaluated under episodic few-shot learning with controlled experimental settings, which may not capture all sources of randomness in real-world deployments. In terms of construct validity, although real-world RF datasets are used, channel impairments such as CFO, AWGN, and IQ imbalance are applied in a controlled in-silico manner to enable reproducible analysis of channel robustness. With respect to external validity, the current evaluation assumes a fixed receiver hardware chain, and cross-receiver heterogeneity is not explicitly modeled. These factors may limit direct generalization to heterogeneous receiver environments and motivate future extensions of the proposed framework.
This study addresses the problem of SEI under highly dynamic channel conditions by proposing a few-shot modeling framework that integrates feature disentanglement, prototypical metric learning, and meta-learning. The proposed architecture adopts a CNN Transformer hybrid backbone with a dual-branch disentanglement module, enabling explicit separation of device-invariant and channel-variant factors at the feature level. Orthogonality constraints are applied to enforce their independence, allowing the model to stably capture the distinctiveness of Radio Frequency Fingerprints under conditions of severe channel variability and data scarcity. Through controlled channel perturbations and parameter sweep-based numerical simulations, the model autonomously learns stable representations under varying channel conditions, demonstrating the feasibility of AI-assisted physical-layer modeling. Experimental results show that FS-SEI achieves high accuracy, low variance, and stable convergence across multiple real-world radio frequency datasets including
Channel perturbations in this study are constructed in a controlled in-silico manner (CFO/AWGN/IQ imbalance), which improves reproducibility but does not fully capture more complex real-world propagation effects and receiver-side hardware heterogeneity; therefore, the conclusions mainly reflect few-shot SEI behavior under a fixed receiver-chain assumption. In addition, prototype-based inference requires computing distances between each query and all class prototypes, leading to inference cost that grows approximately linearly with N-way and may affect real-time scalability when the number of classes becomes very large. Future work will focus on: (i) modeling receiver heterogeneity and enabling online adaptation; (ii) extending channel modeling and/or conducting hardware-in-the-loop validation; and (iii) improving large-scale inference efficiency via hierarchical prototypes, candidate set reduction, or approximate nearest neighbor retrieval.
Acknowledgement: Not applicable.
Funding Statement: The authors received no specific funding for this study.
Author Contributions: The authors confirm contribution to the paper as follows: study conception and design: Chia-Hui Liu; data collection: Chia-Hui Liu and Hao-Feng Liu; analysis and interpretation of results: Chia-Hui Liu; draft manuscript preparation: Chia-Hui Liu. All authors reviewed and approved the final version of the manuscript.
Availability of Data and Materials: The datasets used in this study are publicly available open-source RFF datasets, including the ORACLE Wi-Fi dataset [20] and ADS-B datasets.
Ethics Approval: Not applicable.
Conflicts of Interest: The authors declare no conflicts of interest.
References
1. Ding L, Wang S, Wang F, Zhang W. Specific emitter identification via convolutional neural networks. IEEE Commun Lett. 2018;22(12):2591–4. doi:10.1109/LCOMM.2018.2871465. [Google Scholar] [CrossRef]
2. Merchant K, Revay S, Stantchev G, Nousain B. Deep learning for RF device fingerprinting in cognitive communication networks. IEEE J Sel Top Signal Process. 2018;12(1):160–7. doi:10.1109/JSTSP.2018.2796446. [Google Scholar] [CrossRef]
3. Xie F, Wen H, Wu J, Chen S, Hou W, Jiang Y. Convolution based feature extraction for edge computing access authentication. IEEE Trans Netw Sci Eng. 2020;7(4):2336–46. doi:10.1109/TNSE.2019.2957323. [Google Scholar] [CrossRef]
4. Wang Y, Gui G, Gacanin H, Ohtsuki T, Dobre OA, Poor HV. An efficient specific emitter identification method based on complex-valued neural networks and network compression. IEEE J Sel Areas Commun. 2021;39(8):2305–17. doi:10.1109/JSAC.2021.3087243. [Google Scholar] [CrossRef]
5. He B, Wang F. Cooperative specific emitter identification via multiple distorted receivers. IEEE Trans Inf Forensics Secur. 2020;15:3791–806. doi:10.1109/TIFS.2020.3001721. [Google Scholar] [CrossRef]
6. Li D, Yao B, Sun P, Li P, Yan J, Wang J. Specific emitter identification based on ensemble domain adversarial neural network in multi-domain environments. EURASIP J Adv Signal Process. 2024;2024(42):1–15. doi:10.1186/s13634-024-01138-y. [Google Scholar] [CrossRef]
7. Yang J, Zhu S, Wen Z, Li Q. Cross-receiver radio frequency fingerprint identification: a source-free adaptation approach. Sensor. 2025;25(14):4451. doi:10.3390/s25144451. [Google Scholar] [PubMed] [CrossRef]
8. Liu M, Chai Y, Li M, Wang J, Zhao N. Transfer learning-based specific emitter identification for ADS-B over satellite system. Remote Sens. 2024;16(12):2068. doi:10.3390/rs16122068. [Google Scholar] [CrossRef]
9. Zhao D, Yang J, Liu H, Huang K. A complex-valued self-supervised learning-based method for specific emitter identification. Entropy. 2022;24(7):851. doi:10.3390/e24070851. [Google Scholar] [PubMed] [CrossRef]
10. Yao Z, Fu X, Guo L, Wang Y, Lin Y, Shi S, et al. Few-shot specific emitter identification using asymmetric masked auto-encoder. IEEE Commun Lett. 2023;27(10):2657–61. doi:10.1109/LCOMM.2023.3312669. [Google Scholar] [CrossRef]
11. Liu C, Fu X, Wang Y, Guo L, Liu Y, Lin Y, et al. Overcoming data limitations: a few-shot specific emitter identification method using self-supervised learning and adversarial augmentation. IEEE Trans Inf Forensics Secur. 2024;19:500–13. doi:10.1109/TIFS.2023.3324394. [Google Scholar] [CrossRef]
12. Fu X, Peng Y, Liu Y, Lin Y, Gui G, Gacanin H, et al. Semi-supervised specific emitter identification method using metric-adversarial training. IEEE Internet Things J. 2023;10(12):10778–89. doi:10.1109/JIOT.2023.3240242. [Google Scholar] [CrossRef]
13. Mu S, Zu Y, Chen S, Yang S, Feng Z, Zhang J. Few-shot metric learning with time-frequency fusion for specific emitter identification. Remote Sens. 2024;16(24):4635. doi:10.3390/rs16244635. [Google Scholar] [CrossRef]
14. Zhou Z, Li G, Wang T, Zeng D, Li X, Wang Q. A robust open-set specific emitter identification for complex signals with class-irrelevant features. IEEE Trans Inf Forensics Secur. 2025;20:6058–73. doi:10.1109/TIFS.2025.3576573. [Google Scholar] [CrossRef]
15. Yang S, Yu S, Li Q, Xia K, Zhu H. Few-shot specific emitter identification via asymmetric dual-path masked autoencoder. Digit Signal Process. 2025;163:105201. doi:10.1016/j.dsp.2025.105201. [Google Scholar] [CrossRef]
16. Sun M, Teng J, Liu X, Wang W, Huang X. Few-shot specific emitter identification: a knowledge, data, and model-driven fusion framework. IEEE Trans Inf Forensics Secur. 2025;20:3247–59. doi:10.1109/TIFS.2025.3550080. [Google Scholar] [CrossRef]
17. Sun M, Wei L, Yu C, Qiu Z, Teng J. Self-supervised and semi-supervised learning for few-shot specific emitter identification using CNN-transformer with virtual adversarial training. Appl Intell. 2025;55:799. doi:10.1007/s10489-025-06645-5. [Google Scholar] [CrossRef]
18. Hu N, Li X, Zhao Z, Fang M, Liu J. Multi-binary network with feature disentanglement for open-set specific emitter identification. Signal Process. 2026;240:110344. doi:10.1016/j.sigpro.2025.110344. [Google Scholar] [CrossRef]
19. Xu L, Fu X, Wang Y, Zhang Q, Zhao H, Lin Y, et al. Enhanced few-shot specific emitter identification via phase shift prediction and decoupling. IEEE Trans Cogn Commun Netw. 2025;11(1):145–55. doi:10.1109/TCCN.2024.3435886. [Google Scholar] [CrossRef]
20. Sankhe K, Belgiovine M, Zhou F, Riyaz S, Ioannidis S, Chowdhury K. ORACLE: optimized radio classification through convolutional neural networks. IEEE Infocom. 2019;2019:370–8. doi:10.1109/INFOCOM.2019.8737463. [Google Scholar] [CrossRef]
21. Zhang X. ADS-B real-world data set (DF = 17). Sci Data Bank. 2023. doi:10.57760/sciencedb.o00009.00481. [Google Scholar] [CrossRef]
Cite This Article
Copyright © 2026 The Author(s). Published by Tech Science Press.This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


Submit a Paper
Propose a Special lssue
View Full Text
Download PDF
Downloads
Citation Tools