Open Access
ARTICLE
FNRE: A Novel Approach to Heterogeneous Label Noise Rates Estimation in Federated Learning
1 School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
2 School of Cyber Science and Engineering, Huazhong University of Science and Technology, Wuhan, China
3 School of Software Engineering, Huazhong University of Science and Technology, Wuhan, China
* Corresponding Author: Ling Yuan. Email:
Computers, Materials & Continua 2026, 88(1), 48 https://doi.org/10.32604/cmc.2026.075102
Received 24 October 2025; Accepted 03 March 2026; Issue published 08 May 2026
Abstract
Federated learning (FL) enables collaborative model training across decentralized clients without sharing raw data, thereby preserving privacy. However, in real-world FL deployments—such as sensor-based activity recognition, wearable health monitoring, and industrial Internet of Things, where local training data often suffer from heterogeneous noisy labels due to diverse collection environments, sensor limitations, and labeling errors. These noisy labels, typically distributed unevenly across clients due to differences in client-side annotation, exacerbate Non-Independent and Identically Distributed (non-IID) data issues, leading to biased updates, unstable convergence, and degraded global model performance. Accurate estimation of client-specific noise rates is therefore crucial for adaptive algorithm selection, personalized parameter tuning, noise-aware aggregation, and resource allocation in FL. Existing noise rate estimation methods, primarily developed for centralized settings, require client-specific clean validation sets or prior knowledge of noise, making them impractical in privacy-sensitive federated settings. In this work, we propose a federated noise rate estimation (FNRE) method that eliminates the need for per-client clean datasets or prior knowledge of noise. Our approach requires only a minimal assumption—at least one client with a small clean validation set—and leverages the global model’s predictions to estimate local noise rates across all clients. Specifically, the method computes global prediction accuracy using data from the small, clean subset of clients, broadcasts this accuracy to all participants, and enables each client to infer its noise rate using its own annotated labels and the predicted label sequence. We further provide a theoretical analysis with provable error bounds. Extensive experiments on image classification (CIFAR-10, CIFAR-100) and sensor-based activity recognition (Widar, WISDM-W) under various synthetic and real-world noisy label settings demonstrate that our method achieves a noise rate estimation Mean Absolute Error (MAE) of only 0.82%–2.19%, outperforming state-of-the-art baselines by 29.8%–49.9% on average while maintaining practicality in privacy-sensitive federated environments.Keywords
Federated learning (FL) is a distributed machine learning paradigm that enables collaborative model training across multiple devices or organizations without centralizing data [1–3]. This approach inherently protects data privacy by keeping local data on client devices, while only sharing model updates. In many real-world applications, especially those involving large-scale sensor networks—such as sensor-based activity recognition, wearable health monitoring systems, and industrial Internet of Things infrastructures—FL provides an effective framework for processing the massive volume of distributed sensing data while complying with privacy constraints [4–6]. However, in these practical scenarios, the quality of client-side local data is often difficult to guarantee due to environmental diversity during data collection, sensor limitations or malfunctions, and human or algorithmic errors in the labeling process [7,8]. A particularly prevalent issue among these is ‘noisy labels,’ where some training samples are assigned incorrect class labels. Studies indicate that real-world datasets can contain label noise rates ranging from 8.0% to 38.5% [9,10], which severely degrade model performance [11].
Moreover, in the federated learning paradigm, this noisy label is often heterogeneously distributed across client devices, leading to varied noise rates and mislabeling types [12]. This heterogeneity in noisy labels exacerbates the already existing Non-Independent and Identically Distributed (non-IID) nature of federated data, leading to amplified divergence in local model updates, biased gradient aggregation, and unstable convergence in the global model training process [13,14]. A one-size-fits-all noise-robust learning strategy often fails to effectively address the diverse noise characteristics across clients. Therefore, accurately estimating label noise rates at each client becomes crucial, as it enables data quality diagnosis, supports the design of client-specific noise-robust learning strategies [12,15,16], personalized hyperparameter tuning [17], noise-aware aggregation mechanisms [18], and resource allocation optimization [19]—ultimately enhancing both robustness and generalization in federated settings.
Most existing studies on noise rate estimation have been developed for centralized learning settings, leveraging approaches such as noise transition matrices [29–31], cross-validation [23,32], or mixture proportion estimation (MPE) [20,25,28]. As summarized in Table 1, these centralized methods lack the inherent capability for distributed deployment. Consequently, directly transposing them to a federated environment would necessitate treating each client as an isolated estimation task. However, this imposes a strict constraint: every client must possess either prior knowledge of the specific noise type or a local clean auxiliary dataset. In practical federated scenarios involving hundreds of clients with heterogeneous noise, assuming universal access to such prior knowledge or auxiliary data is unrealistic. Therefore, existing centralized noise rate estimation methods are fundamentally incompatible with the constraints of federated learning.
To address the above challenges, we propose a federated noise rate estimation (FNRE) algorithm that requires only a minimal condition: at least one client possessing a small, clean validation dataset. Specifically, our method first leverages the global model to generate predicted labels for all samples on each client, forming a local sequence of predicted labels, called the auxiliary label sequence. Then, the global model’s prediction accuracy is computed using clean validation data from the small subset of participating clients, and the resulting accuracy values are uploaded to the server. Finally, the server broadcasts the global model’s prediction accuracy to all clients, enabling each client to infer its local noise rate by combining the auxiliary label sequence, the global prediction accuracy, and its own annotated labels.
In summary, the key contributions of this paper are:
• Federated Noise Rate Estimation: We propose a federated noise rate estimation algorithm that achieves privacy-preserving distributed noise rate estimation by transmitting only non-sensitive local accuracy statistics, thereby significantly relaxing constraints on auxiliary data. Specifically, our method enables network-wide estimation provided that at least one client possesses a minimal clean validation set (e.g., as few as 300 samples). This capability substantially reduces barriers to practical deployment in real-world applications.
• Privacy-Preserving Estimation with Error Bounds: To the best of our knowledge, the proposed Federated Noise Rate Estimation (FNRE) algorithm is the first framework specifically tailored for estimating heterogeneous label noise rates in federated settings. It inherently supports distributed deployment while strictly preserving privacy. Furthermore, we provide a rigorous theoretical analysis, establishing convergence error bounds to mathematically guarantee the reliability of our estimator.
• Experimental Validation: Extensive experiments on computer vision tasks (CIFAR-10, CIFAR-100) and sensor-based activity recognition (Widar, WISDM-W) demonstrate superior estimation accuracy compared to baselines.
In this section, we primarily focus on introducing noise estimation methods.
Noise transition matrix. The core idea is to model the probability
Cross-validation. Cross-validation-based methods estimate noisy label rates by partitioning the dataset, training models on one subset, and evaluating their performance on the other. For example, INCV [23] proposed randomly splitting a noisy training set into two disjoint subsets, assuming that their noise transition matrices are identical, and then inferring the noise rate from the test accuracy obtained via cross-validation. While conceptually simple, this approach requires prior knowledge of the type of label noise (e.g., symmetric or asymmetric). Extending this approach to federated learning presents significant challenges, as the estimation process must be executed independently on each client, requiring prior knowledge of the specific type of noisy labels at each client.
Mixture proportion estimation. Mixture proportion estimation (MPE) methods estimate the fraction of examples belonging to a target distribution within a mixture distribution, and have been adapted to estimate noisy label rates. For instance, KM [22] embeds distributions into a reproducing kernel Hilbert space (RKHS) and uses convex optimization to achieve statistically consistent mixture proportion estimates in a non-parametric setting. IR [24] estimates the noise rate by evaluating the conditional probability of noisy samples, but the method’s estimation accuracy is limited by model assumptions, extremely small conditional probabilities, and the choice of parameters in density estimation. By strengthening certain distributional assumptions, the ROC [20] introduces a mixture proportion estimation approach that estimates the true mixture proportion, while also providing a practical convergence rate analysis for MPE. Building on this, TiCE [21] employs decision tree induction to identify high-purity positive examples in positive-unlabeled datasets, leveraging lower bound properties of label frequencies. DEDPUL [26] jointly estimates mixture proportions and performs Positive-Unlabeled (PU) classification by modeling density differences and calibrating posterior probabilities.
Consider a federated learning system with K clients, where each client
For each client
where
4.1 Theoretical Framework for Noise Rate Estimation
To estimate the overall noise rates
4.1.1 Auxiliary Labels Sequence
If the true label sequence
We define
This probability
1. Correct Agreement: Both labels are correct and match the true label
2. Coincidental Error: Both labels are incorrect but erroneously coincide.
Thus, the probability
4.1.3 Mathematical Formulation: Deriving Noise Rate from Agreement
Formally,
where
We interpret
Under the uniform noise assumption, each incorrect label is assigned uniformly at random to one of the other
This simplification reduces the original expression for
Based on the preceding analysis, with
This closed-form solution enables efficient estimation of
4.1.4 Rationality of the Uniform Label Noise Assumption
The assumption of uniform label noise is grounded in the Principle of Maximum Entropy [33]. In the absence of specific prior knowledge about the noise structure (i.e., the noise transition matrix), the uniform distribution is the most conservative choice, as it makes the fewest assumptions and introduces the least bias. In a privacy-preserving federated setting, obtaining client-specific noise transition matrices is often infeasible, making the uniform assumption a practical and principled starting point.
Empirically, as detailed in Section 5.7, we verify that our method is robust to violations of this assumption. Even under non-uniform label noise settings (e.g., asymmetric or instance-dependent noise), adopting the uniform label noise assumption yields virtually no degradation in estimation accuracy. This empirical evidence strongly supports adopting this assumption for practicality and efficiency, without sacrificing significant accuracy.
4.1.5 Practical Implementation and Validation Data Handling
A neural network is trained to generate the auxiliary label sequence
where
In practical implementation, we assume access to a small set
Furthermore, in a federated learning setting, it is sufficient for only a single client to possess a small amount of clean validation data, rather than requiring every client to have such a dataset. From a privacy-preserving perspective, no raw data ever leaves local storage; only the global model’s prediction accuracy on the validation set is communicated, ensuring that sensitive information remains protected while still enabling reliable noise rate estimation.
4.1.6 Error Bound Convergence Analysis
We establish rigorous theoretical foundations for our noise rate estimation methodology by deriving precise error bounds and statistical guarantees.
The error in estimating
The partial derivatives are computed as:
Thus, the error bound is:
The estimation errors satisfy:
where
The error bound converges at a rate of
4.2 Federated Noise Rate Estimation Framework
In the federated learning paradigm, we establish a privacy-preserving framework for noise rate estimation that operates under stringent data locality constraints. The fundamental challenge lies in accurately estimating label noise rates across heterogeneous clients while maintaining data confidentiality. Fig. 1 shows an overview of the Federated Noise Rate Estimation Framework.

Figure 1: An overview of federated noise rate estimation framework.
The key Assumptions underlying our framework are detailed as follows:
• Auxiliary Data Scarcity:
• Distributed Auxiliary Access: Only a subset
Our federated noise estimation protocol (Algorithm 1) operates through four sequential phases, each designed to preserve privacy while enabling accurate quantification of the noise rate.
Phase I: Model Initialization and Local Training
The server initializes a global model
Phase II: Privacy-Preserving Accuracy Aggregation
Calibration clients compute local accuracy statistics on their auxiliary data subsets without exposing raw data:
For each calibration client
where
Phase III: Global Auxiliary Noise Rate Computation
The server aggregates received accuracy statistics to derive the global auxiliary noise rate:
This quantity represents the model’s error rate on clean auxiliary data and serves as a calibration reference for local noise estimation. The server broadcasts
Phase IV: Local Noise Rate Inference
Each client
• Auxiliary Label Generation: Client
• Agreement Rate Computation: The agreement rate between model predictions and observed labels is calculated as:
• Local Noise Rate Estimation: The local noise rate is computed as:
where C denotes the number of classes in the classification task.

4.2.3 Communication Efficiency and Algorithmic Complexity
Our framework achieves exceptional privacy preservation and communication efficiency through a streamlined two-phase protocol. Privacy is guaranteed via strict data locality (raw datasets remain exclusively on client devices), and statistical aggregation (clients transmit only aggregated counts
4.2.4 Extension to Class-Specific Noise Rate Estimation
The proposed federated noise rate estimation framework can be readily extended to estimate class-specific noise rates
Specifically, in Phase II (Privacy-Preserving Accuracy Aggregation), calibration clients would compute and transmit accuracy statistics (
This section details the datasets utilized to evaluate the proposed methodologies. The selected datasets encompass computer vision tasks with natural images (CIFAR-10 [34], CIFAR-100 [34], CIFAR-10N [35], CIFAR-100N [35], ILSVRC 2012 [36]) and human activity recognition tasks using sensor data (Widar [37], WISDM-W [38]), providing a diverse evaluation framework that demonstrates the generalizability of our approach across different modalities and problem domains.
• CIFAR-10 and CIFAR-100 [34] These are standard benchmark datasets for image classification. CIFAR-10 comprises 10 distinct classes, while CIFAR-100 extends this with 100 finer-grained categories. Both consist of
• CIFAR-10N and CIFAR-100N [35] These are noisy variants of CIFAR-10 and CIFAR-100, specifically designed to emulate real-world label noise challenges. Their labels contain various types of human-induced errors, such as aggregation errors and spurious correlations, which commonly arise from crowdsourcing or imperfect annotation processes. This characteristic makes them invaluable for research focused on robust learning in the presence of realistic label noise.
• ILSVRC 2012 [36] The ILSVRC 2012 dataset, a subset of the ImageNet Large Scale Visual Recognition Challenge, consists of approximately 1.2 million training images, 50,000 validation images, and 100,000 test images, spanning 1000 object categories. Each image is labeled with a single class from a diverse set of object types, covering animals, vehicles, everyday objects, and scenes.
• Widar [37] The Widar dataset is designed for contactless gesture recognition based on Wi-Fi signals. The system collects fine-grained Channel State Information (CSI) using an Intel 5300 network interface card in a 3
• WISDM-W [38] The WISDM (Wireless Sensor Data Mining) dataset is a widely used benchmark for human activity recognition (HAR) using accelerometer and gyroscope data collected from smartphones and smartwatches. A total of 51 participants performed 18 daily activities (such as walking, jogging, sitting, standing, climbing stairs, opening doors, eating) for approximately 3 min each. To enhance its utility for machine learning tasks, similar activities were combined (e.g., various eating activities merged into “eating”), and uncommon or problematic activities were removed. Due to differences in data collection methodologies (e.g., smartwatch and smartphone data not always collected simultaneously or precisely synchronized), WISDM-W is treated as an independent dataset from its smartphone counterpart (WISDM-P).
Table 2 provides a comprehensive overview of these datasets, including their key characteristics and statistics.

5.2 Federated Learning Settings
This section details the specific configurations adopted for the federated learning environment, encompassing data distribution, noise injection mechanisms, validation data setup, and other key training parameters.
5.2.1 Data Distribution Configuration
We adopt a non-independent and identically distributed (non-IID) data partitioning strategy following the methodology established by [4]. Specifically, we employ the Dirichlet-based partitioning scheme to create realistic statistical heterogeneity across participating clients.
For classification tasks on datasets with C classes, we generate client-specific data distributions by sampling from a Dirichlet distribution
5.2.2 Label Noise Configuration
We implement a comprehensive noise simulation strategy that encompasses three distinct noise paradigms across participating clients. Specifically, symmetric noise [39] is applied to one-third of clients, with label corruption occurring uniformly across all classes with equal probability. Asymmetric noise [39] is applied to another one-third of clients, featuring class-dependent noise patterns that reflect realistic annotation biases commonly observed in practical scenarios. The remaining one-third of clients are subject to instance-dependent noise [40], where the noise probability varies with sample characteristics and proximity to decision boundaries, thereby simulating the inherent uncertainty in borderline cases.
To simulate realistic noise rate distributions across the federated system, we employ three distinct probabilistic sampling strategies. The first strategy uses a Gaussian distribution
Each client randomly selects one noise rate from these distributions and is subsequently assigned a noise type through random allocation. This dual randomization process ensures diverse noise characteristics across the federated system, creating a heterogeneous noise environment that closely mirrors real-world federated learning deployments. This systematic configuration enables comprehensive evaluation of algorithm robustness under varying noise intensities and patterns, providing a rigorous assessment of the proposed methodology’s effectiveness across diverse, challenging scenarios.
5.2.3 Validation Data Configuration
To maintain realistic evaluation conditions while ensuring fair performance assessment, we implement a limited clean validation setup. Specifically, only five randomly selected clients have access to clean validation data, with a total validation set size of 300 samples distributed across them. This configuration reflects practical federated scenarios where clean, high-quality data is scarce and unevenly distributed among participants.
5.2.4 System and Communication Parameters
The federated learning system parameters are configured according to dataset-specific requirements to ensure optimal convergence and fair comparison across different experimental conditions. Table 3 presents the detailed parameter configurations for each dataset.

We select MPEIA [25], INCV [23], KM [22], ROC [20], TiCE [21], IR [24], DEDPUL [26],
5.4 Comparison of Baseline Methods for Noise Rate Estimation Error
We conduct comprehensive comparative experiments to evaluate the effectiveness of our proposed federated noise rate estimation method against nine state-of-the-art baseline approaches across four benchmark datasets encompassing diverse data modalities: computer vision tasks (CIFAR-10, CIFAR-100) and sensor-based activity recognition (Widar, WISDM-W), as shown in Tables 4 and 5. The experimental framework evaluates performance under three distinct noise rate distribution scenarios: Gaussian distributions with low variance


The experimental results demonstrate substantial performance advantages for our proposed methodology across all evaluated configurations, as shown in Tables 4 and 5. On computer vision datasets (CIFAR-10 and CIFAR-100), our approach achieves MAE values ranging from 0.82% to 1.65%, representing significant improvements over the best-performing baseline method SuDPL, which records MAE values between 1.30% and 2.34%. The performance differential becomes particularly pronounced under high-noise scenarios, where our method maintains consistent accuracy while baseline methods exhibit substantial degradation. On sensor-based datasets (Widar and WISDM-W), our superiority becomes even more evident, achieving exceptional MAE values of 1.18%–2.19%, substantially outperforming all baseline approaches. Traditional methods such as ROC demonstrate particularly poor adaptation to sensor data characteristics, with MAE values exceeding 17% in multiple configurations. Beyond accuracy improvements, our method exhibits superior stability, as evidenced by consistently lower standard deviations ranging from 0.48 to 1.25 across all experimental configurations, substantially lower than those of most baseline approaches. This enhanced stability is particularly critical in federated environments, where client heterogeneity can introduce performance variance. The consistent performance advantages across heterogeneous datasets—from low-resolution natural images to high-dimensional sensor signals—demonstrate the universal applicability and cross-domain generalization capability of our approach.
5.5 Noise Rate Estimation under Federated Configurations
The comprehensive experimental evaluation provides substantive evidence on the performance characteristics of the proposed noise rate estimation algorithm across diverse federated learning paradigms, as shown in Fig. 2. The systematic assessment comprises six distinct experimental configurations, designed to examine algorithmic scalability and estimation precision across varying client participation and categorical scales. Due to computational resource constraints, we were unable to fully train complete training on the large-scale classification dataset ILSVRC 2012. Therefore, we used the publicly available pre-trained ResNet50 model as the trained model to evaluate the noise rate. We employed non-independent and identically distributed (non-IID) data partitioning with a Dirichlet parameter of 0.4 to allocate samples from the ILSVRC 2012 dataset across all participating clients. The experimental configurations for CIFAR-10, CIFAR-10 subset, and CIFAR-100 datasets maintain consistency with the parameters specified in Table 2. In subfigures (a), (b), and (c), client noise rates follow a normal distribution with mean 0.5 and variance 0.2. In subfigures (d), (e), and (f), categorical noise rates follow a uniform distribution ranging from 0.2 to 0.6.

Figure 2: Performance evaluation of the proposed noise rate estimation algorithm across different federated learning scenarios. The top row demonstrates client-wise noise rate estimation performance under varying numbers of participating clients, while the bottom row illustrates class-wise estimation accuracy across different numbers of classes. All experiments show a strong correlation between true and estimated noise rates with consistently low estimation errors.
The empirical findings demonstrate notable consistency in estimation accuracy across different federated learning architectures. Specifically, estimation precision remains remarkably stable regardless of participant scale, with Mean Absolute Error (MAE) values ranging from 0.010 to 0.014 across configurations spanning 10 to 1000 clients. The observed correlation coefficients, consistently exceeding 0.993, suggest strong linear relationships between true and estimated noise rates. This pattern persists across heterogeneous dataset characteristics, from small-scale CIFAR-10 deployments to large-scale classification scenarios comprising 1000 clients and 1.2 million samples. The experimental progression from 5-class to 1000-class scenarios indicates maintained estimation accuracy, with MAE values demonstrating controlled variance (0.010–0.020) despite significant increases in the number of categories. Collectively, this experimental evidence robustly indicates that our federated noise rate estimation method is applicable not only to small and medium-sized federated classification tasks but also to large-scale scenarios, thereby demonstrating substantial practical value.
This section provides a detailed analysis of the experimental results, elucidating the performance characteristics of the proposed federated noise rate evaluation algorithm under various conditions, as shown in Fig. 3. The Mean Absolute Error (MAE) serves as the primary metric for quantifying the algorithm’s accuracy in noise rate estimation.

Figure 3: Comprehensive experimental evaluation of the proposed federated noise rate estimation algorithm across multiple parameter configurations. The experiments demonstrate the algorithm’s performance sensitivity to various factors including the error level of evaluating
Fig. 3a investigates the impact of estimation error in the noise rate of auxiliary label sequence
5.7 Validating the Non-Uniform Label Noise for Simplifying
In this section, we experimentally validate that simplifying the computation of
The experimental results are shown in Tables 6 and 7. Across all datasets (CIFAR-10, CIFAR-100, Widar, WISDM-W), noise types (asymmetric and sample-dependent), and noise rate distributions, FNRE* and FNRE demonstrate consistently similar MAE and standard deviation values with minimal differences. This indicates that introducing the uniform label noise assumption for simplified computation of


Additionally, we generated 500 synthetic classification datasets using sklearn’s make_classification function with random parameters. The number of samples follows a uniform distribution from 2000 to 1,000,000, the number of classes from 2 to 1000, feature dimensions from 4 to 256, and class separation from 1 to 5. We computed the absolute error between FNRE* and FNRE estimates, with results shown in Fig. 4. The absolute errors remain within a very small range.

Figure 4: Distribution of absolute error between FNRE* and FNRE on 500 synthetic classification datasets.
Overall, our adoption of the uniform label noise assumption to simplify the computation of
We presented Federated Noise Rate Estimation (FNRE), a practical and privacy-preserving method for estimating client-specific label noise rates in federated learning. Unlike conventional centralized approaches, FNRE requires only a small clean validation set from a few clients and avoids per-client clean data or prior noise information. Our approach offers theoretical error guarantees and achieves superior accuracy across diverse datasets and noise conditions. These results demonstrate FNRE’s effectiveness in enabling adaptive.
Limitations and Future Work
The proposed Federated Noise Rate Estimation (FNRE) method significantly improves label noise rate estimation in federated learning, but there are several limitations to address in future work:
Dependency on Clean Validation Data:Our approach relies on the prerequisite that at least one client possesses a small, clean validation set to guide the noise estimation.
Assumptions behind the Noise Model:We employ a uniform label noise assumption to simplify the configuration of the parameter
Practical Considerations:From a practical deployment perspective, in cases where such complex noise-error correlations exist, the parameter
In future work, we plan to reference anchor-based noise matrix estimation methods to construct a non-uniform correction term for the calculation of
Acknowledgement: Not applicable.
Funding Statement: This research was funded by the National Natural Science Foundation of China under Grant Numbers 62272180 (for Ling Yuan) and 62272176 (for Guohui Li).
Author Contributions: Qian Rong was responsible for the conception and design of the study and for drafting the paper. Lu Zhang contributed to the drafting of the paper. Ling Yuan was responsible for revising the manuscript critically for intellectual content and for the final approval of the version to be published. Zhong Yang contributed to the drafting of the paper. Guohui Li was involved in revising the manuscript critically for intellectual content and in the final approval of the version to be published. All authors reviewed and approved the final version of the manuscript.
Availability of Data and Materials: All datasets supporting the results or analyses of this study are publicly available in the following repositories:
• CIFAR-10 and CIFAR-100 datasets: Available at https://www.cs.toronto.edu/~kriz/cifar.html (No DOI or Accession Number).
• CIFAR-10N and CIFAR-100N datasets: Available at http://competition.noisylabels.com/ (No DOI or Accession Number).
• ILSVRC 2012 datasets: Available at https://image-net.org/challenges/LSVRC/2012/index.php (No DOI or Accession Number).
• Widar datasets: Available at https://tns.thss.tsinghua.edu.cn/widar3.0/index.html with DOI: https://doi.org/10.21227/7znf-qp86.
• WISDM datasets: Available at https://www.cis.fordham.edu/wisdm/dataset.php (No DOI or Accession Number).
Ethics Approval:: Not applicable.
Conflicts of Interest:: The authors declare no conflicts of interest.
References
1. Kairouz P, McMahan HB, Avent B, Bellet A, Bennis M, Bhagoji AN, et al. Advances and open problems in federated learning. Found Trends Mach Learn. 2021;14(1–2):1–210. [Google Scholar]
2. Li T, Sahu AK, Talwalkar A, Smith V. Federated learning: challenges, methods, and future directions. IEEE Signal Process Magaz. 2020;37(3):50–60. [Google Scholar]
3. Li G, Cai J, Lu J, Chen H. Incentive mechanism design for cross-device federated learning: a reinforcement auction approach. IEEE Trans Mobile Comput. 2025;24(4):3059–75. doi:10.1109/tmc.2024.3508260. [Google Scholar] [CrossRef]
4. Alam S, Zhang T, Feng T, Shen H, Cao Z, Zhao D, et al. FedAIoT: a federated learning benchmark for artificial intelligence of things. arXiv:2310.00109. 2023. [Google Scholar]
5. Nguyen DC, Ding M, Pathirana PN, Seneviratne A, Li J, Poor HV. Federated learning for internet of things: a comprehensive survey. IEEE Commun Surv Tut. 2021;23(3):1622–58. doi:10.1109/comst.2021.3075439. [Google Scholar] [CrossRef]
6. Li G, Cai J, He C, Zhang X, Chen H. Online incentive mechanism designs for asynchronous federated learning in edge computing. IEEE Internet Things J. 2024;11(5):7787–804. doi:10.1109/jiot.2023.3316470. [Google Scholar] [CrossRef]
7. Wang H, Jiang T, Guo Y, Guo F, Bie R, Jia X. Label noise correction for federated learning: a secure, efficient and reliable realization. In: 2024 IEEE 40th International Conference on Data Engineering (ICDE). Piscataway, NJ, USA: IEEE; 2024. p. 3600–12. [Google Scholar]
8. Liang S, Huang J, Hong J, Zeng D, Zhou J, Xu Z. FedNoisy: federated noisy label learning benchmark. arXiv:2306.11650. 2023. [Google Scholar]
9. Xiao T, Xia T, Yang Y, Huang C, Wang X. Learning from massive noisy labeled data for image classification. In: Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition. Piscataway, NJ, USA: IEEE; 2015. p. 2691–9. [Google Scholar]
10. Li W, Wang L, Li W, Agustsson E, Van Gool L. Webvision database: visual learning and understanding from web data. arXiv:1708.02862. 2017. [Google Scholar]
11. Giap T-T, Kieu T-D, Le T-L, Tran T-H. FedDC: label noise correction with dynamic clients for federated learning. IEEE Internet Things J. 2025;12(8):10266–77. [Google Scholar]
12. Ji X, Zhu Z, Xi W, Gadyatskaya O, Song Z, Cai Y, et al. FedFixer: mitigating heterogeneous label noise in federated learning. In: Proceedings of the AAAI Conference on Artificial Intelligence. Palo Alto, CA, USA: AAAI Press; 2024. Vol. 38, p. 12830–8. [Google Scholar]
13. Ke S, Huang C, Liu X. Quantifying the impact of label noise on federated learning. arXiv:2211.07816. 2022. [Google Scholar]
14. Wu N, Yu L, Jiang X, Cheng K-T, Yan Z. FedNoRo: towards noise-robust federated learning by addressing class imbalance and label noise heterogeneity. arXiv:2305.05230. 2023. [Google Scholar]
15. Zeng B, Yang X, Chen Y, Shen Z, Yu H, Zhang Y. FedES: federated early-stopping for hindering memorizing heterogeneous label noise. In: Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence; 2024 Aug 3–9; Jeju, Republic of Korea. p. 5416–24. [Google Scholar]
16. Chen L, Ang F, Chen Y, Wang W. Robust federated learning with noisy labeled data through loss function correction. IEEE Trans Netw Sci Eng. 2022;10(3):1501–11. doi:10.1109/tnse.2022.3227287. [Google Scholar] [CrossRef]
17. Cheng A, Wang Z, Li Y, Cheng J. HPN: Personalized federated hyperparameter optimization. arXiv:2304.05195. 2023. [Google Scholar]
18. Jiang X, Wen T, Yang Z, Wu L, Chen Y, Sun S, et al. Robust federated learning against noisy clients via masked optimization. arXiv:2506.02079. 2025. [Google Scholar]
19. Lari E, Arablouei R, Gogineni VC, Werner S. Noise-robust and resource-efficient ADMM-based federated learning for WLS regression. Signal Process. 2025;241(10):110387. doi:10.2139/ssrn.5256354. [Google Scholar] [CrossRef]
20. Scott C. A rate of convergence for mixture proportion estimation, with application to learning from noisy labels. In: Artificial Intelligence and Statistics. London, UK: PMLR; 2015. p. 838–46. doi:10.3233/978-1-61499-672-9-1618. [Google Scholar] [CrossRef]
21. Bekker J, Davis J. Estimating the class prior in positive and unlabeled data through decision tree induction. In: AAAI’18/IAAI’18/EAAI’18: Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence. Palo Alto, CA, USA: AAAI Press; 2018. Vol. 32, p. 2712–9. [Google Scholar]
22. Ramaswamy H, Scott C, Tewari A. Mixture proportion estimation via kernel embeddings of distributions. In: ICML’16: Proceedings of the 33rd International Conference on Machine Learning. London, UK: PMLR; 2016. p. 2052–60. [Google Scholar]
23. Chen P, Liao BB, Chen G, Zhang S. Understanding and utilizing deep neural networks trained with noisy labels. In: Proceedings of the 36 th International Conference on Machine Learning. London, UK: PMLR; 2019. p. 1062–70. [Google Scholar]
24. Liu T, Tao D. Classification with noisy labels by importance reweighting. IEEE Trans Pattern Anal Mach Intell. 2016;38(3):447–61. doi:10.1109/tpami.2015.2456899. [Google Scholar] [PubMed] [CrossRef]
25. Yu X, Liu T, Gong M, Batmanghelich K, Tao D. An efficient and provable approach for mixture proportion estimation using linear independence assumption. In: Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition. Piscataway, NJ, USA: IEEE; 2018. p. 4480–9. [Google Scholar]
26. Ivanov D. Dedpul: difference-of-estimated-densities-based positive-unlabeled learning. In: 2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA). Piscataway, NJ, USA: IEEE; 2020. p. 782–90. [Google Scholar]
27. Garg S, Wu Y, Smola AJ, Balakrishnan S, Lipton Z. Mixture proportion estimation and PU learning: a modern approach. Adv Neural Inf Process Systs. 2021;34:8532–44. [Google Scholar]
28. Zhu Y, Fjeldsted A, Holland D, Landon G, Lintereur A, Scott C. Mixture proportion estimation beyond irreducibility. In: ICML’23: Proceedings of the 40th International Conference on Machine Learning. London, UK: PMLR; 2023. p. 42962–82. [Google Scholar]
29. Xia X, Liu T, Wang N, Han B, Gong C, Niu G, et al. Are anchor points really indispensable in label-noise learning? In: Advances in neural information processing systems. Red Hook, NY, USA: Curran Associates, Inc.; 2019. [Google Scholar]
30. Yao Y, Liu T, Han B, Gong M, Deng J, Niu G, et al. Dual t: reducing estimation error for transition matrix in label-noise learning. Adv Neural Infn Process Syst. 2020;33:7260–71. [Google Scholar]
31. Li X, Liu T, Han B, Niu G, Sugiyama M. Provably end-to-end label-noise learning without anchor points. In: Proceedings of the 38 th International Conference on Machine Learning. London, UK: PMLR; 2021. p. 6403–13. [Google Scholar]
32. Menon A, Van Rooyen B, Ong CS, Williamson B. Learning from corrupted binary labels via class-probability estimation. In: Proceedings of the 32nd International Conference on Machine Learning. London, UK: PMLR; 2015. p. 125–34. [Google Scholar]
33. Jaynes ET. Probability theory: the logic of science. Cambridge, UK: Cambridge University Press; 2003. [Google Scholar]
34. Krizhevsky A, Hinton G. Learning multiple layers of features from tiny images [master’s thesis]. Toronto, ON, Canada: University of Toronto; 2009. [Google Scholar]
35. Wei J, Zhu Z, Cheng H, Liu T, Niu G, Liu Y. Learning with noisy labels revisited: a study using real-world human annotations. arXiv:2110.12088. 2021. [Google Scholar]
36. Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, et al. Imagenet large scale visual recognition challenge. Intl J Comput Vision. 2015;115(3):211–52. doi:10.1007/s11263-015-0816-y. [Google Scholar] [CrossRef]
37. Zheng Y, Zhang Y, Qian K, Zhang G, Liu Y, Wu C, et al. Zero-effort cross-domain gesture recognition with Wi-Fi. In: Proceedings of the 17th Annual International Conference on Mobile Systems, Applications, and Services. New York, NY, USA: ACM; 2019. p. 313–25. [Google Scholar]
38. Kwapisz JR, Weiss GM, Moore SA. Activity recognition using cell phone accelerometers. ACM SigKDD Explor Newsletter. 2011;12(2):74–82. [Google Scholar]
39. Song H, Kim M, Park D, Shin Y, Lee JG. Learning from noisy labels with deep neural networks: a survey. IEEE Trans Neural Netw Learn Syst. 2022;34(11):8135–53. doi:10.1109/tnnls.2022.3152527. [Google Scholar] [PubMed] [CrossRef]
40. Cheng H, Zhu Z, Li X, Gong Y, Sun X, Liu Y. Learning with instance-dependent label noise: a sample sieve approach. arXiv:2010.02347. 2020. [Google Scholar]
Cite This Article
Copyright © 2026 The Author(s). Published by Tech Science Press.This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


Submit a Paper
Propose a Special lssue
View Full Text
Download PDF
Downloads
Citation Tools