Open Access
ARTICLE
Robust Deep One-Class Classification Time Series Anomaly Detection
1 State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang, 550000, China
2 College of Computer Science and Technology, Weifang University of Science and Technology, Weifang, 261000, China
3 School of Mathematics and Big Data, Guizhou Education University, Guiyang, 550018, China
* Corresponding Author: Yuling Chen. Email:
(This article belongs to the Special Issue: Artificial Intelligence Algorithms and Applications)
Computers, Materials & Continua 2025, 83(3), 5181-5197. https://doi.org/10.32604/cmc.2025.060564
Received 04 November 2024; Accepted 04 March 2025; Issue published 19 May 2025
Abstract
Anomaly detection (AD) in time series data is widely applied across various industries for monitoring and security applications, emerging as a key research focus within the field of deep learning. While many methods based on different normality assumptions perform well in specific scenarios, they often neglected the overall normality issue. Some feature extraction methods incorporate pre-training processes but they may not be suitable for time series anomaly detection, leading to decreased performance. Additionally, real-world time series samples are rarely free from noise, making them susceptible to outliers, which further impacts detection accuracy. To address these challenges, we propose a novel anomaly detection method called Robust One-Class Classification Detection (ROC). This approach utilizes an autoencoder (AE) to learn features while constraining the context vectors from the AE within a sufficiently small hypersphere, akin to One-Class Classification (OC) methods. By simultaneously optimizing two hypothetical objective functions, ROC captures various aspects of normality. We categorize the input raw time series into clean and outlier sequences, reducing the impact of outliers on compressed feature representation. Experimental results on public datasets indicate that our approach outperforms existing baseline methods and substantially improves model robustness.Keywords
Analyzing time series allows us to understand the underlying processes that generate these sequences, thereby enhancing our comprehension of these processes. Anomaly detection in time series is a critical challenge in data mining, with applications across diverse fields such as transportation and manufacturing, where it serves to monitor system behavior. For instance, in aircraft engine fault detection, monitoring the time series of engine RPM, fuel pressure, temperature, and vibration signals enables the identification of abnormal fluctuations or patterns, allowing for early detection of mechanical failures and preventing accidents during flight. In railway transportation systems, monitoring the status of trains involves real-time analysis of data such as wheel temperature, axle vibration, and electrical signals. By detecting anomalies, potential hazards that could lead to derailments or failures can be identified and mitigated promptly. In manufacturing production lines, equipment health management is facilitated through the monitoring of vibration, temperature, and current of machinery via time series data, enabling the identification of potential failures and timely maintenance, thus minimizing downtime and ensuring production efficiency.
Time Series Anomaly Detection (TSAD) is also essential in applications like health monitoring and fraud detection, where it involves identifying unique time series instances that deviate from typical patterns [1]. For example, in health monitoring, detecting abnormal heart rates and brain waves in electrocardiogram (ECG) and electroencephalogram (EEG) data can reveal health issues such as arrhythmias and epilepsy. In credit card fraud detection, analyzing user transaction patterns and identifying unusual behaviors, such as large purchases in a short time or transactions at atypical locations, helps in detecting and preventing fraudulent activities. Additionally, in network traffic anomaly detection, analyzing time series data of network traffic can uncover abnormal data transmission behaviors, such as DDoS attacks, data breaches, or malicious software activities.
In recent years, deep learning-based approaches have achieved impressive results in time series anomaly detection, particularly with complex datasets. These methods excel at modeling long-term and nonlinear temporal patterns within the data, surpassing traditional methods such as similarity search and density-based clustering [2–5]. Neural network-based approaches commonly use an encoder to compress the input time series into a compact latent representation, which is then decoded to reconstruct the original series. This encoder-decoder framework, commonly referred to as an autoencoder [6], compresses the initial input through a bottleneck layer, promoting the formation of compact latent representations. This structure enables the model to capture essential patterns within the time series while filtering out irrelevant or atypical patterns, such as anomalies [7]. This approach aids in the detection of anomalies by evaluating the reconstruction error between the original time series and its reconstructed counterpart. A larger reconstruction error suggests an increased probability that the related observations are anomalies. One-class (OC) techniques consolidate normal instances into a single category by minimizing the volume of the hypersphere that encompasses the feature representations. However, despite the strong performance demonstrated by deep learning techniques [8–10] such as autoencoders, they encounter two major challenges.
Single Hypothesis: A single hypothesis often captures only a limited aspect of sample normality, while anomalies can manifest in diverse forms, such as point anomalies, subsequence anomalies, and anomalies across entire time series. Inspired by the principles of ensemble anomaly detection methods, we suggest that detectors relying on a singular hypothesis may be insufficient to capture this diversity, potentially limiting their effectiveness in detecting varied anomaly types [11]. Consequently, while these methods may excel in detecting specific types of anomalies, their effectiveness often diminishes when encountering others.
An alternative approach divides the process into two phases: pre-training on the overall time series, followed by fine-tuning for anomaly detection (AD). For instance, deep SVDD [12] initially utilizes an autoencoder for feature extraction and subsequently refines these features for anomaly detection through a one-class loss function. Similarly, Reference [13] applies contrastive learning in the first phase and utilize one-class methods for detection in the second phase. Formally, these approaches combine feature extraction methods with the assumptions of one-class learning; however, the objectives of the two phases are often misaligned. While the pre-training phase may yield representations that align with typical patterns, these representations can also be influenced by extraneous features unrelated to anomaly detection. As a result, the performance of such methods may be constrained by these pre-trained features.
Robustness: In unsupervised learning, training data often includes anomalies. Because the encoder compresses all observations in the input time series, even those that are anomalous, the hidden representation may become susceptible to these outliers. Particularly when anomalies are high in amplitude, even a few can compromise the latent information, leading to a risk that the latent representation itself reflects anomalous patterns from the training data.
Deep learning-based soft sensors for time series anomaly detection exhibit significant vulnerabilities to adversarial attacks. Knowledge-guided adversarial perturbations can be designed to subtly manipulate the input data distribution without causing noticeable changes, resulting in a substantial degradation of model performance. This exposes the insufficient robustness of current deep learning methods in time series anomaly detection [14]. Reference [15] proposes an adversarial training strategy using historical gradients and domain adaptation. By leveraging historical information to capture temporal dynamics and mapping input samples to a shared feature space, this approach effectively enhances model robustness against adversarial examples. This is particularly critical for time series anomaly detection, where data distribution changes introduce additional uncertainty. As a result, the model may produce lower reconstruction errors for specific anomalies, making it difficult to distinguish them from normal data, which in turn negatively affects accuracy. As shown in Fig. 1, training set data contaminated by anomalies can lead to the model generating smaller reconstruction errors for similar anomalies during testing, complicating their detection. To mitigate this issue, a robust solution is required to ensure that the latent representation is not influenced by anomalies present in the training data.

Figure 1: A portion of the time series from the training set for anomaly detection, where the orange shading represents the actual anomalous segments and the red solid line represents the anomaly score. In unsupervised anomaly detection, the training set typically relies on benign samples, which results in the model learning characteristics associated with anomalies. Consequently, the model becomes less sensitive to certain anomalies present in the test set
We introduce a novel autoencoder framework to address the issues of robustness and single hypothesis in anomaly detection. In this study, we introduce a single-stage anomaly detection approach termed the Robust One-Class Classification (ROC) autoencoder. We hypothesize that normal samples will be more accurately reconstructed, with their projection vectors in the latent space forming a compact hypersphere.
Rather than reconstructing the input time series T directly, we decompose it into two components: the clean time series
In summary, the main contributions of this paper are as follows:
• The proposed method differs from traditional single-hypothesis autoencoder approaches by integrating a multi-hypothesis anomaly detection framework that combines autoencoders with OC methods, capturing richer feature representations in the time series.
• Unlike standard denoising autoencoders, our method does not require additional noise-free training data. By optimizing the sparsity of
• The effectiveness of the proposed approach has been validated on publicly available time series datasets, where it demonstrates superior performance compared to existing methods.
A time series
For a specified time series
2.2 Deep Learning-Based Anomaly Detection
Traditional anomaly detection methods rely on statistical features, but these approaches are often unsuitable for time series data, leads to the feature information not being correctly represented. With the increase in data dimensionality and volume, deep learning methods have emerged. OC methods can capture complex features that represent “normality.” For instance, methods based on Generative Adversarial Networks (GANs) [16] and autoencoders [17] assume that normal samples can be reconstructed better by the model. Clustering methods, on the other hand, posit that normal samples cluster together as a large group, while outlier data points are classified as anomalies [18]. In contrast, contrastive learning approaches enhance the data by making positive samples closer to each other while causing the negative samples to be far apart from each other [19]. However, these assumptions can often be overly simplistic or effective only for specific types of anomalies. Additionally, another class of deep learning methods, such as Deep SVDD, employs a two-stage OC classifier. This involves feature extraction using a pre-trained encoder model, which is then classified using OC-SVM [12]. Nevertheless, this approach tends to separate the training objective from the downstream task, hindering the effective learning of diverse time series features.
2.3 Robust Principal Component Analysis
Given a matrix M, Principal Component Analysis (PCA) is capable of discovering a low-rank matrix that serves as an approximation of M. However, as PCA generally employs Singular Value Decomposition (SVD) to determine the low-rank matrix, it exhibits similar sensitivity to outliers as that observed in SVD. To enhance the effectiveness of Principal Component Analysis (PCA) in the presence of outliers, Robust Principal Component Analysis (RPCA) [20] has been introduced. The aim of RPCA is to decompose the matrix M into two components: a low-rank matrix L that represents the underlying clean data and a sparse matrix S that captures the outliers. Specifically, RPCA expresses the original matrix M as follows:
In this decomposition, L serves as a low-rank matrix that approximates the clean data within the original matrix M, while S represents a sparse matrix composed of elements identified as outliers, which are not encapsulated by the low-rank matrix L. RPCA accomplishes this decomposition by resolving the optimization problem outlined in the following formula.
In this context,
In unsupervised learning, training data may already contain anomalies. As the encoder compresses the time series, the hidden representation becomes highly sensitive to these anomalies. This means that the anomalous information in the training data can contaminate the latent representation, causing the model to learn abnormal features. As a result, the model may exhibit low reconstruction errors for anomalous samples, making it difficult to distinguish them from clean samples.
To address this issue, we propose a robust method that ensures the latent representation is not affected by anomalies in the training data. Drawing on the principles of RPCA, our proposed neural network is designed to partition the input data during training into two segments: the clean time series

Figure 2: The architecture of the proposed ROC model is depicted, where
In our proposed deep one-class (OC) method, we combine two assumptions: that normal samples can be better reconstructed and that when mapped to a high-dimensional space, they will form a hypersphere with a smaller radius. This approach allows for a more comprehensive learning of the features of clean samples. The motivation behind this is that under a single assumption, the normality features learned by the model may be one-sided, leading to a situation where certain features of abnormal samples differ from those of normal samples, but those features are not recognized by the model. As a result, specific types of anomalies may go undetected by a model based on a single assumption. The specific objective function is as follows:
For a given set of N time series training samples, the objective function consists of two parts: one for learning the features of the time series from the perspective of the autoencoder and the other from the one-class (OC) method. The parameter
The
The data point
where
where
where
In Principal Component Analysis (PCA), a given matrix M can be approximated by identifying a low-rank matrix. To obtain a low-rank representation, PCA applies Singular Value Decomposition (SVD), which makes it inherently sensitive to outliers. To enhance robustness in the presence of outliers, Robust Principal Component Analysis (RPCA) has been proposed. RPCA aims to break down the original matrix M into two parts: a low-rank matrix L that represents the underlying clean structure of M and a sparse matrix S that contains the elements identified as anomalies.
Inspired by the approach of RPCA, we can separate the anomalous parts from the input time series and focus solely on learning the benign features from the samples. In this context, the clean time series
In Eqs. (1) and (9), the loss functions include an
The optimization problems of ROC have constraints and thus cannot be solved by gradient descent based back-propagation (BACKPROP). The optimization task may instead be reformulated into two segments and approached using the Alternating Direction Method of Multipliers (ADMM). ADMM fundamentally works by breaking down the main objective into several sub-objectives, enabling the iterative optimization of each sub-objective while holding the remaining ones constant. Upon optimizing a given sub-objective, the method applies constraints to ensure consistency with the overall objective [22]. Furthermore, the Proximal Algorithm [23] is employed to address elements involving the
As shown in Algorithm 1, when optimizing the ROC, the process is as follows: first, optimize the integrated autoencoder part by minimizing

In terms of datasets, the AIOps dataset is a collection of 29 sub-datasets designed for detecting anomalies in web services based on business cloud KPIs. It includes 29 KPI time series collected from several large technology companies (such as Alibaba, Sogou, Tencent, Baidu, and eBay). These time series are sampled at 1 or 5-min intervals and divided into training and testing portions.
Another dataset we use the UCR Time Series Anomaly Archive, a recently launched repository containing 250 different time series datasets specifically for time series anomaly detection research. Each dataset contains anomalous events of varying lengths, ranging from 1 to 1700. Furthermore, these datasets cover various fields such as health, industry, and biology, exhibiting different types of anomalies with specific characteristics [24].
We also use the multivariate time series dataset SMAP, which comes from a real-world expert-labeled dataset provided by NASA. Each dataset includes a training set and a testing set, with anomalies labeled in the testing set. It consists of data from 27 entities, each monitored by 55 metrics (variables). In all datasets, both point anomalies and collective anomalies are present, and true anomaly labels are available. Moreover, all methods are trained using time series data that contains anomalies, as the datasets do not provide clean time series without anomalies for training purposes. This configuration enables an investigation into the robustness of various algorithms when confronted with anomalies.
Table 1 provides a systematic comparison of the key characteristics of the three datasets (AIOps, UCR, and SMAP) used in this study. It outlines the configurations of the sliding window parameters (window size and time step), the total number of samples, the data splits across training, validation, and testing sets, as well as the anomaly proportions in the training and testing datasets. Notably, the anomaly proportions vary significantly across datasets, with AIOps containing a small proportion of anomalies in both training and testing sets, UCR having no anomalies in the training set and a low proportion in the testing set, and SMAP featuring no anomalies in the training set but a relatively higher anomaly proportion in the testing set. This highlights the diverse nature of the datasets and their suitability for different anomaly detection tasks.

Regarding baseline methods, we select two shallow machine learning methods, including OC-SVM [25] and Random Cut Forest (RCF) [26]. In deep learning algorithms, we compare our approach with five algorithms, including the deep one-class method SVDD [12], context-based anomaly detection for time series (TS-TCC) [27], and Ensemble Detection Method AOC [28]. Lastly, we chose two variants of the ROC method for ablation experiments: NoOC, which sets
In our experiments, we primarily employed PA, PW, and Affiliation (precision recall and F1-score) [29] as evaluation metrics, as they align well with the unique requirements of time series anomaly detection tasks.
PA measures the ratio of correctly classified points (both normal and anomalous) to the total number of points in the time series, offering a global perspective on the model’s overall performance. Its formula is as follows:
PA is particularly suitable for scenarios where the primary goal is to assess the model’s general classification accuracy across both normal and anomalous data. However, it may have limitations in datasets where normal points significantly outnumber anomalies, as the metric can dilute the model’s anomaly detection performance by emphasizing overall accuracy.
PW on the other hand, is designed to focus specifically on anomaly detection by emphasizing the precision and recall of the model when identifying anomalies. It provides a more refined measure of the model’s effectiveness in distinguishing anomalous data from normal data [30,31]. The formula for PW Precision is:
PW is particularly well-suited for time series anomaly detection tasks where the primary focus is on ensuring the accurate identification of anomalous samples. This metric is valuable in applications such as fault detection in industrial systems, where missing an anomaly (false negative) or misclassifying a normal event (false positive) can lead to significant consequences.
The choice of PA and PW as evaluation metrics reflects their ability to complement each other in time series anomaly detection scenarios. PA offers a holistic view of the model’s classification accuracy, while PW ensures the model’s stability and effectiveness in specifically detecting anomalies. This dual perspective allows for a balanced evaluation of the model’s performance in time series tasks, particularly when addressing real-world applications with imbalanced data distributions or critical anomaly detection requirements.
By integrating these metrics, we can better assess the trade-offs between general accuracy and the precision-recall balance in anomaly detection, ensuring the model’s applicability across diverse time series datasets and real-world tasks.
In the ROC framework, we utilize two identical three-layer LSTMs (with a dropout rate of 0.45) as Seq2Seq autoencoders. The Adam optimizer was employed with a learning rate of
In our study, we presented the prediction accuracy of affiliation, along with the corresponding point accuracy (PA) and range prediction (PW) scores. The results indicate that our method performs well across multiple datasets. Although it did not achieve the best results on the UCR dataset, it maintained a strong competitive edge relative to other methods. It is noteworthy that RCF and LSTM exhibited excellent performance on the UCR dataset, but their accuracy significantly declined on the AIOps dataset. This phenomenon can be attributed to the fact that the UCR dataset typically contains only a single anomaly segment and lacks anomalous samples in the training set. In contrast, the AIOps dataset features multiple anomaly segments, and the training set includes some anomalous characteristics, leading to higher false negative rates for methods that lack robustness.
To enhance the robustness of the model, the AOC method employs a soft-boundary strategy, which effectively improves the model’s adaptability to anomalies. In contrast, our approach integrates the RPCA method to successfully filter out a significant portion of anomalous features in the training set. This approach also yielded favorable results on the multivariate time series dataset SMAP, further validating the broad applicability of our method.
As shown in Table 2, in comparisons with various baseline models, we draw the following conclusions: RCF, as a shallow model, demonstrated outstanding performance on the UCR dataset, even surpassing some deep learning models. Meanwhile, two-stage anomaly detection methods, including SVDD and TS-TCC, did not achieve ideal results in time series anomaly detection, revealing the limitations of staged approaches for time series data and thereby constraining the performance of deep models. Additionally, our proposed ROC method performed well across all three datasets, confirming the effectiveness of the ensemble approach and its robustness against contaminated training sets.

On some datasets, although our methods are not the best, they do not fall behind much. The reason why our method does not achieve the best performance on the UCR dataset may lie in the fact that many time series in the UCR dataset exhibit strong contextual dependencies [32,33]. For instance, in motion sensor data, transitions between different actions, or in weather data, long-term trends and seasonal patterns play a significant role. Models like LSTM, which excel at handling strongly time-dependent sequential data, can effectively capture critical patterns through learning temporal transitions between states. This capability allows LSTM to achieve higher accuracy in tasks such as behavior classification and anomaly detection. As a result, methods like LSTM are more suitable for datasets with strong temporal dependencies, such as UCR.
However, for datasets like AIOps, which encompass rich operational data and diverse task scenarios, our method demonstrates superior performance. This is due to its ability to handle large-scale data with highly diverse anomaly samples and to tackle more complex tasks. In such cases, our method significantly outperforms LSTM and other approaches that rely solely on temporal dependencies.
Finally, the results from NoOC and NoRPCA indicate that combining multiple normality assumptions with anomaly filtering models significantly enhances anomaly detection (AD) performance, this further validates the efficacy and importance of the different elements within our model.
In this section, we perform a hyperparameter analysis on the AIOps dataset, with a specific focus on examining two key parameters:

Figure 3: We conducted a hyperparameter analysis on AIOps for
We also conducted detailed experiments to investigate the reasons behind the performance decline associated with varying

Figure 4: The figure illustrates how the anomaly scores change with variations in
4.5 Optimization Algorithm Analysis
We employ the gradient descent method as the core optimization algorithm to minimize the objective function. Gradient descent iteratively updates the parameters by computing the gradient of the loss function with respect to the model’s parameters, ensuring a systematic approach to minimizing loss.
In addition, we incorporate a dynamically adjusted learning rate during the optimization process. The dynamic adjustment of the learning rate allows the algorithm to take larger steps when far from the optimal solution to accelerate convergence, while automatically reducing the step size as it approaches the optimal solution. This mechanism helps to avoid overshooting the minimum and improves stability near the global optimum. More importantly, the dynamically adjusted learning rate mitigates the risk of the algorithm getting trapped in local minima, a common issue in non-convex optimization problems.
By comparing the effects of dynamic and fixed learning rates, Fig. 5 provides a visual representation of how the convergence behavior differs under different learning rate strategies. The dynamic learning rate strategy demonstrates faster convergence and better adaptability to the optimization landscape, particularly in complex scenarios.

Figure 5: Figure demonstrates the convergence process of the optimization algorithm
In addition, we visualized the actual convergence process of the model, demonstrating the step-by-step reduction in the loss function values during optimization. Fig. 6 not only provides an intuitive comparison between the performance of dynamic learning rate adjustment and fixed learning rate strategies but also strongly supports the effectiveness of the dynamic learning rate. Specifically, the dynamic learning rate facilitates a faster reduction in loss values and exhibits greater stability as it approaches the global optimum. This indicates that the dynamic learning rate has significant advantages in optimizing non-convex problems and handling complex objective functions. Furthermore, it validates the applicability of this approach in addressing challenging tasks.

Figure 6: The optimization process with dynamic learning rate and fixed learning rate as a function of epochs
This paper introduces a robust time series anomaly detection method, ROC, which is grounded in multiple hypotheses and eliminates the need for pre-training. The proposed method projects the hidden representation layer of the autoencoder and integrates the objectives of both the autoencoder and one-class (OC) methods. By filtering out anomalous segments of the input time series, ROC avoids the contamination of the compression layer by anomalous features during training and resolves potential inconsistencies between the two hypotheses. This approach effectively captures normal patterns from multiple perspectives, allowing the model to learn a more comprehensive representation of typical time series data. As a result, the method demonstrates an enhanced ability to detect diverse types of anomalies. Experimental evaluations on three real-world datasets validate the superior performance of the proposed approach.
In future work, we plan to further enhance the method’s robustness against adversarial attacks in time series anomaly detection. Drawing inspiration from state-of-the-art techniques, we aim to explore feature learning from various forms of time series representations, such as residuals and frequency domains. Additionally, we intend to combine these advanced feature extraction techniques with our robust approach to filter anomalous features from multiple perspectives, ultimately improving the model’s effectiveness and adaptability in complex scenarios.
Acknowledgement: The authors appreciate the reviewers and editors for their valuable feedback on this work. We also acknowledge the providers of datasets, including AIOps and UCR.
Funding Statement: This research was supported by the National Natural Science Foundation (62202118), Guizhou Province Major Project (Qiankehe Major Project [2024]014), Science and Scientific and Technological Research Projects from Guizhou Education Department (Qianiao ji [2023]003), Hundred-level Innovative Talent Project of Guizhou Provincial Science and Technology Department (Qiankehe Platform Talent-GCC[2023]018) and Guizhou Province Major Project (Qiankehe Major Project [2024]003), Foundation of Chongqing Key Laboratory of Public Big Data Security Technology (CQKL-QJ202300001).
Author Contributions: The authors confirm contribution to the paper as follows: method proposal and implementation, experimental proof, and manuscript writing: Zhengdao Yang; experimental setting and grant support: Yuling Chen, Xuewei Wang; manuscript revision: Hui Dou, Haiwei Sang. All authors reviewed the results and approved the final version of the manuscript.
Availability of Data and Materials: The data used to support the findings of this study are available from the corresponding author upon request.
Ethics Approval: Not applicable.
Conflicts of Interest: The authors declare no conflicts of interest to report regarding the present study.
Nomenclature
| AE | Autoencoder |
| AD | Anomaly Detection |
| ADMM | Alternating Direction Method of Multipliers |
| AIOps | Artificial Intelligence for IT (Information Technology) Operations Performance Score |
| AOC | deep Autoencoding One Class |
| BACKPROP | Backpropagation |
| DDoS | Distributed Denial of Service |
| DAGMM | Deep Autoencoding Gaussian Mixture Model |
| EMA | Exponential Moving Average |
| ECG | Electrocardiogram |
| EEG | Electroencephalogram |
| GAN | Generative Adversarial Network |
| KPI | Key Performance Indicator |
| LSTM | Long Short-Term Memory |
| MP | Matrix Profile |
| OC | One-Class Classification |
| OS | Outlier Score |
| PCA | Principal Component Analysis |
| PA | Point-adjusted metrics |
| PW | Point-wise metrics |
| RCF | Random Cut Forest |
| RPCA | Robust Principal Component Analysis |
| ROC | Robust One-Class Classification Detection |
| SVD | Singular Value Decomposition |
| SVM | Support Vector Machine |
| SSA | Singular Spectrum Analysis |
| STL | Seasonal and Trend decomposition using Loess |
| SVDD | Support Vector Data Description |
| TSAD | Time Series Anomaly Detection |
| TS-TCC | Time-Series representation learning via Temporal and Contextual Contrasting |
| UCR | University of California, Riverside Time Series Anomaly Archive |
References
1. Grubbs FE. Procedures for detecting outlying observations in samples. Technometrics. 1969;11(1):1–21. doi:10.1080/00401706.1969.10490657. [Google Scholar] [CrossRef]
2. Pang G, Shen C, Cao L, Hengel AVD. Deep learning for anomaly detection: a review. ACM Comput Surv. 2021;54(2):1–38. doi:10.1145/3439950. [Google Scholar] [CrossRef]
3. Sun Y, Chen Y, Wu P, Wang X, Wang Q. DRL: dynamic rebalance learning for adversarial robustness of UAV with long-tailed distribution. Comput Commun. 2023;205(6):14–23. doi:10.1016/j.comcom.2023.04.002. [Google Scholar] [CrossRef]
4. Yang H, Wang X, Chen Y, Dou H, Zhang Y. RPU-PVB: robust object detection based on a unified metric perspective with bilinear interpolation. J Cloud Comput. 2023;12(1):169. doi:10.1186/s13677-023-00534-3. [Google Scholar] [CrossRef]
5. Yang X, Chen Y, Qian X, Li T, Lv X. BCEAD: a blockchain-empowered ensemble anomaly detection for wireless sensor network via isolation forest. Secur Commun Netw. 2021;2021(1):9430132. doi:10.1155/2021/9430132. [Google Scholar] [CrossRef]
6. Hinton GE, Salakhutdinov RR. Reducing the dimensionality of data with neural networks. Science. 2006;313(5786):504–7. doi:10.1126/science.1127647. [Google Scholar] [PubMed] [CrossRef]
7. Tishby N, Pereira FC, Bialek W. The information bottleneck method. arXiv:physics/0004057. 2000. doi:10.48550/arXiv.physics/0004057. [Google Scholar] [CrossRef]
8. Lee B, Kim S, Maqsood M, Moon J, Rho S. Advancing autoencoder architectures for enhanced anomaly detection in multivariate industrial time series. Comput Mater Contin. 2024;81(1):1275–300. doi:10.32604/cmc.2024.054826. [Google Scholar] [CrossRef]
9. Bar S, Prasad P, Sayeed MS. Enhancing internet of things intrusion detection using artificial intelligence. Comput, Mater Contin. 2024;81(1):1–23. doi:10.32604/cmc.2024.053861. [Google Scholar] [CrossRef]
10. Long G, Zhang Z. PUNet: a semi-supervised anomaly detection model for network anomaly detection based on positive unlabeled data. Comput Mater Contin. 2024;81(1):327–43. doi:10.32604/cmc.2024.054558. [Google Scholar] [CrossRef]
11. Blázquez-García A, Conde A, Mori U, Lozano JA. A review on outlier/anomaly detection in time series data. ACM Comput Surv. 2021;54(3):1–33. doi:10.1145/3444690. [Google Scholar] [CrossRef]
12. Ruff L, Vandermeulen R, Goernitz N, Deecke L, Siddiqui SA, Binder A, et al. Deep one-class classification. In: International Conference on Machine Learning; 2018; PMLR. p. 4393–402. [Google Scholar]
13. Sohn K, Li CL, Yoon J, Jin M, Pfister T. Learning and evaluating representations for deep one-class classification. arXiv:2011.02578. 2020. doi:10.48550/arXiv.2011.02578. [Google Scholar] [CrossRef]
14. Guo R, Liu H, Liu D. When deep learning-based soft sensors encounter reliability challenges: a practical knowledge-guided adversarial attack and its defense. IEEE Trans Ind Inform. 2023;20(2):2702–14. doi:10.1109/TII.2023.3297663. [Google Scholar] [CrossRef]
15. Guo R, Chen Q, Liu H, Wang W. Adversarial robustness enhancement for deep learning-based soft sensors: an adversarial training strategy using historical gradients and domain adaptation. Sensors. 2024;24(12):3909. doi:10.3390/s24123909. [Google Scholar] [PubMed] [CrossRef]
16. Schlegl T, Seeböck P, Waldstein SM, Schmidt-Erfurth U, Langs G. Unsupervised anomaly detection with generative adversarial networks to guide marker discovery. In: International Conference on Information Processing in Medical Imaging; 2017; Springer. p. 146–57. [Google Scholar]
17. Malhotra P, Ramakrishnan A, Anand G, Vig L, Agarwal P, Shroff G. LSTM-based encoder-decoder for multi-sensor anomaly detection. arXiv:1607.00148. 2016. doi:10.48550/arXiv.1607.00148. [Google Scholar] [CrossRef]
18. Zong B, Song Q, Min MR, Cheng W, Lumezanu C, Cho D, et al. Deep autoencoding gaussian mixture model for unsupervised anomaly detection. In: International Conference on Learning Representations; 2018. [Google Scholar]
19. Qiu C, Pfrommer T, Kloft M, Mandt S, Rudolph M. Neural transformation learning for deep anomaly detection beyond images. In: International Conference on Machine Learning; 2021; PMLR. p. 8703–14. [Google Scholar]
20. Candès EJ, Li X, Ma Y, Wright J. Robust principal component analysis? J ACM. 2011;58(3):1–37. doi:10.1145/1970392.197039. [Google Scholar] [CrossRef]
21. Donoho DL. For most large underdetermined systems of linear equations the minimal l1-norm solution is also the sparsest solution. Commun Pure Appl Math: A J Issued Courant Institute Math Sci. 2006;59(6):797–829. doi:10.1002/cpa.20132. [Google Scholar] [CrossRef]
22. Boyd S, Parikh N, Chu E, Peleato B, Eckstein J. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found Trends® Mach Learn. 2011;3(1):1–122. doi:10.1561/2200000016. [Google Scholar] [CrossRef]
23. Parikh N, Boyd S, et al. Proximal algorithms. Found Trends® Optim. 2014;1(3):127–239. doi:10.1561/2400000003. [Google Scholar] [CrossRef]
24. Dau HA, Bagnall A, Kamgar K, Yeh CCM, Zhu Y, Gharghabi S, et al. The UCR time series archive. IEEE/CAA J Autom Sin. 2019;6(6):1293–305. doi:10.1109/JAS.2019.1911747. [Google Scholar] [CrossRef]
25. Schölkopf B, Williamson RC, Smola A, Shawe-Taylor J, Platt J. Support vector method for novelty detection. In: Solla S, Leen T, Müller K, editors. Advances in neural information processing systems. MIT Press; 1999. [Google Scholar]
26. Guha S, Mishra N, Roy G, Schrijvers O. Robust random cut forest based anomaly detection on streams. In: International Conference on Machine Learning; 2016; PMLR. p. 2712–21. [Google Scholar]
27. Eldele E, Ragab M, Chen Z, Wu M, Kwoh CK, Li X, et al. Time-series representation learning via temporal and contextual contrasting. arXiv:2106.14112. 2021. doi:10.24963/ijcai.2021. [Google Scholar] [CrossRef]
28. Mou X, Wang R, Wang T, Sun J, Li B, Wo T, et al. Deep autoencoding one-class time series anomaly detection. In: ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); 2023, IEEE. p. 1–5. [Google Scholar]
29. Huet A, Navarro JM, Rossi D. Local evaluation of time series anomaly detection algorithms. In: Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining; 2022. p. 635–45. [Google Scholar]
30. Schmidl S, Wenig P, Papenbrock T. Anomaly detection in time series: a comprehensive evaluation. Proc VLDB Endow. 2022;15(9):1779–97. doi:10.14778/3538598.3538602. [Google Scholar] [CrossRef]
31. Su Y, Zhao Y, Niu C, Liu R, Sun W, Pei D. Robust anomaly detection for multivariate time series through stochastic recurrent neural network. In: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining; 2019. p. 2828–37. [Google Scholar]
32. Wu R, Keogh EJ. Current time series anomaly detection benchmarks are flawed and are creating the illusion of progress. IEEE Trans Knowl Data Eng. 2021;35(3):2421–9. doi:10.1109/TKDE.2021.3112126. [Google Scholar] [CrossRef]
33. Lai KH, Zha D, Xu J, Zhao Y, Wang G, Hu X. Revisiting time series outlier detection: definitions and benchmarks. In: Thirty-Fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1); 2021. [Google Scholar]
Cite This Article
Copyright © 2025 The Author(s). Published by Tech Science Press.This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


Submit a Paper
Propose a Special lssue
View Full Text
Download PDF
Downloads
Citation Tools