33 datasets found
  1. d

    Data from: Anomaly Detection in a Fleet of Systems

    • datasets.ai
    • data.nasa.gov
    • +2more
    33
    Updated Sep 10, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    National Aeronautics and Space Administration (2024). Anomaly Detection in a Fleet of Systems [Dataset]. https://datasets.ai/datasets/anomaly-detection-in-a-fleet-of-systems
    Explore at:
    33Available download formats
    Dataset updated
    Sep 10, 2024
    Dataset authored and provided by
    National Aeronautics and Space Administration
    Description

    A fleet is a group of systems (e.g., cars, aircraft) that are designed and manufactured the same way and are intended to be used the same way. For example, a fleet of delivery trucks may consist of one hundred instances of a particular model of truck, each of which is intended for the same type of service—almost the same amount of time and distance driven every day, approximately the same total weight carried, etc. For this reason, one may imagine that data mining for fleet monitoring may merely involve collecting operating data from the multiple systems in the fleet and developing some sort of model, such as a model of normal operation that can be used for anomaly detection. However, one then may realize that each member of the fleet will be unique in some ways—there will be minor variations in manufacturing, quality of parts, and usage. For this reason, the typical machine learning and statis- tics algorithm’s assumption that all the data are independent and identically distributed is not correct. One may realize that data from each system in the fleet must be treated as unique so that one can notice significant changes in the operation of that system.

  2. Anomaly Detection Market Analysis North America, Europe, APAC, South...

    • technavio.com
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Technavio, Anomaly Detection Market Analysis North America, Europe, APAC, South America, Middle East and Africa - US, Germany, UK, China, Japan - Size and Forecast 2024-2028 [Dataset]. https://www.technavio.com/report/anomaly-detection-market-industry-analysis
    Explore at:
    Dataset provided by
    TechNavio
    Authors
    Technavio
    Time period covered
    2021 - 2025
    Area covered
    United Kingdom, United States, Global
    Description

    Snapshot img

    Anomaly Detection Market Size 2024-2028

    The anomaly detection market size is forecast to increase by USD 3.71 billion at a CAGR of 13.63% between 2023 and 2028. Anomaly detection is a critical aspect of cybersecurity, particularly in sectors like healthcare where abnormal patient conditions or unusual network activity can have significant consequences. The market for anomaly detection solutions is experiencing significant growth due to several factors. Firstly, the increasing incidence of internal threats and cyber frauds has led organizations to invest in advanced tools for detecting and responding to anomalous behavior. Secondly, the infrastructural requirements for implementing these solutions are becoming more accessible, making them a viable option for businesses of all sizes. Data science and machine learning algorithms play a crucial role in anomaly detection, enabling accurate identification of anomalies and minimizing the risk of incorrect or misleading conclusions.

    However, data quality is a significant challenge in this field, as poor quality data can lead to false positives or false negatives, undermining the effectiveness of the solution. Overall, the market for anomaly detection solutions is expected to grow steadily in the coming years, driven by the need for enhanced cybersecurity and the increasing availability of advanced technologies.

    What will be the Anomaly Detection Market Size During the Forecast Period?

    Request Free Sample

    Anomaly detection, also known as outlier detection, is a critical data analysis technique used to identify observations or events that deviate significantly from the normal behavior or expected patterns in data. These deviations, referred to as anomalies or outliers, can indicate infrastructure failures, breaking changes, manufacturing defects, equipment malfunctions, or unusual network activity. In various industries, including manufacturing, cybersecurity, healthcare, and data science, anomaly detection plays a crucial role in preventing incorrect or misleading conclusions. Artificial intelligence and machine learning algorithms, such as statistical tests (Grubbs test, Kolmogorov-Smirnov test), decision trees, isolation forest, naive Bayesian, autoencoders, local outlier factor, and k-means clustering, are commonly used for anomaly detection.

    Furthermore, these techniques help identify anomalies by analyzing data points and their statistical properties using charts, visualization, and ML models. For instance, in manufacturing, anomaly detection can help identify defective products, while in cybersecurity, it can detect unusual network activity. In healthcare, it can be used to identify abnormal patient conditions. By applying anomaly detection techniques, organizations can proactively address potential issues and mitigate risks, ensuring optimal performance and security.

    Market Segmentation

    The market research report provides comprehensive data (region-wise segment analysis), with forecasts and estimates in 'USD billion' for the period 2024-2028, as well as historical data from 2018-2022 for the following segments.

    Deployment
    
      Cloud
      On-premise
    
    
    Geography
    
      North America
    
        US
    
    
      Europe
    
        Germany
        UK
    
    
      APAC
    
        China
        Japan
    
    
      South America
    
    
    
      Middle East and Africa
    

    By Deployment Insights

    The cloud segment is estimated to witness significant growth during the forecast period. The market is witnessing a notable shift towards cloud-based solutions due to their numerous advantages over traditional on-premises systems. Cloud-based anomaly detection offers breaking changes such as quicker deployment, enhanced flexibility, and scalability, real-time data visibility, and customization capabilities. These features are provided by service providers with flexible payment models like monthly subscriptions and pay-as-you-go, making cloud-based software a cost-effective and economical choice. Anodot, Ltd, Cisco Systems Inc, IBM Corp, and SAS Institute Inc are some prominent companies offering cloud-based anomaly detection solutions in addition to on-premise alternatives. In the context of security threats, architectural optimization, marketing strategies, finance, fraud detection, manufacturing, and defects, equipment malfunctions, cloud-based anomaly detection is becoming increasingly popular due to its ability to provide real-time insights and swift response to anomalies.

    Get a glance at the market share of various segments Request Free Sample

    The cloud segment accounted for USD 1.59 billion in 2018 and showed a gradual increase during the forecast period.

    Regional Insights

    When it comes to Anomaly Detection Market growth, North America is estimated to contribute 37% to the global market during the forecast period. Technavio's analysts have elaborately explained the regional trends and drivers that shape the market during the forecast per

  3. Data from: Automated, Unsupervised, and Auto-parameterized Inference of Data...

    • zenodo.org
    pdf, zip
    Updated Jan 22, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Qiaolin Qin; Qiaolin Qin (2025). Automated, Unsupervised, and Auto-parameterized Inference of Data Patterns and Anomaly Detection [Dataset]. http://doi.org/10.48550/arxiv.2412.05240
    Explore at:
    pdf, zipAvailable download formats
    Dataset updated
    Jan 22, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Qiaolin Qin; Qiaolin Qin
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This repository contains the replication package for the paper "Automated, Unsupervised, and Auto-parameterized Inference of Data Patterns and Anomaly Detection".

    "Discover-Data-Quality-With-RIOLU-A-Replication-Package" Folder structure:

    ├── ablation_study
    ├── 20_subsampling.py
    ├── no_selection.py
    ├── static_rEM_1.py
    ├── static_rcov_95.py
    ├── static_selection_threshold.py
    └── readme.md
    ├── ground_truth_anomaly_detection (Data ground truths)
    ├── images
    ├── java_repo_exploration
    ├── java_names
    ├── java_naming_anomalies
    └── readme.md
    ├── sensitivity_analysis
    ├── Auto_RIOLU_alt_inircov.py
    ├── Auto_RIOLU_alt_nsubset.py
    └── readme.md
    ├── test_anomaly_detection
    ├── chatgpt_sampled (Data sampled for ChatGPT & the extracted regexes)
    ├── flights
    ├── hosp_1k
    ├── hosp_10k
    ├── hosp_100k
    ├── movies
    └── readme.md
    ├── test_data_profiling
    ├── hetero
    ├── homo.simple
    ├── homo
    ├── GPT_responses.csv (ChatGPT profiling responses & the extracted regexes)
    └── readme.md
    ├── Auto-RIOLU.py (Auto-RIOLU for anomaly detection)
    ├── Guided-RIOLU.py (Guided-RIOLU for anomaly detection)
    ├── pattern_generator.py
    ├── pattern_selector.py
    ├── pattern_summarizer.py
    ├── test_profiling.py (RIOLU for data profiling)
    ├── utils.py
    ├── LICENSE
    └── readme.md

  4. Discovering Anomalous Aviation Safety Events Using Scalable Data Mining...

    • data.nasa.gov
    • datadiscoverystudio.org
    • +6more
    application/rdfxml +5
    Updated Jun 26, 2018
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2018). Discovering Anomalous Aviation Safety Events Using Scalable Data Mining Algorithms [Dataset]. https://data.nasa.gov/dataset/Discovering-Anomalous-Aviation-Safety-Events-Using/2bmi-rg8d
    Explore at:
    tsv, csv, json, application/rdfxml, application/rssxml, xmlAvailable download formats
    Dataset updated
    Jun 26, 2018
    License

    U.S. Government Workshttps://www.usa.gov/government-works
    License information was derived automatically

    Description

    The worldwide civilian aviation system is one of the most complex dynamical systems created. Most modern commercial aircraft have onboard flight data recorders that record several hundred discrete and continuous parameters at approximately 1Hz for the entire duration of the flight. These data contain information about the flight control systems, actuators, engines, landing gear, avionics, and pilot commands. In this paper, recent advances in the development of a novel knowledge discovery process consisting of a suite of data mining techniques for identifying precursors to aviation safety incidents are discussed. The data mining techniques include scalable multiple-kernel learning for large-scale distributed anomaly detection. A novel multivariate time-series search algorithm is used to search for signatures of discovered anomalies on massive datasets. The process can identify operationally significant events due to environmental, mechanical, and human factors issues in the high-dimensional flight operations quality assurance data. All discovered anomalies are validated by a team of independent domain experts. This novel automated knowledge discovery process is aimed at complementing the state-of-the-art human-generated exceedance-based analysis that fails to discover previously unknown aviation safety incidents. In this paper, the discovery pipeline, the methods used, and some of the significant anomalies detected on real-world commercial aviation data are discussed.

  5. d

    pyhydroqc Sensor Data QC: Single Site Example

    • search.dataone.org
    • hydroshare.org
    • +1more
    Updated Dec 30, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Amber Spackman Jones (2023). pyhydroqc Sensor Data QC: Single Site Example [Dataset]. http://doi.org/10.4211/hs.92f393cbd06b47c398bdd2bbb86887ac
    Explore at:
    Dataset updated
    Dec 30, 2023
    Dataset provided by
    Hydroshare
    Authors
    Amber Spackman Jones
    Time period covered
    Jan 1, 2017 - Dec 31, 2017
    Description

    This resource contains an example script for using the software package pyhydroqc. pyhydroqc was developed to identify and correct anomalous values in time series data collected by in situ aquatic sensors. For more information, see the code repository: https://github.com/AmberSJones/pyhydroqc and the documentation: https://ambersjones.github.io/pyhydroqc/. The package may be installed from the Python Package Index.

    This script applies the functions to data from a single site in the Logan River Observatory, which is included in the repository. The data collected in the Logan River Observatory are sourced at http://lrodata.usu.edu/tsa/ or on HydroShare: https://www.hydroshare.org/search/?q=logan%20river%20observatory.

    Anomaly detection methods include ARIMA (AutoRegressive Integrated Moving Average) and LSTM (Long Short Term Memory). These are time series regression methods that detect anomalies by comparing model estimates to sensor observations and labeling points as anomalous when they exceed a threshold. There are multiple possible approaches for applying LSTM for anomaly detection/correction. - Vanilla LSTM: uses past values of a single variable to estimate the next value of that variable. - Multivariate Vanilla LSTM: uses past values of multiple variables to estimate the next value for all variables. - Bidirectional LSTM: uses past and future values of a single variable to estimate a value for that variable at the time step of interest. - Multivariate Bidirectional LSTM: uses past and future values of multiple variables to estimate a value for all variables at the time step of interest.

    The correction approach uses piecewise ARIMA models. Each group of consecutive anomalous points is considered as a unit to be corrected. Separate ARIMA models are developed for valid points preceding and following the anomalous group. Model estimates are blended to achieve a correction.

    The anomaly detection and correction workflow involves the following steps: 1. Retrieving data 2. Applying rules-based detection to screen data and apply initial corrections 3. Identifying and correcting sensor drift and calibration (if applicable) 4. Developing a model (i.e., ARIMA or LSTM) 5. Applying model to make time series predictions 6. Determining a threshold and detecting anomalies by comparing sensor observations to modeled results 7. Widening the window over which an anomaly is identified 8. Aggregating detections resulting from multiple models 9. Making corrections for anomalous events

    Instructions to run the notebook through the CUAHSI JupyterHub: 1. Click "Open with..." at the top of the resource and select the CUAHSI JupyterHub. You may need to sign into CUAHSI JupyterHub using your HydroShare credentials. 2. Select 'Python 3.8 - Scientific' as the server and click Start. 2. From your JupyterHub directory, click on the ExampleNotebook.ipynb file. 3. Execute each cell in the code by clicking the Run button.

  6. GECCO Industrial Challenge 2018 Dataset: A water quality dataset for the...

    • zenodo.org
    • data.niaid.nih.gov
    csv, pdf, zip
    Updated Jul 19, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Steffen Moritz; Steffen Moritz; Frederik Rehbach; Sowmya Chandrasekaran; Margarita Rebolledo; Thomas Bartz-Beielstein; Thomas Bartz-Beielstein; Frederik Rehbach; Sowmya Chandrasekaran; Margarita Rebolledo (2024). GECCO Industrial Challenge 2018 Dataset: A water quality dataset for the 'Internet of Things: Online Anomaly Detection for Drinking Water Quality' competition at the Genetic and Evolutionary Computation Conference 2018, Kyoto, Japan. [Dataset]. http://doi.org/10.5281/zenodo.3884398
    Explore at:
    zip, csv, pdfAvailable download formats
    Dataset updated
    Jul 19, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Steffen Moritz; Steffen Moritz; Frederik Rehbach; Sowmya Chandrasekaran; Margarita Rebolledo; Thomas Bartz-Beielstein; Thomas Bartz-Beielstein; Frederik Rehbach; Sowmya Chandrasekaran; Margarita Rebolledo
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Dataset of the 'Internet of Things: Online Anomaly Detection for Drinking Water Quality' competition hosted at The Genetic and Evolutionary Computation Conference (GECCO) July 15th-19th 2018, Kyoto, Japan

    The task of the competition was to develop an anomaly detection algorithm for a water- and environmental data set.

    Included in zenodo:

    - dataset of water quality data

    - additional material and descriptions provided for the competition

    The competition was organized by:

    F. Rehbach, M. Rebolledo, S. Moritz, S. Chandrasekaran, T. Bartz-Beielstein (TH Köln)

    The dataset was provided by:

    Thüringer Fernwasserversorgung and IMProvT research project

    GECCO Industrial Challenge: 'Internet of Things: Online Anomaly Detection for Drinking Water Quality'

    Description:

    For the 7th time in GECCO history, the SPOTSeven Lab is hosting an industrial challenge in cooperation with various industry partners. This years challenge, based on the 2017 challenge, is held in cooperation with "Thüringer Fernwasserversorgung" which provides their real-world data set. The task of this years competition is to develop an anomaly detection algorithm for the water- and environmental data set. Early identification of anomalies in water quality data is a challenging task. It is important to identify true undesirable variations in the water quality. At the same time, false alarm rates have to be very low.
    Additionally to the competition, for the first time in GECCO history we are now able to provide the opportunity for all participants to submit 2-page algorithm descriptions for the GECCO Companion. Thus, it is now possible to create publications in a similar procedure to the Late Breaking Abstracts (LBAs) directly through competition participation!

    Accepted Competition Entry Abstracts
    - Online Anomaly Detection for Drinking Water Quality Using a Multi-objective Machine Learning Approach (Victor Henrique Alves Ribeiro and Gilberto Reynoso Meza from the Pontifical Catholic University of Parana)
    - Anomaly Detection for Drinking Water Quality via Deep BiLSTM Ensemble (Xingguo Chen, Fan Feng, Jikai Wu, and Wenyu Liu from the Nanjing University of Posts and Telecommunications and Nanjing University)
    - Automatic vs. Manual Feature Engineering for Anomaly Detection of Drinking-Water Quality (Valerie Aenne Nicola Fehst from idatase GmbH)

    Official webpage:

    http://www.spotseven.de/gecco/gecco-challenge/gecco-challenge-2018/

  7. Data from: Multi-Source Distributed System Data for AI-powered Analytics

    • zenodo.org
    • data.niaid.nih.gov
    zip
    Updated Nov 10, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sasho Nedelkoski; Jasmin Bogatinovski; Ajay Kumar Mandapati; Soeren Becker; Jorge Cardoso; Odej Kao; Sasho Nedelkoski; Jasmin Bogatinovski; Ajay Kumar Mandapati; Soeren Becker; Jorge Cardoso; Odej Kao (2022). Multi-Source Distributed System Data for AI-powered Analytics [Dataset]. http://doi.org/10.5281/zenodo.3549604
    Explore at:
    zipAvailable download formats
    Dataset updated
    Nov 10, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Sasho Nedelkoski; Jasmin Bogatinovski; Ajay Kumar Mandapati; Soeren Becker; Jorge Cardoso; Odej Kao; Sasho Nedelkoski; Jasmin Bogatinovski; Ajay Kumar Mandapati; Soeren Becker; Jorge Cardoso; Odej Kao
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Abstract:

    In recent years there has been an increased interest in Artificial Intelligence for IT Operations (AIOps). This field utilizes monitoring data from IT systems, big data platforms, and machine learning to automate various operations and maintenance (O&M) tasks for distributed systems.
    The major contributions have been materialized in the form of novel algorithms.
    Typically, researchers took the challenge of exploring one specific type of observability data sources, such as application logs, metrics, and distributed traces, to create new algorithms.
    Nonetheless, due to the low signal-to-noise ratio of monitoring data, there is a consensus that only the analysis of multi-source monitoring data will enable the development of useful algorithms that have better performance.
    Unfortunately, existing datasets usually contain only a single source of data, often logs or metrics. This limits the possibilities for greater advances in AIOps research.
    Thus, we generated high-quality multi-source data composed of distributed traces, application logs, and metrics from a complex distributed system. This paper provides detailed descriptions of the experiment, statistics of the data, and identifies how such data can be analyzed to support O&M tasks such as anomaly detection, root cause analysis, and remediation.

    General Information:

    This repository contains the simple scripts for data statistics, and link to the multi-source distributed system dataset.

    You may find details of this dataset from the original paper:

    Sasho Nedelkoski, Jasmin Bogatinovski, Ajay Kumar Mandapati, Soeren Becker, Jorge Cardoso, Odej Kao, "Multi-Source Distributed System Data for AI-powered Analytics".

    If you use the data, implementation, or any details of the paper, please cite!

    BIBTEX:

    _

    @inproceedings{nedelkoski2020multi,
     title={Multi-source Distributed System Data for AI-Powered Analytics},
     author={Nedelkoski, Sasho and Bogatinovski, Jasmin and Mandapati, Ajay Kumar and Becker, Soeren and Cardoso, Jorge and Kao, Odej},
     booktitle={European Conference on Service-Oriented and Cloud Computing},
     pages={161--176},
     year={2020},
     organization={Springer}
    }
    

    _

    The multi-source/multimodal dataset is composed of distributed traces, application logs, and metrics produced from running a complex distributed system (Openstack). In addition, we also provide the workload and fault scripts together with the Rally report which can serve as ground truth. We provide two datasets, which differ on how the workload is executed. The sequential_data is generated via executing workload of sequential user requests. The concurrent_data is generated via executing workload of concurrent user requests.

    The raw logs in both datasets contain the same files. If the user wants the logs filetered by time with respect to the two datasets, should refer to the timestamps at the metrics (they provide the time window). In addition, we suggest to use the provided aggregated time ranged logs for both datasets in CSV format.

    Important: The logs and the metrics are synchronized with respect time and they are both recorded on CEST (central european standard time). The traces are on UTC (Coordinated Universal Time -2 hours). They should be synchronized if the user develops multimodal methods. Please read the IMPORTANT_experiment_start_end.txt file before working with the data.

    Our GitHub repository with the code for the workloads and scripts for basic analysis can be found at: https://github.com/SashoNedelkoski/multi-source-observability-dataset/

  8. Solving a prisoner's dilemma in distributed anomaly detection

    • data.staging.idas-ds1.appdat.jsc.nasa.gov
    • datasets.ai
    • +5more
    Updated Feb 18, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    nasa.gov (2025). Solving a prisoner's dilemma in distributed anomaly detection [Dataset]. https://data.staging.idas-ds1.appdat.jsc.nasa.gov/dataset/solving-a-prisoners-dilemma-in-distributed-anomaly-detection
    Explore at:
    Dataset updated
    Feb 18, 2025
    Dataset provided by
    NASAhttp://nasa.gov/
    Description

    Anomaly detection has recently become an important problem in many industrial and financial applications. In several instances, the data to be analyzed for possible anomalies is located at multiple sites and cannot be merged due to practical constraints such as bandwidth limitations and proprietary concerns. At the same time, the size of data sets affects prediction quality in almost all data mining applications. In such circumstances, distributed data mining algorithms may be used to extract information from multiple data sites in order to make better predictions. In the absence of theoretical guarantees, however, the degree to which data decentralization affects the performance of these algorithms is not known, which reduces the data providing participants' incentive to cooperate.This creates a metaphorical 'prisoners' dilemma' in the context of data mining. In this work, we propose a novel general framework for distributed anomaly detection with theoretical performance guarantees. Our algorithmic approach combines existing anomaly detection procedures with a novel method for computing global statistics using local sufficient statistics. We show that the performance of such a distributed approach is indistinguishable from that of a centralized instantiation of the same anomaly detection algorithm, a condition that we call zero information loss. We further report experimental results on synthetic as well as real-world data to demonstrate the viability of our approach. The remaining content of this presentation is presented in Fig. 1.

  9. Data from: Fleet Level Anomaly Detection of Aviation Safety Data

    • data.nasa.gov
    • data.staging.idas-ds1.appdat.jsc.nasa.gov
    • +1more
    application/rdfxml +5
    Updated Jun 26, 2018
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2018). Fleet Level Anomaly Detection of Aviation Safety Data [Dataset]. https://data.nasa.gov/w/n7yu-ua2x/_variation_?cur=kQyfdqGHa1-&from=root
    Explore at:
    application/rdfxml, application/rssxml, xml, csv, json, tsvAvailable download formats
    Dataset updated
    Jun 26, 2018
    License

    U.S. Government Workshttps://www.usa.gov/government-works
    License information was derived automatically

    Description

    For the purposes of this paper, the National Airspace System (NAS) encompasses the operations of all aircraft which are subject to air traffic control procedures. The NAS is a highly complex dynamic system that is sensitive to aeronautical decision-making and risk management skills. In order to ensure a healthy system with safe flights a systematic approach to anomaly detection is very important when evaluating a given set of circumstances and for determination of the best possible course of action. Given the fact that the NAS is a vast and loosely integrated network of systems, it requires improved safety assurance capabilities to maintain an extremely low accident rate under increasingly dense operating conditions. Data mining based tools and techniques are required to support and aid operators’ (such as pilots, management, or policy makers) overall decision-making capacity. Within the NAS, the ability to analyze fleetwide aircraft data autonomously is still considered a significantly challenging task. For our purposes a fleet is defined as a group of aircraft sharing generally compatible parameter lists. Here, in this effort, we aim at developing a system level analysis scheme. In this paper we address the capability for detection of fleetwide anomalies as they occur, which itself is an important initiative toward the safety of the real-world flight operations. The flight data recorders archive millions of data points with valuable information on flights everyday. The operational parameters consist of both continuous and discrete (binary & categorical) data from several critical subsystems and numerous complex procedures. In this paper, we discuss a system level anomaly detection approach based on the theory of kernel learning to detect potential safety anomalies in a very large data base of commercial aircraft. We also demonstrate that the proposed approach uncovers some operationally significant events due to environmental, mechanical, and human factors issues in high dimensional, multivariate Flight Operations Quality Assurance (FOQA) data. We present the results of our detection algorithms on real FOQA data from a regional carrier.

  10. Numbers of palindromes as a function of Mason_variator iterations.

    • plos.figshare.com
    xls
    Updated Jun 14, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Alan F. Karr; Jason Hauzel; Adam A. Porter; Marcel Schaefer (2023). Numbers of palindromes as a function of Mason_variator iterations. [Dataset]. http://doi.org/10.1371/journal.pone.0271970.t003
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 14, 2023
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Alan F. Karr; Jason Hauzel; Adam A. Porter; Marcel Schaefer
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The genome is E. coli. Half lengths of 6, 8, 10, 12, 14 and 16 are columns and Mason_variator iterations are rows.

  11. z

    OPSSAT-AD - anomaly detection dataset for satellite telemetry

    • zenodo.org
    • data.niaid.nih.gov
    csv
    Updated Jul 9, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ruszczak Bogdan; Ruszczak Bogdan (2024). OPSSAT-AD - anomaly detection dataset for satellite telemetry [Dataset]. http://doi.org/10.5281/zenodo.12588359
    Explore at:
    csvAvailable download formats
    Dataset updated
    Jul 9, 2024
    Dataset provided by
    Ruszczak
    Authors
    Ruszczak Bogdan; Ruszczak Bogdan
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This is the AI-ready benchmark dataset (OPSSAT-AD) containing the telemetry data acquired on board OPS-SAT---a CubeSat mission that has been operated by the European Space Agency.

    It is accompanied by the paper with baseline results obtained using 30 supervised and unsupervised classic and deep machine learning algorithms for anomaly detection. They were trained and validated using the training-test dataset split introduced in this work, and we present a suggested set of quality metrics that should always be calculated to confront the new algorithms for anomaly detection while exploiting OPSSAT-AD. We believe that this work may become an important step toward building a fair, reproducible, and objective validation procedure that can be used to quantify the capabilities of the emerging anomaly detection techniques in an unbiased and fully transparent way.

    The two included files are:

    • segments.csv with the acquired telemetry signals from ESA OPS-SAT aircraft,
    • dataset.csv with the extracted, synthetic features are computed for each manually split and labeled telemetry segment.

    Please have a look at our two papers commenting on this dataset:

    • The benchmark paper with results of 30 supervised and unsupervised anomaly detection models for this collection:
      Ruszczak, B., Kotowski. K., Nalepa, J., Evans, D.: The OPS-SAT benchmark for detecting anomalies in satellite telemetry, 2024, preprint arxiv: 2407.04730,
    • the conference paper in which we presented some preliminary results for this dataset:
      Ruszczak, B., Kotowski. K., Andrzejewski, J., et al.: (2023). Machine Learning Detects Anomalies in OPS-SAT Telemetry. Computational Science – ICCS 2023. LNCS, vol 14073. Springer, Cham, DOI:10.1007/978-3-031-35995-8_21.
  12. d

    Supporting data and tools for "Toward automating post processing of aquatic...

    • search.dataone.org
    • hydroshare.org
    • +1more
    Updated Dec 30, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Amber Spackman Jones; Tanner Jones; Jeffery S. Horsburgh (2023). Supporting data and tools for "Toward automating post processing of aquatic sensor data" [Dataset]. https://search.dataone.org/view/sha256%3A5ebbad6b4db49f3718e386c4bfc53a14c5d5f5b6b1d030f7b518320cc5311a4c
    Explore at:
    Dataset updated
    Dec 30, 2023
    Dataset provided by
    Hydroshare
    Authors
    Amber Spackman Jones; Tanner Jones; Jeffery S. Horsburgh
    Time period covered
    Jan 1, 2013 - Dec 31, 2019
    Area covered
    Description

    This resource contains the supporting data and code files for the analyses presented in "Toward automating post processing of aquatic sensor data," an article published in the journal Environmental Modelling and Software. This paper describes pyhydroqc, a Python package developed to identify and correct anomalous values in time series data collected by in situ aquatic sensors. For more information on pyhydroqc, see the code repository (https://github.com/AmberSJones/pyhydroqc) and the documentation (https://ambersjones.github.io/pyhydroqc/). The package may be installed from the Python Package Index (more info: https://packaging.python.org/tutorials/installing-packages/).

    Included in this resource are input data, Python scripts to run the package on the input data (anomaly detection and correction), results from running the algorithm, and Python scripts for generating the figures in the manuscript. The organization and structure of the files are described in detail in the readme file. The input data were collected as part of the Logan River Observatory (LRO). The data in this resource represent a subset of data available for the LRO and were compiled by querying the LRO’s operational database. All available data for the LRO can be sourced at http://lrodata.usu.edu/tsa/ or on HydroShare: https://www.hydroshare.org/search/?q=logan%20river%20observatory.

    There are two sets of scripts in this resource: 1.) Scripts that reproduce plots for the paper using saved results, and 2.) Code used to generate the complete results for the series in the case study. While all figures can be reproduced, there are challenges to running the code for the complete results (it is computationally intensive, different results will be generated due to the stochastic nature of the models, and the code was developed with an early version of the package), which is why the saved results are included in this resource. For a simple example of running pyhydroqc functions for anomaly detection and correction on a subset of data, see this resource: https://www.hydroshare.org/resource/92f393cbd06b47c398bdd2bbb86887ac/.

  13. d

    Developing Standardized Testing Datasets for Benchmarking Automated QC...

    • search.dataone.org
    Updated Mar 15, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ehsan Kahrizi (2025). Developing Standardized Testing Datasets for Benchmarking Automated QC Algorithm Performance [Dataset]. https://search.dataone.org/view/sha256%3A6e32c460ddc4bc00c48ad1582047e7509bfb09f11efa4171d5a144e48ac8c893
    Explore at:
    Dataset updated
    Mar 15, 2025
    Dataset provided by
    Hydroshare
    Authors
    Ehsan Kahrizi
    Description

    Diagnose Aquatic Sensor Data for Temperature and Water Quality Events

    Overview

    This project is designed to diagnose and flag events in aquatic sensor data based on various conditions and thresholds. It processes raw data from aquatic sites and applies thresholds and logical conditions to identify different types of anomalies. The primary focus is to flag events that may indicate sensor anomalies, environmental conditions (e.g., frozen water), or technician site visits.

    Key Features

    1. Event Detection: Detects and flags various event types, such as MNT (maintenance), LWT (low water table), ICE (frozen water), SLM (sensor logger malfunction), PF (power failure), and VIN (visual inspection).
    2. Data Quality Control: Uses thresholds to validate sensor readings, ensuring accurate representation of water conditions.
    3. Automated Labelling: Automatically labels events using a set of predefined indicators for anomaly detection.

    Workflow of the model: https://ibb.co/8BDFjsv

  14. GECCO Industrial Challenge 2017 Dataset: A water quality dataset for the...

    • zenodo.org
    • data.niaid.nih.gov
    csv, pdf, zip
    Updated Jul 19, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Steffen Moritz; Steffen Moritz; Martina Friese; Jörg Stork; Margarita Rebolledo; Andreas Fischbach; Thomas Bartz-Beielstein; Thomas Bartz-Beielstein; Martina Friese; Jörg Stork; Margarita Rebolledo; Andreas Fischbach (2024). GECCO Industrial Challenge 2017 Dataset: A water quality dataset for the 'Monitoring of drinking-water quality' competition at the Genetic and Evolutionary Computation Conference 2017, Berlin, Germany. [Dataset]. http://doi.org/10.5281/zenodo.3884465
    Explore at:
    pdf, csv, zipAvailable download formats
    Dataset updated
    Jul 19, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Steffen Moritz; Steffen Moritz; Martina Friese; Jörg Stork; Margarita Rebolledo; Andreas Fischbach; Thomas Bartz-Beielstein; Thomas Bartz-Beielstein; Martina Friese; Jörg Stork; Margarita Rebolledo; Andreas Fischbach
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    Berlin
    Description

    Dataset of the 'Industrial Challenge: Monitoring of drinking-water quality' competition hosted at The Genetic and Evolutionary Computation Conference (GECCO) July 15th-19th 2017, Berlin, Germany

    The task of the competition was to develop an anomaly detection algorithm for a water- and environmental data set.

    Included in zenodo:

    - dataset of water quality data

    - additional material and descriptions provided for the competition

    The competition was organized by:

    M. Friese, J. Stork, A. Fischbach, M. Rebolledo, T. Bartz-Beielstein (TH Köln)

    The dataset was provided and prepared by:

    Thüringer Fernwasserversorgung,

    IMProvT research project (S. Moritz)


    Industrial Challenge: Monitoring of drinking-water quality

    Description:

    Water covers 71% of the Earth's surface and is vital to all known forms of life. The provision of safe and clean drinking water to protect public health is a natural aim. Performing regular monitoring of the water-quality is essential to achieve this aim.

    Goal of the GECCO 2017 Industrial Challenge is to analyze drinking-water data and to develop a highly efficient algorithm that most accurately recognizes diverse kinds of changes in the quality of our drinking-water.

    Submission deadline:

    June 30, 2017

    Official webpage:

    http://www.spotseven.de/gecco-challenge/gecco-challenge-2017/

  15. H

    Techniques for Increased Automation of Aquatic Sensor Data Post Processing...

    • beta.hydroshare.org
    • hydroshare.org
    • +1more
    zip
    Updated Sep 7, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Amber Spackman Jones; Jeffery S. Horsburgh; Tannner Jones (2021). Techniques for Increased Automation of Aquatic Sensor Data Post Processing in Python: Video Presentation [Dataset]. https://beta.hydroshare.org/resource/bc5c616426214b60b068352ae028d963/
    Explore at:
    zip(351.0 MB)Available download formats
    Dataset updated
    Sep 7, 2021
    Dataset provided by
    HydroShare
    Authors
    Amber Spackman Jones; Jeffery S. Horsburgh; Tannner Jones
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This resource contains a video recording for a presentation given as part of the National Water Quality Monitoring Council conference in April 2021. The presentation covers the motivation for performing quality control for sensor data, the development of PyHydroQC, a Python package with functions for automating sensor quality control including anomaly detection and correction, and the performance of the algorithms applied to data from multiple sites in the Logan River Observatory.

    The initial abstract for the presentation: Water quality sensors deployed to aquatic environments make measurements at high frequency and commonly include artifacts that do not represent the environmental phenomena targeted by the sensor. Sensors are subject to fouling from environmental conditions, often exhibit drift and calibration shifts, and report anomalies and erroneous readings due to issues with datalogging, transmission, and other unknown causes. The suitability of data for analyses and decision making often depend on subjective and time-consuming quality control processes consisting of manual review and adjustment of data. Data driven and machine learning techniques have the potential to automate identification and correction of anomalous data, streamlining the quality control process. We explored documented approaches and selected several for implementation in a reusable, extensible Python package designed for anomaly detection for aquatic sensor data. Implemented techniques include regression approaches that estimate values in a time series, flag a point as anomalous if the difference between the sensor measurement exceeds a threshold, and offer replacement values for correcting anomalies. Additional algorithms that scaffold the central regression approaches include rules-based preprocessing, thresholds for determining anomalies that adjust with data variability, and the ability to detect and correct anomalies using forecasted and backcasted estimation. The techniques were developed and tested based on several years of data from aquatic sensors deployed at multiple sites in the Logan River Observatory in northern Utah, USA. Performance was assessed based on labels and corrections applied previously by trained technicians. In this presentation, we describe the techniques for detection and correction, report their performance, illustrate the workflow for applying to high frequency aquatic sensor data, and demonstrate the possibility for additional approaches to help increase automation of aquatic sensor data post processing.

  16. Z

    GECCO Industrial Challenge 2019 Dataset: A water quality dataset for the...

    • data.niaid.nih.gov
    • zenodo.org
    Updated Jul 19, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Rehbach, Frederik (2024). GECCO Industrial Challenge 2019 Dataset: A water quality dataset for the 'Internet of Things: Online Event Detection for Drinking Water Quality Control' competition at the Genetic and Evolutionary Computation Conference 2019, Prague, Czech Republic. [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_3884443
    Explore at:
    Dataset updated
    Jul 19, 2024
    Dataset provided by
    Moritz, Steffen
    Rehbach, Frederik
    Bartz-Beielstein, Thomas
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    Czechia, Prague
    Description

    Dataset of the 'Internet of Things: Online Event Detection for Drinking Water Quality Control' competition hosted at The Genetic and Evolutionary Computation Conference (GECCO) July 13th-17th 2019, Prague, Czech Republic

    The task of the competition was to develop an anomaly detection algorithm for a water- and environmental data set.

    Included in zenodo:

    1. Original train dataset of water quality data provided to participants (identical to gecco2019_train_water_quality.csv)

    2. Call for Participation

    3. Rules and Description of the Challenge

    4. Resource Package provided to participants

    5. The complete dataset, consisting of train, test and validation merged together (gecco2019_all_water_quality.csv)

    6. The test dataset, which was used for creating the leaderboard on the server (gecco2019_test_water_quality.csv)

    7. The train dataset, which participants had available for training their models (gecco2019_train_water_quality.csv)

    8. The validation dataset, which was used for the end results for the challenge (gecco2019_valid_water_quality.csv)

    The challenge required the participants to submit a program for event detection. A training dataset was available to the participants (gecco2019_train_water_quality.csv). During the challenge the participants were able to upload a version of their program to out online platform, where this version was scored against the testing dataset (gecco2019_test_water_quality.csv), thus an intermediate leaderboard was available. To avoid overfitting against this dataset, at the end of the challenge, the end result was created from scoring with the validation dataset (gecco2019_valid_water_quality.csv).

    Train, Test, Validation dataset are from the same measuring station and are in chronological order. So the timestamps from the test dataset begin directly after the train timestamps, while the validation timestamps begin directly after the test timestamps.

    The competition was organized by:

    F. Rehbach, S. Moritz, T. Bartz-Beielstein (TH Köln)

    The dataset was provided by:

    Thüringer Fernwasserversorgung and IMProvT research project

    Internet of Things: Online Event Detection for Drinking Water Quality Control

    Description:

    For the 8th time in GECCO history, the SPOTSeven Lab is hosting an industrial challenge in cooperation with various industry partners. This years challenge, based on the 2018 challenge, is held in cooperation with "Thüringer Fernwasserversorgung" which provides their real-world data set. The task of this years competition is to develop an anomaly detection algorithm for the water- and environmental data set. Early identification of anomalies in water quality data is a challenging task. It is important to identify true undesirable variations in the water quality. At the same time, false alarm rates have to be very low.

    Competition Opens: End of January/Start of February 2019 Final Submission: 30 June 2019

    Official webpage:

    https://www.th-koeln.de/informatik-und-ingenieurwissenschaften/gecco-challenge-2019_63244.php

  17. f

    Table2_A system suitability testing platform for untargeted, high-resolution...

    • frontiersin.figshare.com
    docx
    Updated Jun 13, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Andrei Dmitrenko; Michelle Reid; Nicola Zamboni (2023). Table2_A system suitability testing platform for untargeted, high-resolution mass spectrometry.DOCX [Dataset]. http://doi.org/10.3389/fmolb.2022.1026184.s005
    Explore at:
    docxAvailable download formats
    Dataset updated
    Jun 13, 2023
    Dataset provided by
    Frontiers
    Authors
    Andrei Dmitrenko; Michelle Reid; Nicola Zamboni
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The broad coverage of untargeted metabolomics poses fundamental challenges for the harmonization of measurements along time, even if they originate from the very same instrument. Internal isotopic standards can hardly cover the chemical complexity of study samples. Therefore, they are insufficient for normalizing data a posteriori as done for targeted metabolomics. Instead, it is crucial to verify instrument’s performance a priori, that is, before samples are injected. Here, we propose a system suitability testing platform for time-of-flight mass spectrometers independent of liquid chromatography. It includes a chemically defined quality control mixture, a fast acquisition method, software for extracting ca. 3,000 numerical features from profile data, and a simple web service for monitoring. We ran a pilot for 21 months and present illustrative results for anomaly detection or learning causal relationships between the spectral features and machine settings. Beyond mere detection of anomalies, our results highlight several future applications such as 1) recommending instrument retuning strategies to achieve desired values of quality indicators, 2) driving preventive maintenance, and 3) using the obtained, detailed spectral features for posterior data harmonization.

  18. d

    MKAD (Open Sourced Code)

    • catalog.data.gov
    • data.nasa.gov
    • +1more
    Updated Dec 6, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dashlink (2023). MKAD (Open Sourced Code) [Dataset]. https://catalog.data.gov/dataset/mkad-open-sourced-code
    Explore at:
    Dataset updated
    Dec 6, 2023
    Dataset provided by
    Dashlink
    Area covered
    MKAD
    Description

    The Multiple Kernel Anomaly Detection (MKAD) algorithm is designed for anomaly detection over a set of files. It combines multiple kernels into a single optimization function using the One Class Support Vector Machine (OCSVM) framework. Any kernel function can be combined in the algorithm as long as it meets the Mercer conditions, however for the purposes of this code the data preformatting and kernel type is specific to the Flight Operations Quality Assurance (FOQA) data and has been integrated into the coding steps. For this domain, discrete binary switch sequences are used in the discrete kernel, and discretized continuous parameter features are used to form the continuous kernel. The OCSVM uses a training set of nominal examples (in this case flights) and evaluates test examples for anomaly detection to determine whether they are anomalous or not. After completing this analysis the algorithm reports the anomalous examples and determines whether there is a contribution from either or both continuous and discrete elements.

  19. f

    The pseudocode of the isolated Forests.

    • plos.figshare.com
    xls
    Updated Mar 10, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zhibo Xie; Heng Long; Chengyi Ling; Yingjun Zhou; Yan Luo (2025). The pseudocode of the isolated Forests. [Dataset]. http://doi.org/10.1371/journal.pone.0315322.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Mar 10, 2025
    Dataset provided by
    PLOS ONE
    Authors
    Zhibo Xie; Heng Long; Chengyi Ling; Yingjun Zhou; Yan Luo
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Anomaly detection is widely used in cold chain logistics (CCL). But, because of the high cost and technical problem, the anomaly detection performance is poor, and the anomaly can not be detected in time, which affects the quality of goods. To solve these problems, the paper presents a new anomaly detection scheme for CCL. At first, the characteristics of the collected data of CCL are analyzed, the mathematical model of data flow is established, and the sliding window and correlation coefficient are defined. Then the abnormal events in CCL are summarized, and three types of abnormal judgment conditions based on cor-relation coefficient ρjk are deduced. A measurement anomaly detection algorithm based on the improved isolated forest algorithm is proposed. Subsampling and cross factor are designed and used to overcome the shortcomings of the isolated forest algorithm (iForest). Experiments have shown that as the dimensionality of the data increases, the performance indicators of the new scheme, such as P (precision), R (recall), F1 score, and AUC (area under the curve), become increasingly superior to commonly used support vector machines (SVM), local outlier factors (LOF), and iForests. Its average P is 0.8784, average R is 0.8731, average F1 score is 0.8639, and average AUC is 0.9064. However, the execution time of the improved algorithm is slightly longer than that of the iForest.

  20. Fault Detection and Classification Market Analysis APAC, North America,...

    • technavio.com
    Updated Jan 9, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Technavio (2024). Fault Detection and Classification Market Analysis APAC, North America, Europe, Middle East and Africa, South America - US, China, Japan, Germany, UK - Size and Forecast 2024-2028 [Dataset]. https://www.technavio.com/report/fault-detection-and-classification-market-industry-analysis
    Explore at:
    Dataset updated
    Jan 9, 2024
    Dataset provided by
    TechNavio
    Authors
    Technavio
    Time period covered
    2021 - 2025
    Area covered
    Germany, Japan, China, United Kingdom, United States, Global
    Description

    Snapshot img

    Fault Detection and Classification Market Size 2024-2028

    The fault detection and classification market size is forecast to increase by USD 2.49 billion at a CAGR of 8.83% between 2023 and 2028.

    The market is experiencing significant growth due to the increasing complexity in manufacturing processes and the integration of advanced technologies, such as artificial intelligence (AI), into industrial processes. Complex system integration, particularly in the manufacturing industry, is driving the need for robust fault detection and classification systems to ensure operational efficiency and maintain high-quality production processes. Machine vision systems are increasingly being used for inspection and fault diagnosis, employing various classification algorithms to identify and respond to faults in real-time. Fortunately, advancements in computer software, automation, and data-driven technologies offer data-driven solutions for early fault detection and classification. Real-time monitoring of vehicle components and industrial systems enables manufacturers to respond to abnormalities before they escalate into costly repairs.
    Moreover, these systems enable quality control to be performed more accurately and efficiently than with human senses alone. Throughput and response times are critical factors in the manufacturing industry, making the ability to quickly and accurately identify and classify faults essential for maintaining production and minimizing downtime. Baselines are established to monitor normal operating conditions and detect deviations, enabling proactive intervention and reducing the risk of more significant issues arising. Overall, the market for fault detection and classification solutions is poised for continued growth as industries seek to optimize their operations and improve the reliability of their complex systems.
    

    What will be the Size of the Market During the Forecast Period?

    Request Free Sample

    The industrial sector's increasing reliance on complex systems, such as engine monitoring, cybersecurity, and equipment reliability, has led to a significant demand for advanced fault detection and classification solutions. These systems, which include brake monitoring, remote monitoring, and transmission monitoring, among others, are integral to ensuring operational efficiency, system integration, and industrial data analysis. Fault detection and classification play a crucial role in optimizing cost and improving asset management in industries, such as semiconductor manufacturing and supply chain resilience.
    With the advent of Industry 4.0 and the Internet of Things (IoT), real-time diagnostics and predictive maintenance have become essential components of industrial operations. Error identification and anomaly detection are the foundation of fault detection and classification. Machine health monitoring and automated fault diagnosis enable proactive maintenance, reducing production downtime and increasing operational efficiency. Advanced analytics, condition monitoring, and sensor data analysis are key components of these solutions, providing data-driven decision-making capabilities. Cybersecurity is a critical aspect of fault detection and classification, ensuring the protection of complex systems from potential threats. Malfunction investigation and production line monitoring are essential for maintaining industrial data analysis and optimizing processes.
    Additionally, the market for fault detection and classification solutions is driven by several factors. The need for system integration and operational efficiency is a significant driver, as is the increasing complexity of industrial systems and the requirement for preventive maintenance. The adoption of digital twins and smart manufacturing is also fueling growth in this market. The cost optimization benefits of fault detection and classification are substantial. By identifying and addressing issues before they escalate, organizations can save on maintenance costs and minimize production downtime. Additionally, these solutions enable predictive maintenance, allowing for more efficient use of resources and improved asset management.
    Further, fault classification is a critical component of fault detection and classification, enabling organizations to prioritize maintenance activities and allocate resources effectively. By categorizing faults based on their severity and impact, organizations can focus on critical issues and prevent minor issues from escalating into major problems. In conclusion, the market for fault detection and classification solutions is growing rapidly, driven by the increasing complexity of industrial systems, the need for operational efficiency, and the adoption of Industry 4.0 and IoT technologies. These solutions provide significant benefits, including cost optimization, improved asset management, and enhanced cybersecurity. By investing in advanced fault de
    
Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
National Aeronautics and Space Administration (2024). Anomaly Detection in a Fleet of Systems [Dataset]. https://datasets.ai/datasets/anomaly-detection-in-a-fleet-of-systems

Data from: Anomaly Detection in a Fleet of Systems

Related Article
Explore at:
33Available download formats
Dataset updated
Sep 10, 2024
Dataset authored and provided by
National Aeronautics and Space Administration
Description

A fleet is a group of systems (e.g., cars, aircraft) that are designed and manufactured the same way and are intended to be used the same way. For example, a fleet of delivery trucks may consist of one hundred instances of a particular model of truck, each of which is intended for the same type of service—almost the same amount of time and distance driven every day, approximately the same total weight carried, etc. For this reason, one may imagine that data mining for fleet monitoring may merely involve collecting operating data from the multiple systems in the fleet and developing some sort of model, such as a model of normal operation that can be used for anomaly detection. However, one then may realize that each member of the fleet will be unique in some ways—there will be minor variations in manufacturing, quality of parts, and usage. For this reason, the typical machine learning and statis- tics algorithm’s assumption that all the data are independent and identically distributed is not correct. One may realize that data from each system in the fleet must be treated as unique so that one can notice significant changes in the operation of that system.

Search
Clear search
Close search
Google apps
Main menu