# tods **Repository Path**: tomdev/tods ## Basic Information - **Project Name**: tods - **Description**: 自动化时间序列异常检测系统 | 自动机器学习 | AutoML|异常检测 - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: master - **Homepage**: https://tods-doc.github.io/ - **GVP Project**: No ## Statistics - **Stars**: 1 - **Forks**: 24 - **Created**: 2021-06-29 - **Last Updated**: 2023-06-05 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # TODS: Automated Time-series Outlier Detection System Logo [![Build Status](https://travis-ci.org/datamllab/tods.svg?branch=master)](https://travis-ci.org/datamllab/tods) [中文文档](README.zh-CN.md) TODS is a full-stack automated machine learning system for outlier detection on multivariate time-series data. TODS provides exhaustive modules for building machine learning-based outlier detection systems, including: data processing, time series processing, feature analysis (extraction), detection algorithms, and reinforcement module. The functionalities provided via these modules include data preprocessing for general purposes, time series data smoothing/transformation, extracting features from time/frequency domains, various detection algorithms, and involving human expertise to calibrate the system. Three common outlier detection scenarios on time-series data can be performed: point-wise detection (time points as outliers), pattern-wise detection (subsequences as outliers), and system-wise detection (sets of time series as outliers), and a wide-range of corresponding algorithms are provided in TODS. This package is developed by [DATA Lab @ Texas A&M University](https://people.engr.tamu.edu/xiahu/index.html). TODS is featured for: * **Full Stack Machine Learning System** which supports exhaustive components from preprocessings, feature extraction, detection algorithms and also human-in-the loop interface. * **Wide-range of Algorithms**, including all of the point-wise detection algorithms supported by [PyOD](https://github.com/yzhao062/pyod), state-of-the-art pattern-wise (collective) detection algorithms such as [DeepLog](https://www.cs.utah.edu/~lifeifei/papers/deeplog.pdf), [Telemanon](https://arxiv.org/pdf/1802.04431.pdf), and also various ensemble algorithms for performing system-wise detection. * **Automated Machine Learning** aims to provide knowledge-free process that construct optimal pipeline based on the given data by automatically searching the best combination from all of the existing modules. ## Resources * API Documentations: [http://tods-doc.github.io](http://tods-doc.github.io) * Paper: [https://arxiv.org/abs/2009.09822](https://arxiv.org/abs/2009.09822) ## Cite this Work: If you find this work useful, you may cite this work: ``` @article{Lai_Zha_Wang_Xu_Zhao_Kumar_Chen_Zumkhawaka_Wan_Martinez_Hu_2021, title={TODS: An Automated Time Series Outlier Detection System}, volume={35}, number={18}, journal={Proceedings of the AAAI Conference on Artificial Intelligence}, author={Lai, Kwei-Herng and Zha, Daochen and Wang, Guanchu and Xu, Junjie and Zhao, Yue and Kumar, Devesh and Chen, Yile and Zumkhawaka, Purav and Wan, Minyang and Martinez, Diego and Hu, Xia}, year={2021}, month={May}, pages={16060-16062} } ``` ## Installation This package works with **Python 3.6** and pip 19+. You need to have the following packages installed on the system (for Debian/Ubuntu): ``` sudo apt-get install libssl-dev libcurl4-openssl-dev libyaml-dev build-essential libopenblas-dev libcap-dev ffmpeg ``` Clone the repository: ``` git clone https://github.com/datamllab/tods.git ``` Install locally with `pip`: ``` cd tods pip install -e . ``` # Examples Examples are available in [/examples](examples/). For basic usage, you can evaluate a pipeline on a given datasets. Here, we provide example to load our default pipeline and evaluate it on a subset of yahoo dataset. ```python import pandas as pd from tods import schemas as schemas_utils from tods import generate_dataset, evaluate_pipeline table_path = 'datasets/anomaly/raw_data/yahoo_sub_5.csv' target_index = 6 # what column is the target metric = 'F1_MACRO' # F1 on both label 0 and 1 # Read data and generate dataset df = pd.read_csv(table_path) dataset = generate_dataset(df, target_index) # Load the default pipeline pipeline = schemas_utils.load_default_pipeline() # Run the pipeline pipeline_result = evaluate_pipeline(dataset, pipeline, metric) print(pipeline_result) ``` We also provide AutoML support to help you automatically find a good pipeline for your data. ```python import pandas as pd from axolotl.backend.simple import SimpleRunner from tods import generate_dataset, generate_problem from tods.searcher import BruteForceSearch # Some information table_path = 'datasets/yahoo_sub_5.csv' target_index = 6 # what column is the target time_limit = 30 # How many seconds you wanna search metric = 'F1_MACRO' # F1 on both label 0 and 1 # Read data and generate dataset and problem df = pd.read_csv(table_path) dataset = generate_dataset(df, target_index=target_index) problem_description = generate_problem(dataset, metric) # Start backend backend = SimpleRunner(random_seed=0) # Start search algorithm search = BruteForceSearch(problem_description=problem_description, backend=backend) # Find the best pipeline best_runtime, best_pipeline_result = search.search_fit(input_data=[dataset], time_limit=time_limit) best_pipeline = best_runtime.pipeline best_output = best_pipeline_result.output # Evaluate the best pipeline best_scores = search.evaluate(best_pipeline).scores ``` # Acknowledgement We gratefully acknowledge the Data Driven Discovery of Models (D3M) program of the Defense Advanced Research Projects Agency (DARPA)