pypots


Namepypots JSON
Version 0.8.1 PyPI version JSON
download
home_pageNone
SummaryA Python Toolbox for Machine Learning on Partially-Observed Time Series
upload_time2024-09-26 05:57:24
maintainerNone
docs_urlNone
authorNone
requires_python>=3.8
licenseCopyright (c) 2023-present, Wenjie Du All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. 3. Neither the name of the copyright holder nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
keywords data science data mining neural networks machine learning deep learning artificial intelligence time-series analysis time series imputation interpolation classification clustering forecasting partially observed irregular sampled partially-observed time series incomplete time series missing data missing values
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage
            <a href="https://github.com/WenjieDu/PyPOTS">
    <img src="https://pypots.com/figs/pypots_logos/PyPOTS/logo_FFBG.svg" width="200" align="right">
</a>

<h3 align="center">Welcome to PyPOTS</h3>

<p align="center"><i>a Python toolbox for machine learning on Partially-Observed Time Series</i></p>

<p align="center">
    <a href="https://docs.pypots.com/en/latest/install.html#reasons-of-version-limitations-on-dependencies">
       <img alt="Python version" src="https://img.shields.io/badge/Python-v3.8+-E97040?logo=python&logoColor=white">
    </a>
    <a href="https://www.google.com/search?q=%22PyPOTS%22+site%3Apytorch.org">
        <img alt="powered by Pytorch" src="https://img.shields.io/badge/PyTorch-%E2%9D%A4%EF%B8%8F-F8C6B5?logo=pytorch&logoColor=white">
    </a>
    <a href="https://github.com/WenjieDu/PyPOTS/releases">
        <img alt="the latest release version" src="https://img.shields.io/github/v/release/wenjiedu/pypots?color=EE781F&include_prereleases&label=Release&logo=github&logoColor=white">
    </a>
    <a href="https://github.com/WenjieDu/PyPOTS/blob/main/LICENSE">
        <img alt="BSD-3 license" src="https://img.shields.io/badge/License-BSD--3-E9BB41?logo=opensourceinitiative&logoColor=white">
    </a>
    <a href="https://github.com/WenjieDu/PyPOTS#-community">
        <img alt="Community" src="https://img.shields.io/badge/join_us-community!-C8A062">
    </a>
    <a href="https://github.com/WenjieDu/PyPOTS/graphs/contributors">
        <img alt="GitHub contributors" src="https://img.shields.io/github/contributors/wenjiedu/pypots?color=D8E699&label=Contributors&logo=GitHub">
    </a>
    <a href="https://star-history.com/#wenjiedu/pypots">
        <img alt="GitHub Repo stars" src="https://img.shields.io/github/stars/wenjiedu/pypots?logo=None&color=6BB392&label=%E2%98%85%20Stars">
    </a>
    <a href="https://github.com/WenjieDu/PyPOTS/network/members">
        <img alt="GitHub Repo forks" src="https://img.shields.io/github/forks/wenjiedu/pypots?logo=forgejo&logoColor=black&label=Forks">
    </a>
    <a href="https://codeclimate.com/github/WenjieDu/PyPOTS">
        <img alt="Code Climate maintainability" src="https://img.shields.io/codeclimate/maintainability-percentage/WenjieDu/PyPOTS?color=3C7699&label=Maintainability&logo=codeclimate">
    </a>
    <a href="https://coveralls.io/github/WenjieDu/PyPOTS">
        <img alt="Coveralls coverage" src="https://img.shields.io/coverallsCoverage/github/WenjieDu/PyPOTS?branch=main&logo=coveralls&color=75C1C4&label=Coverage">
    </a>
    <a href="https://github.com/WenjieDu/PyPOTS/actions/workflows/testing_ci.yml">
        <img alt="GitHub Testing" src="https://img.shields.io/github/actions/workflow/status/wenjiedu/pypots/testing_ci.yml?logo=circleci&color=C8D8E1&label=CI">
    </a>
    <a href="https://docs.pypots.com">
        <img alt="Docs building" src="https://img.shields.io/readthedocs/pypots?logo=readthedocs&label=Docs&logoColor=white&color=395260">
    </a>
    <a href="https://anaconda.org/conda-forge/pypots">
        <img alt="Conda downloads" src="https://img.shields.io/endpoint?url=https://pypots.com/figs/downloads_badges/conda_pypots_downloads.json">
    </a>
    <a href="https://pepy.tech/project/pypots">
        <img alt="PyPI downloads" src="https://img.shields.io/endpoint?url=https://pypots.com/figs/downloads_badges/pypi_pypots_downloads.json">
    </a>
    <a href="https://arxiv.org/abs/2305.18811">
        <img alt="arXiv DOI" src="https://img.shields.io/badge/DOI-10.48550/arXiv.2305.18811-F8F7F0">
    </a>
    <a href="https://github.com/WenjieDu/PyPOTS/blob/main/README_zh.md">
        <img alt="README in Chinese" src="https://pypots.com/figs/pypots_logos/readme/CN.svg">
    </a>
   <a href="https://github.com/WenjieDu/PyPOTS/blob/main/README.md">
        <img alt="README in English" src="https://pypots.com/figs/pypots_logos/readme/US.svg">
    </a>
    <a href="https://github.com/WenjieDu/PyPOTS">
        <img alt="PyPOTS Hits" src="https://hits.seeyoufarm.com/api/count/incr/badge.svg?url=https%3A%2F%2Fgithub.com%2FPyPOTS%2FPyPOTS&count_bg=%23009A0A&title_bg=%23555555&icon=&icon_color=%23E7E7E7&title=Hits&edge_flat=false">
    </a>
</p>

⦿ `Motivation`: Due to all kinds of reasons like failure of collection sensors, communication error,
and unexpected malfunction, missing values are common to see in time series from the real-world environment.
This makes partially-observed time series (POTS) a pervasive problem in open-world modeling and prevents advanced
data analysis. Although this problem is important, the area of machine learning on POTS still lacks a dedicated toolkit.
PyPOTS is created to fill in this blank.

⦿ `Mission`: PyPOTS (pronounced "Pie Pots") is born to become a handy toolbox that is going to make machine learning on
POTS easy rather than tedious, to help engineers and researchers focus more on the core problems in their hands rather
than on how to deal with the missing parts in their data. PyPOTS will keep integrating classical and the latest
state-of-the-art machine learning algorithms for partially-observed multivariate time series. For sure, besides various
algorithms, PyPOTS is going to have unified APIs together with detailed documentation and interactive examples across
algorithms as tutorials.

🤗 **Please** star this repo to help others notice PyPOTS if you think it is a useful toolkit.
**Please** kindly [cite PyPOTS](https://github.com/WenjieDu/PyPOTS#-citing-pypots) in your publications if it helps with
your research.
This really means a lot to our open-source research. Thank you!

The rest of this readme file is organized as follows:
[**❖ Available Algorithms**](#-available-algorithms),
[**❖ PyPOTS Ecosystem**](#-pypots-ecosystem),
[**❖ Installation**](#-installation),
[**❖ Usage**](#-usage),
[**❖ Citing PyPOTS**](#-citing-pypots),
[**❖ Contribution**](#-contribution),
[**❖ Community**](#-community).

## ❖ Available Algorithms

PyPOTS supports imputation, classification, clustering, forecasting, and anomaly detection tasks on multivariate
partially-observed time series with missing values. The table below shows the availability of each algorithm
(sorted by Year) in PyPOTS for different tasks. The symbol `✅` indicates the algorithm is available for the
corresponding task (note that models will be continuously updated in the future to handle tasks that are not
currently supported. Stay tuned❗️).

🌟 Since **v0.2**, all neural-network models in PyPOTS has got hyperparameter-optimization support.
This functionality is implemented with the [Microsoft NNI](https://github.com/microsoft/nni) framework. You may want to
refer to our time-series imputation survey repo [Awesome_Imputation](https://github.com/WenjieDu/Awesome_Imputation)
to see how to config and tune the hyperparameters.

🔥 Note that all models whose name with `🧑‍🔧` in the table (e.g. Transformer, iTransformer, Informer etc.) are not
originally proposed as algorithms for POTS data in their papers, and they cannot directly accept time series with
missing values as input, let alone imputation. **To make them applicable to POTS data, we specifically apply the
embedding strategy and training approach (ORT+MIT) the same as we did in
[the SAITS paper](https://arxiv.org/pdf/2202.08516)[^1].**

The task types are abbreviated as follows:
**`IMPU`**: Imputation;
**`FORE`**: Forecasting;
**`CLAS`**: Classification;
**`CLUS`**: Clustering;
**`ANOD`**: Anomaly Detection.
The paper references and links are all listed at the bottom of this file.

| **Type**      | **Algo**                                                                                                                         | **IMPU** | **FORE** | **CLAS** | **CLUS** | **ANOD** | **Year - Venue**   |
|:--------------|:---------------------------------------------------------------------------------------------------------------------------------|:--------:|:--------:|:--------:|:--------:|:--------:|:-------------------|
| LLM           | <a href="https://time-series.ai"><img src="https://time-series.ai/static/figs/robot.svg" width="26px"> Time-Series.AI</a>  [^36] |    ✅     |    ✅     |    ✅     |    ✅     |    ✅     | `Later in 2024`    |
| Neural Net    | TEFN🧑‍🔧[^39]                                                                                                                   |    ✅     |          |          |          |          | `2024 - arXiv`     |
| Neural Net    | TimeMixer[^37]                                                                                                                   |    ✅     |          |          |          |          | `2024 - ICLR`      |
| Neural Net    | iTransformer🧑‍🔧[^24]                                                                                                           |    ✅     |          |          |          |          | `2024 - ICLR`      |
| Neural Net    | ModernTCN[^38]                                                                                                                   |    ✅     |          |          |          |          | `2024 - ICLR`      |
| Neural Net    | ImputeFormer🧑‍🔧[^34]                                                                                                           |    ✅     |          |          |          |          | `2024 - KDD`       |
| Neural Net    | SAITS[^1]                                                                                                                        |    ✅     |          |          |          |          | `2023 - ESWA`      |
| Neural Net    | FreTS🧑‍🔧[^23]                                                                                                                  |    ✅     |          |          |          |          | `2023 - NeurIPS`   |
| Neural Net    | Koopa🧑‍🔧[^29]                                                                                                                  |    ✅     |          |          |          |          | `2023 - NeurIPS`   |
| Neural Net    | Crossformer🧑‍🔧[^16]                                                                                                            |    ✅     |          |          |          |          | `2023 - ICLR`      |
| Neural Net    | TimesNet[^14]                                                                                                                    |    ✅     |          |          |          |          | `2023 - ICLR`      |
| Neural Net    | PatchTST🧑‍🔧[^18]                                                                                                               |    ✅     |          |          |          |          | `2023 - ICLR`      |
| Neural Net    | ETSformer🧑‍🔧[^19]                                                                                                              |    ✅     |          |          |          |          | `2023 - ICLR`      |
| Neural Net    | MICN🧑‍🔧[^27]                                                                                                                   |    ✅     |          |          |          |          | `2023 - ICLR`      |
| Neural Net    | DLinear🧑‍🔧[^17]                                                                                                                |    ✅     |          |          |          |          | `2023 - AAAI`      |
| Neural Net    | TiDE🧑‍🔧[^28]                                                                                                                   |    ✅     |          |          |          |          | `2023 - TMLR`      |
| Neural Net    | SCINet🧑‍🔧[^30]                                                                                                                 |    ✅     |          |          |          |          | `2022 - NeurIPS`   |
| Neural Net    | Nonstationary Tr.🧑‍🔧[^25]                                                                                                      |    ✅     |          |          |          |          | `2022 - NeurIPS`   |
| Neural Net    | FiLM🧑‍🔧[^22]                                                                                                                   |    ✅     |          |          |          |          | `2022 - NeurIPS`   |
| Neural Net    | RevIN_SCINet🧑‍🔧[^31]                                                                                                           |    ✅     |          |          |          |          | `2022 - ICLR`      |
| Neural Net    | Pyraformer🧑‍🔧[^26]                                                                                                             |    ✅     |          |          |          |          | `2022 - ICLR`      |
| Neural Net    | Raindrop[^5]                                                                                                                     |          |          |    ✅     |          |          | `2022 - ICLR`      |
| Neural Net    | FEDformer🧑‍🔧[^20]                                                                                                              |    ✅     |          |          |          |          | `2022 - ICML`      |
| Neural Net    | Autoformer🧑‍🔧[^15]                                                                                                             |    ✅     |          |          |          |          | `2021 - NeurIPS`   |
| Neural Net    | CSDI[^12]                                                                                                                        |    ✅     |    ✅     |          |          |          | `2021 - NeurIPS`   |
| Neural Net    | Informer🧑‍🔧[^21]                                                                                                               |    ✅     |          |          |          |          | `2021 - AAAI`      |
| Neural Net    | US-GAN[^10]                                                                                                                      |    ✅     |          |          |          |          | `2021 - AAAI`      |
| Neural Net    | CRLI[^6]                                                                                                                         |          |          |          |    ✅     |          | `2021 - AAAI`      |
| Probabilistic | BTTF[^8]                                                                                                                         |          |    ✅     |          |          |          | `2021 - TPAMI`     |
| Neural Net    | StemGNN🧑‍🔧[^33]                                                                                                                |    ✅     |          |          |          |          | `2020 - NeurIPS`   |
| Neural Net    | Reformer🧑‍🔧[^32]                                                                                                               |    ✅     |          |          |          |          | `2020 - ICLR`      |
| Neural Net    | GP-VAE[^11]                                                                                                                      |    ✅     |          |          |          |          | `2020 - AISTATS`   |
| Neural Net    | VaDER[^7]                                                                                                                        |          |          |          |    ✅     |          | `2019 - GigaSci.`  |
| Neural Net    | M-RNN[^9]                                                                                                                        |    ✅     |          |          |          |          | `2019 - TBME`      |
| Neural Net    | BRITS[^3]                                                                                                                        |    ✅     |          |    ✅     |          |          | `2018 - NeurIPS`   |
| Neural Net    | GRU-D[^4]                                                                                                                        |    ✅     |          |    ✅     |          |          | `2018 - Sci. Rep.` |
| Neural Net    | TCN🧑‍🔧[^35]                                                                                                                    |    ✅     |          |          |          |          | `2018 - arXiv`     |
| Neural Net    | Transformer🧑‍🔧[^2]                                                                                                             |    ✅     |          |          |          |          | `2017 - NeurIPS`   |
| Naive         | Lerp[^40]                                                                                                                        |    ✅     |          |          |          |          |                    |
| Naive         | LOCF/NOCB                                                                                                                        |    ✅     |          |          |          |          |                    |
| Naive         | Mean                                                                                                                             |    ✅     |          |          |          |          |                    |
| Naive         | Median                                                                                                                           |    ✅     |          |          |          |          |                    |

💯 Contribute your model right now to increase your research impact! PyPOTS downloads are increasing rapidly
(**[300K+ in total and 1K+ daily on PyPI so far](https://www.pepy.tech/projects/pypots)**),
and your work will be widely used and cited by the community.
Refer to the [contribution guide](https://github.com/WenjieDu/PyPOTS#-contribution) to see how to include your model in
PyPOTS.

## ❖ PyPOTS Ecosystem

At PyPOTS, things are related to coffee, which we're familiar with. Yes, this is a coffee universe!
As you can see, there is a coffee pot in the PyPOTS logo. And what else? Please read on ;-)

<a href="https://github.com/WenjieDu/TSDB">
    <img src="https://pypots.com/figs/pypots_logos/TSDB/logo_FFBG.svg" align="left" width="140" alt="TSDB logo"/>
</a>

👈 Time series datasets are taken as coffee beans at PyPOTS, and POTS datasets are incomplete coffee beans with missing
parts that have their own meanings. To make various public time-series datasets readily available to users,
<i>Time Series Data Beans (TSDB)</i> is created to make loading time-series datasets super easy!
Visit [TSDB](https://github.com/WenjieDu/TSDB) right now to know more about this handy tool 🛠, and it now supports a
total of 172 open-source datasets!

<a href="https://github.com/WenjieDu/PyGrinder">
    <img src="https://pypots.com/figs/pypots_logos/PyGrinder/logo_FFBG.svg" align="right" width="140" alt="PyGrinder logo"/>
</a>

👉 To simulate the real-world data beans with missingness, the ecosystem library
[PyGrinder](https://github.com/WenjieDu/PyGrinder), a toolkit helping grind your coffee beans into incomplete ones, is
created. Missing patterns fall into three categories according to Robin's theory[^13]:
MCAR (missing completely at random), MAR (missing at random), and MNAR (missing not at random).
PyGrinder supports all of them and additional functionalities related to missingness.
With PyGrinder, you can introduce synthetic missing values into your datasets with a single line of code.

<a href="https://github.com/WenjieDu/BenchPOTS">
    <img src="https://pypots.com/figs/pypots_logos/BenchPOTS/logo_FFBG.svg" align="left" width="140" alt="BenchPOTS logo"/>
</a>

👈 To fairly evaluate the performance of PyPOTS algorithms, the benchmarking suite
[BenchPOTS](https://github.com/WenjieDu/BenchPOTS) is created, which provides standard and unified data-preprocessing
pipelines to prepare datasets for measuring the performance of different POTS algorithms on various tasks.

<a href="https://github.com/WenjieDu/BrewPOTS">
    <img src="https://pypots.com/figs/pypots_logos/BrewPOTS/logo_FFBG.svg" align="right" width="140" alt="BrewPOTS logo"/>
</a>

👉 Now the beans, grinder, and pot are ready, please have a seat on the bench and let's think about how to brew us a cup
of coffee. Tutorials are necessary! Considering the future workload, PyPOTS tutorials are released in a single repo,
and you can find them in [BrewPOTS](https://github.com/WenjieDu/BrewPOTS).
Take a look at it now, and learn how to brew your POTS datasets!

<p align="center">
<a href="https://pypots.com/ecosystem/">
    <img src="https://pypots.com/figs/pypots_logos/Ecosystem/PyPOTS_Ecosystem_Pipeline.png" width="95%"/>
</a>
<br>
<b> ☕️ Welcome to the universe of PyPOTS. Enjoy it and have fun!</b>
</p>

## ❖ Installation

You can refer to [the installation instruction](https://docs.pypots.com/en/latest/install.html) in PyPOTS documentation
for a guideline with more details.

PyPOTS is available on both [PyPI](https://pypi.python.org/pypi/pypots)
and [Anaconda](https://anaconda.org/conda-forge/pypots).
You can install PyPOTS like below as well as
[TSDB](https://github.com/WenjieDu/TSDB),[PyGrinder](https://github.com/WenjieDu/PyGrinder),
[BenchPOTS](https://github.com/WenjieDu/BenchPOTS), and [AI4TS](https://github.com/WenjieDu/AI4TS):

``` bash
# via pip
pip install pypots            # the first time installation
pip install pypots --upgrade  # update pypots to the latest version
# install from the latest source code with the latest features but may be not officially released yet
pip install https://github.com/WenjieDu/PyPOTS/archive/main.zip

# via conda
conda install conda-forge::pypots  # the first time installation
conda update  conda-forge::pypots  # update pypots to the latest version
```

## ❖ Usage

Besides [BrewPOTS](https://github.com/WenjieDu/BrewPOTS), you can also find a simple and quick-start tutorial notebook
on Google Colab
<a href="https://colab.research.google.com/drive/1HEFjylEy05-r47jRy0H9jiS_WhD0UWmQ">
<img src="https://img.shields.io/badge/GoogleColab-PyPOTS_Tutorials-F9AB00?logo=googlecolab&logoColor=white" alt="Colab tutorials" align="center"/>
</a>. If you have further questions, please refer to PyPOTS documentation [docs.pypots.com](https://docs.pypots.com).
You can also [raise an issue](https://github.com/WenjieDu/PyPOTS/issues) or [ask in our community](#-community).

We present you a usage example of imputing missing values in time series with PyPOTS below, you can click it to view.

<details open>
<summary><b>Click here to see an example applying SAITS on PhysioNet2012 for imputation:</b></summary>

``` python
# Data preprocessing. Tedious, but PyPOTS can help.
import numpy as np
from sklearn.preprocessing import StandardScaler
from pygrinder import mcar
from pypots.data import load_specific_dataset
data = load_specific_dataset('physionet_2012')  # PyPOTS will automatically download and extract it.
X = data['X']
num_samples = len(X['RecordID'].unique())
X = X.drop(['RecordID', 'Time'], axis = 1)
X = StandardScaler().fit_transform(X.to_numpy())
X = X.reshape(num_samples, 48, -1)
X_ori = X  # keep X_ori for validation
X = mcar(X, 0.1)  # randomly hold out 10% observed values as ground truth
dataset = {"X": X}  # X for model input
print(X.shape)  # (11988, 48, 37), 11988 samples and each sample has 48 time steps, 37 features

# Model training. This is PyPOTS showtime.
from pypots.imputation import SAITS
from pypots.utils.metrics import calc_mae
saits = SAITS(n_steps=48, n_features=37, n_layers=2, d_model=256, n_heads=4, d_k=64, d_v=64, d_ffn=128, dropout=0.1, epochs=10)
# Here I use the whole dataset as the training set because ground truth is not visible to the model, you can also split it into train/val/test sets
saits.fit(dataset)  # train the model on the dataset
imputation = saits.impute(dataset)  # impute the originally-missing values and artificially-missing values
indicating_mask = np.isnan(X) ^ np.isnan(X_ori)  # indicating mask for imputation error calculation
mae = calc_mae(imputation, np.nan_to_num(X_ori), indicating_mask)  # calculate mean absolute error on the ground truth (artificially-missing values)
saits.save("save_it_here/saits_physionet2012.pypots")  # save the model for future use
saits.load("save_it_here/saits_physionet2012.pypots")  # reload the serialized model file for following imputation or training
```

</details>

## ❖ Citing PyPOTS

> [!TIP]
> **[Updates in Jun 2024]** 😎 The 1st comprehensive time-seres imputation benchmark paper
[TSI-Bench: Benchmarking Time Series Imputation](https://arxiv.org/abs/2406.12747) now is public available.
> The code is open source in the repo [Awesome_Imputation](https://github.com/WenjieDu/Awesome_Imputation).
> With nearly 35,000 experiments, we provide a comprehensive benchmarking study on 28 imputation methods, 3 missing
> patterns (points, sequences, blocks),
> various missing rates, and 8 real-world datasets.
>
> **[Updates in Feb 2024]** 🎉 Our survey
> paper [Deep Learning for Multivariate Time Series Imputation: A Survey](https://arxiv.org/abs/2402.04059) has been
> released on arXiv.
> We comprehensively review the literature of the state-of-the-art deep-learning imputation methods for time series,
> provide a taxonomy for them, and discuss the challenges and future directions in this field.

The paper introducing PyPOTS is available [on arXiv](https://arxiv.org/abs/2305.18811),
and a short version of it is accepted by the 9th SIGKDD international workshop on Mining and Learning from Time
Series ([MiLeTS'23](https://kdd-milets.github.io/milets2023/))).
**Additionally**, PyPOTS has been included as a [PyTorch Ecosystem](https://pytorch.org/ecosystem/) project.
We are pursuing to publish it in prestigious academic venues, e.g. JMLR (track for
[Machine Learning Open Source Software](https://www.jmlr.org/mloss/)). If you use PyPOTS in your work,
please cite it as below and 🌟star this repository to make others notice this library. 🤗

There are scientific research projects using PyPOTS and referencing in their papers.
Here is [an incomplete list of them](https://scholar.google.com/scholar?as_ylo=2022&q=%E2%80%9CPyPOTS%E2%80%9D&hl=en).

```bibtex
@article{du2023pypots,
    title = {{PyPOTS: a Python toolbox for data mining on Partially-Observed Time Series}},
    author = {Wenjie Du},
    journal = {arXiv preprint arXiv:2305.18811},
    year = {2023},
}
```

or
> Wenjie Du.
> PyPOTS: a Python toolbox for data mining on Partially-Observed Time Series.
> arXiv, abs/2305.18811, 2023.

## ❖ Contribution

You're very welcome to contribute to this exciting project!

By committing your code, you'll

1. make your well-established model out-of-the-box for PyPOTS users to run,
   and help your work obtain more exposure and impact.
   Take a look at our [inclusion criteria](https://docs.pypots.com/en/latest/faq.html#inclusion-criteria).
   You can utilize the `template` folder in each task package (e.g.
   [pypots/imputation/template](https://github.com/WenjieDu/PyPOTS/tree/main/pypots/imputation/template)) to quickly
   start;
2. become one of [PyPOTS contributors](https://github.com/WenjieDu/PyPOTS/graphs/contributors) and
   be listed as a volunteer developer [on the PyPOTS website](https://pypots.com/about/#volunteer-developers);
3. get mentioned in PyPOTS [release notes](https://github.com/WenjieDu/PyPOTS/releases);

You can also contribute to PyPOTS by simply staring🌟 this repo to help more people notice it.
Your star is your recognition to PyPOTS, and it matters!

<details open>
<summary>
    <b><i>
    👏 Click here to view PyPOTS stargazers and forkers.<br>
    We're so proud to have more and more awesome users, as well as more bright ✨stars:
    </i></b>
</summary>
<a href="https://github.com/WenjieDu/PyPOTS/stargazers">
    <img alt="PyPOTS stargazers" src="http://reporoster.com/stars/dark/WenjieDu/PyPOTS">
</a>
<br>
<a href="https://github.com/WenjieDu/PyPOTS/network/members">
    <img alt="PyPOTS forkers" src="http://reporoster.com/forks/dark/WenjieDu/PyPOTS">
</a>
</details>

👀 Check out a full list of our users' affiliations [on PyPOTS website here](https://pypots.com/users/)!

## ❖ Community

We care about the feedback from our users, so we're building PyPOTS community on

- [Slack](https://join.slack.com/t/pypots-org/shared_invite/zt-1gq6ufwsi-p0OZdW~e9UW_IA4_f1OfxA). General discussion,
  Q&A, and our development team are here;
- [LinkedIn](https://www.linkedin.com/company/pypots). Official announcements and news are here;
- [WeChat (微信公众号)](https://mp.weixin.qq.com/s/X3ukIgL1QpNH8ZEXq1YifA). We also run a group chat on WeChat,
  and you can get the QR code from the official account after following it;

If you have any suggestions or want to contribute ideas or share time-series related papers, join us and tell.
PyPOTS community is open, transparent, and surely friendly. Let's work together to build and improve PyPOTS!


[//]: # (Use APA reference style below)
[^1]: Du, W., Cote, D., & Liu, Y. (2023).
[SAITS: Self-Attention-based Imputation for Time Series](https://doi.org/10.1016/j.eswa.2023.119619).
*Expert systems with applications*.
[^2]: Vaswani, A., Shazeer, N.M., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., & Polosukhin, I. (
2017).
[Attention is All you Need](https://papers.nips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html).
*NeurIPS 2017*.
[^3]: Cao, W., Wang, D., Li, J., Zhou, H., Li, L., & Li, Y. (2018).
[BRITS: Bidirectional Recurrent Imputation for Time Series](https://papers.nips.cc/paper/2018/hash/734e6bfcd358e25ac1db0a4241b95651-Abstract.html).
*NeurIPS 2018*.
[^4]: Che, Z., Purushotham, S., Cho, K., Sontag, D.A., & Liu, Y. (2018).
[Recurrent Neural Networks for Multivariate Time Series with Missing Values](https://www.nature.com/articles/s41598-018-24271-9).
*Scientific Reports*.
[^5]: Zhang, X., Zeman, M., Tsiligkaridis, T., & Zitnik, M. (2022).
[Graph-Guided Network for Irregularly Sampled Multivariate Time Series](https://arxiv.org/abs/2110.05357).
*ICLR 2022*.
[^6]: Ma, Q., Chen, C., Li, S., & Cottrell, G. W. (2021).
[Learning Representations for Incomplete Time Series Clustering](https://ojs.aaai.org/index.php/AAAI/article/view/17070).
*AAAI 2021*.
[^7]: Jong, J.D., Emon, M.A., Wu, P., Karki, R., Sood, M., Godard, P., Ahmad, A., Vrooman, H.A., Hofmann-Apitius, M., &
Fröhlich, H. (2019).
[Deep learning for clustering of multivariate clinical patient trajectories with missing values](https://academic.oup.com/gigascience/article/8/11/giz134/5626377).
*GigaScience*.
[^8]: Chen, X., & Sun, L. (2021).
[Bayesian Temporal Factorization for Multidimensional Time Series Prediction](https://arxiv.org/abs/1910.06366).
*IEEE transactions on pattern analysis and machine intelligence*.
[^9]: Yoon, J., Zame, W. R., & van der Schaar, M. (2019).
[Estimating Missing Data in Temporal Data Streams Using Multi-Directional Recurrent Neural Networks](https://ieeexplore.ieee.org/document/8485748).
*IEEE Transactions on Biomedical Engineering*.
[^10]: Miao, X., Wu, Y., Wang, J., Gao, Y., Mao, X., & Yin, J. (2021).
[Generative Semi-supervised Learning for Multivariate Time Series Imputation](https://ojs.aaai.org/index.php/AAAI/article/view/17086).
*AAAI 2021*.
[^11]: Fortuin, V., Baranchuk, D., Raetsch, G. & Mandt, S. (2020).
[GP-VAE: Deep Probabilistic Time Series Imputation](https://proceedings.mlr.press/v108/fortuin20a.html).
*AISTATS 2020*.
[^12]: Tashiro, Y., Song, J., Song, Y., & Ermon, S. (2021).
[CSDI: Conditional Score-based Diffusion Models for Probabilistic Time Series Imputation](https://proceedings.neurips.cc/paper/2021/hash/cfe8504bda37b575c70ee1a8276f3486-Abstract.html).
*NeurIPS 2021*.
[^13]: Rubin, D. B. (1976).
[Inference and missing data](https://academic.oup.com/biomet/article-abstract/63/3/581/270932).
*Biometrika*.
[^14]: Wu, H., Hu, T., Liu, Y., Zhou, H., Wang, J., & Long, M. (2023).
[TimesNet: Temporal 2d-variation modeling for general time series analysis](https://openreview.net/forum?id=ju_Uqw384Oq).
*ICLR 2023*
[^15]: Wu, H., Xu, J., Wang, J., & Long, M. (2021).
[Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting](https://proceedings.neurips.cc/paper/2021/hash/bcc0d400288793e8bdcd7c19a8ac0c2b-Abstract.html).
*NeurIPS 2021*.
[^16]: Zhang, Y., & Yan, J. (2023).
[Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting](https://openreview.net/forum?id=vSVLM2j9eie).
*ICLR 2023*.
[^17]: Zeng, A., Chen, M., Zhang, L., & Xu, Q. (2023).
[Are transformers effective for time series forecasting?](https://ojs.aaai.org/index.php/AAAI/article/view/26317).
*AAAI 2023*
[^18]: Nie, Y., Nguyen, N. H., Sinthong, P., & Kalagnanam, J. (2023).
[A time series is worth 64 words: Long-term forecasting with transformers](https://openreview.net/forum?id=Jbdc0vTOcol).
*ICLR 2023*
[^19]: Woo, G., Liu, C., Sahoo, D., Kumar, A., & Hoi, S. (2023).
[ETSformer: Exponential Smoothing Transformers for Time-series Forecasting](https://openreview.net/forum?id=5m_3whfo483).
*ICLR 2023*
[^20]: Zhou, T., Ma, Z., Wen, Q., Wang, X., Sun, L., & Jin, R. (2022).
[FEDformer: Frequency enhanced decomposed transformer for long-term series forecasting](https://proceedings.mlr.press/v162/zhou22g.html).
*ICML 2022*.
[^21]: Zhou, H., Zhang, S., Peng, J., Zhang, S., Li, J., Xiong, H., & Zhang, W. (2021).
[Informer: Beyond efficient transformer for long sequence time-series forecasting](https://ojs.aaai.org/index.php/AAAI/article/view/17325).
*AAAI 2021*.
[^22]: Zhou, T., Ma, Z., Wen, Q., Sun, L., Yao, T., Yin, W., & Jin, R. (2022).
[FiLM: Frequency improved Legendre Memory Model for Long-term Time Series Forecasting](https://proceedings.neurips.cc/paper_files/paper/2022/hash/524ef58c2bd075775861234266e5e020-Abstract-Conference.html).
*NeurIPS 2022*.
[^23]: Yi, K., Zhang, Q., Fan, W., Wang, S., Wang, P., He, H., An, N., Lian, D., Cao, L., & Niu, Z. (2023).
[Frequency-domain MLPs are More Effective Learners in Time Series Forecasting](https://proceedings.neurips.cc/paper_files/paper/2023/hash/f1d16af76939f476b5f040fd1398c0a3-Abstract-Conference.html).
*NeurIPS 2023*.
[^24]: Liu, Y., Hu, T., Zhang, H., Wu, H., Wang, S., Ma, L., & Long, M. (2024).
[iTransformer: Inverted Transformers Are Effective for Time Series Forecasting](https://openreview.net/forum?id=JePfAI8fah).
*ICLR 2024*.
[^25]: Liu, Y., Wu, H., Wang, J., & Long, M. (2022).
[Non-stationary Transformers: Exploring the Stationarity in Time Series Forecasting](https://proceedings.neurips.cc/paper_files/paper/2022/hash/4054556fcaa934b0bf76da52cf4f92cb-Abstract-Conference.html).
*NeurIPS 2022*.
[^26]: Liu, S., Yu, H., Liao, C., Li, J., Lin, W., Liu, A. X., & Dustdar, S. (2022).
[Pyraformer: Low-Complexity Pyramidal Attention for Long-Range Time Series Modeling and Forecasting](https://openreview.net/forum?id=0EXmFzUn5I).
*ICLR 2022*.
[^27]: Wang, H., Peng, J., Huang, F., Wang, J., Chen, J., & Xiao, Y. (2023).
[MICN: Multi-scale Local and Global Context Modeling for Long-term Series Forecasting](https://openreview.net/forum?id=zt53IDUR1U).
*ICLR 2023*.
[^28]: Das, A., Kong, W., Leach, A., Mathur, S., Sen, R., & Yu, R. (2023).
[Long-term Forecasting with TiDE: Time-series Dense Encoder](https://openreview.net/forum?id=pCbC3aQB5W).
*TMLR 2023*.
[^29]: Liu, Y., Li, C., Wang, J., & Long, M. (2023).
[Koopa: Learning Non-stationary Time Series Dynamics with Koopman Predictors](https://proceedings.neurips.cc/paper_files/paper/2023/hash/28b3dc0970fa4624a63278a4268de997-Abstract-Conference.html).
*NeurIPS 2023*.
[^30]: Liu, M., Zeng, A., Chen, M., Xu, Z., Lai, Q., Ma, L., & Xu, Q. (2022).
[SCINet: Time Series Modeling and Forecasting with Sample Convolution and Interaction](https://proceedings.neurips.cc/paper_files/paper/2022/hash/266983d0949aed78a16fa4782237dea7-Abstract-Conference.html).
*NeurIPS 2022*.
[^31]: Kim, T., Kim, J., Tae, Y., Park, C., Choi, J. H., & Choo, J. (2022).
[Reversible Instance Normalization for Accurate Time-Series Forecasting against Distribution Shift](https://openreview.net/forum?id=cGDAkQo1C0p).
*ICLR 2022*.
[^32]: Kitaev, N., Kaiser, Ł., & Levskaya, A. (2020).
[Reformer: The Efficient Transformer](https://openreview.net/forum?id=rkgNKkHtvB).
*ICLR 2020*.
[^33]: Cao, D., Wang, Y., Duan, J., Zhang, C., Zhu, X., Huang, C., Tong, Y., Xu, B., Bai, J., Tong, J., & Zhang, Q. (
2020).
[Spectral Temporal Graph Neural Network for Multivariate Time-series Forecasting](https://proceedings.neurips.cc/paper/2020/hash/cdf6581cb7aca4b7e19ef136c6e601a5-Abstract.html).
*NeurIPS 2020*.
[^34]: Nie, T., Qin, G., Mei, Y., & Sun, J. (2024).
[ImputeFormer: Low Rankness-Induced Transformers for Generalizable Spatiotemporal Imputation](https://arxiv.org/abs/2312.01728).
*KDD 2024*.
[^35]: Bai, S., Kolter, J. Z., & Koltun, V. (2018).
[An empirical evaluation of generic convolutional and recurrent networks for sequence modeling](https://arxiv.org/abs/1803.01271).
*arXiv 2018*.
[^36]: Project Gungnir, the world 1st LLM for time-series multitask modeling, will meet you soon. 🚀 Missing values and
variable lengths in your datasets?
Hard to perform multitask learning with your time series? Not problems no longer. We'll open application for public beta
test recently ;-) Follow us, and stay tuned!
<a href="https://time-series.ai"><img src="http://time-series.ai/static/figs/robot.svg" width="20px">
Time-Series.AI</a>
[^37]: Wang, S., Wu, H., Shi, X., Hu, T., Luo, H., Ma, L., ... & ZHOU, J. (2024).
[TimeMixer: Decomposable Multiscale Mixing for Time Series Forecasting](https://openreview.net/forum?id=7oLshfEIC2).
*ICLR 2024*.
[^38]: Luo, D., & Wang X. (2024).
[ModernTCN: A Modern Pure Convolution Structure for General Time Series Analysis](https://openreview.net/forum?id=vpJMJerXHU).
*ICLR 2024*.
[^39]: Zhan, T., He, Y., Li, Z., & Deng, Y. (2024).
[Time Evidence Fusion Network: Multi-source View in Long-Term Time Series Forecasting](https://arxiv.org/abs/2405.06419).
*arXiv 2024*.
[^40]: [Wikipedia: Linear interpolation](https://en.wikipedia.org/wiki/Linear_interpolation)

            

Raw data

            {
    "_id": null,
    "home_page": null,
    "name": "pypots",
    "maintainer": null,
    "docs_url": null,
    "requires_python": ">=3.8",
    "maintainer_email": null,
    "keywords": "data science, data mining, neural networks, machine learning, deep learning, artificial intelligence, time-series analysis, time series, imputation, interpolation, classification, clustering, forecasting, partially observed, irregular sampled, partially-observed time series, incomplete time series, missing data, missing values",
    "author": null,
    "author_email": "Wenjie Du <wenjay.du@gmail.com>",
    "download_url": "https://files.pythonhosted.org/packages/5a/aa/a5d57881ffdd959bbfb1d167bc9353076b875af5f9f67362f95e915bbec3/pypots-0.8.1.tar.gz",
    "platform": null,
    "description": "<a href=\"https://github.com/WenjieDu/PyPOTS\">\n    <img src=\"https://pypots.com/figs/pypots_logos/PyPOTS/logo_FFBG.svg\" width=\"200\" align=\"right\">\n</a>\n\n<h3 align=\"center\">Welcome to PyPOTS</h3>\n\n<p align=\"center\"><i>a Python toolbox for machine learning on Partially-Observed Time Series</i></p>\n\n<p align=\"center\">\n    <a href=\"https://docs.pypots.com/en/latest/install.html#reasons-of-version-limitations-on-dependencies\">\n       <img alt=\"Python version\" src=\"https://img.shields.io/badge/Python-v3.8+-E97040?logo=python&logoColor=white\">\n    </a>\n    <a href=\"https://www.google.com/search?q=%22PyPOTS%22+site%3Apytorch.org\">\n        <img alt=\"powered by Pytorch\" src=\"https://img.shields.io/badge/PyTorch-%E2%9D%A4%EF%B8%8F-F8C6B5?logo=pytorch&logoColor=white\">\n    </a>\n    <a href=\"https://github.com/WenjieDu/PyPOTS/releases\">\n        <img alt=\"the latest release version\" src=\"https://img.shields.io/github/v/release/wenjiedu/pypots?color=EE781F&include_prereleases&label=Release&logo=github&logoColor=white\">\n    </a>\n    <a href=\"https://github.com/WenjieDu/PyPOTS/blob/main/LICENSE\">\n        <img alt=\"BSD-3 license\" src=\"https://img.shields.io/badge/License-BSD--3-E9BB41?logo=opensourceinitiative&logoColor=white\">\n    </a>\n    <a href=\"https://github.com/WenjieDu/PyPOTS#-community\">\n        <img alt=\"Community\" src=\"https://img.shields.io/badge/join_us-community!-C8A062\">\n    </a>\n    <a href=\"https://github.com/WenjieDu/PyPOTS/graphs/contributors\">\n        <img alt=\"GitHub contributors\" src=\"https://img.shields.io/github/contributors/wenjiedu/pypots?color=D8E699&label=Contributors&logo=GitHub\">\n    </a>\n    <a href=\"https://star-history.com/#wenjiedu/pypots\">\n        <img alt=\"GitHub Repo stars\" src=\"https://img.shields.io/github/stars/wenjiedu/pypots?logo=None&color=6BB392&label=%E2%98%85%20Stars\">\n    </a>\n    <a href=\"https://github.com/WenjieDu/PyPOTS/network/members\">\n        <img alt=\"GitHub Repo forks\" src=\"https://img.shields.io/github/forks/wenjiedu/pypots?logo=forgejo&logoColor=black&label=Forks\">\n    </a>\n    <a href=\"https://codeclimate.com/github/WenjieDu/PyPOTS\">\n        <img alt=\"Code Climate maintainability\" src=\"https://img.shields.io/codeclimate/maintainability-percentage/WenjieDu/PyPOTS?color=3C7699&label=Maintainability&logo=codeclimate\">\n    </a>\n    <a href=\"https://coveralls.io/github/WenjieDu/PyPOTS\">\n        <img alt=\"Coveralls coverage\" src=\"https://img.shields.io/coverallsCoverage/github/WenjieDu/PyPOTS?branch=main&logo=coveralls&color=75C1C4&label=Coverage\">\n    </a>\n    <a href=\"https://github.com/WenjieDu/PyPOTS/actions/workflows/testing_ci.yml\">\n        <img alt=\"GitHub Testing\" src=\"https://img.shields.io/github/actions/workflow/status/wenjiedu/pypots/testing_ci.yml?logo=circleci&color=C8D8E1&label=CI\">\n    </a>\n    <a href=\"https://docs.pypots.com\">\n        <img alt=\"Docs building\" src=\"https://img.shields.io/readthedocs/pypots?logo=readthedocs&label=Docs&logoColor=white&color=395260\">\n    </a>\n    <a href=\"https://anaconda.org/conda-forge/pypots\">\n        <img alt=\"Conda downloads\" src=\"https://img.shields.io/endpoint?url=https://pypots.com/figs/downloads_badges/conda_pypots_downloads.json\">\n    </a>\n    <a href=\"https://pepy.tech/project/pypots\">\n        <img alt=\"PyPI downloads\" src=\"https://img.shields.io/endpoint?url=https://pypots.com/figs/downloads_badges/pypi_pypots_downloads.json\">\n    </a>\n    <a href=\"https://arxiv.org/abs/2305.18811\">\n        <img alt=\"arXiv DOI\" src=\"https://img.shields.io/badge/DOI-10.48550/arXiv.2305.18811-F8F7F0\">\n    </a>\n    <a href=\"https://github.com/WenjieDu/PyPOTS/blob/main/README_zh.md\">\n        <img alt=\"README in Chinese\" src=\"https://pypots.com/figs/pypots_logos/readme/CN.svg\">\n    </a>\n   <a href=\"https://github.com/WenjieDu/PyPOTS/blob/main/README.md\">\n        <img alt=\"README in English\" src=\"https://pypots.com/figs/pypots_logos/readme/US.svg\">\n    </a>\n    <a href=\"https://github.com/WenjieDu/PyPOTS\">\n        <img alt=\"PyPOTS Hits\" src=\"https://hits.seeyoufarm.com/api/count/incr/badge.svg?url=https%3A%2F%2Fgithub.com%2FPyPOTS%2FPyPOTS&count_bg=%23009A0A&title_bg=%23555555&icon=&icon_color=%23E7E7E7&title=Hits&edge_flat=false\">\n    </a>\n</p>\n\n\u29bf `Motivation`: Due to all kinds of reasons like failure of collection sensors, communication error,\nand unexpected malfunction, missing values are common to see in time series from the real-world environment.\nThis makes partially-observed time series (POTS) a pervasive problem in open-world modeling and prevents advanced\ndata analysis. Although this problem is important, the area of machine learning on POTS still lacks a dedicated toolkit.\nPyPOTS is created to fill in this blank.\n\n\u29bf `Mission`: PyPOTS (pronounced \"Pie Pots\") is born to become a handy toolbox that is going to make machine learning on\nPOTS easy rather than tedious, to help engineers and researchers focus more on the core problems in their hands rather\nthan on how to deal with the missing parts in their data. PyPOTS will keep integrating classical and the latest\nstate-of-the-art machine learning algorithms for partially-observed multivariate time series. For sure, besides various\nalgorithms, PyPOTS is going to have unified APIs together with detailed documentation and interactive examples across\nalgorithms as tutorials.\n\n\ud83e\udd17 **Please** star this repo to help others notice PyPOTS if you think it is a useful toolkit.\n**Please** kindly [cite PyPOTS](https://github.com/WenjieDu/PyPOTS#-citing-pypots) in your publications if it helps with\nyour research.\nThis really means a lot to our open-source research. Thank you!\n\nThe rest of this readme file is organized as follows:\n[**\u2756 Available Algorithms**](#-available-algorithms),\n[**\u2756 PyPOTS Ecosystem**](#-pypots-ecosystem),\n[**\u2756 Installation**](#-installation),\n[**\u2756 Usage**](#-usage),\n[**\u2756 Citing PyPOTS**](#-citing-pypots),\n[**\u2756 Contribution**](#-contribution),\n[**\u2756 Community**](#-community).\n\n## \u2756 Available Algorithms\n\nPyPOTS supports imputation, classification, clustering, forecasting, and anomaly detection tasks on multivariate\npartially-observed time series with missing values. The table below shows the availability of each algorithm\n(sorted by Year) in PyPOTS for different tasks. The symbol `\u2705` indicates the algorithm is available for the\ncorresponding task (note that models will be continuously updated in the future to handle tasks that are not\ncurrently supported. Stay tuned\u2757\ufe0f).\n\n\ud83c\udf1f Since **v0.2**, all neural-network models in PyPOTS has got hyperparameter-optimization support.\nThis functionality is implemented with the [Microsoft NNI](https://github.com/microsoft/nni) framework. You may want to\nrefer to our time-series imputation survey repo [Awesome_Imputation](https://github.com/WenjieDu/Awesome_Imputation)\nto see how to config and tune the hyperparameters.\n\n\ud83d\udd25 Note that all models whose name with `\ud83e\uddd1\u200d\ud83d\udd27` in the table (e.g. Transformer, iTransformer, Informer etc.) are not\noriginally proposed as algorithms for POTS data in their papers, and they cannot directly accept time series with\nmissing values as input, let alone imputation. **To make them applicable to POTS data, we specifically apply the\nembedding strategy and training approach (ORT+MIT) the same as we did in\n[the SAITS paper](https://arxiv.org/pdf/2202.08516)[^1].**\n\nThe task types are abbreviated as follows:\n**`IMPU`**: Imputation;\n**`FORE`**: Forecasting;\n**`CLAS`**: Classification;\n**`CLUS`**: Clustering;\n**`ANOD`**: Anomaly Detection.\nThe paper references and links are all listed at the bottom of this file.\n\n| **Type**      | **Algo**                                                                                                                         | **IMPU** | **FORE** | **CLAS** | **CLUS** | **ANOD** | **Year - Venue**   |\n|:--------------|:---------------------------------------------------------------------------------------------------------------------------------|:--------:|:--------:|:--------:|:--------:|:--------:|:-------------------|\n| LLM           | <a href=\"https://time-series.ai\"><img src=\"https://time-series.ai/static/figs/robot.svg\" width=\"26px\"> Time-Series.AI</a>  [^36] |    \u2705     |    \u2705     |    \u2705     |    \u2705     |    \u2705     | `Later in 2024`    |\n| Neural Net    | TEFN\ud83e\uddd1\u200d\ud83d\udd27[^39]                                                                                                                   |    \u2705     |          |          |          |          | `2024 - arXiv`     |\n| Neural Net    | TimeMixer[^37]                                                                                                                   |    \u2705     |          |          |          |          | `2024 - ICLR`      |\n| Neural Net    | iTransformer\ud83e\uddd1\u200d\ud83d\udd27[^24]                                                                                                           |    \u2705     |          |          |          |          | `2024 - ICLR`      |\n| Neural Net    | ModernTCN[^38]                                                                                                                   |    \u2705     |          |          |          |          | `2024 - ICLR`      |\n| Neural Net    | ImputeFormer\ud83e\uddd1\u200d\ud83d\udd27[^34]                                                                                                           |    \u2705     |          |          |          |          | `2024 - KDD`       |\n| Neural Net    | SAITS[^1]                                                                                                                        |    \u2705     |          |          |          |          | `2023 - ESWA`      |\n| Neural Net    | FreTS\ud83e\uddd1\u200d\ud83d\udd27[^23]                                                                                                                  |    \u2705     |          |          |          |          | `2023 - NeurIPS`   |\n| Neural Net    | Koopa\ud83e\uddd1\u200d\ud83d\udd27[^29]                                                                                                                  |    \u2705     |          |          |          |          | `2023 - NeurIPS`   |\n| Neural Net    | Crossformer\ud83e\uddd1\u200d\ud83d\udd27[^16]                                                                                                            |    \u2705     |          |          |          |          | `2023 - ICLR`      |\n| Neural Net    | TimesNet[^14]                                                                                                                    |    \u2705     |          |          |          |          | `2023 - ICLR`      |\n| Neural Net    | PatchTST\ud83e\uddd1\u200d\ud83d\udd27[^18]                                                                                                               |    \u2705     |          |          |          |          | `2023 - ICLR`      |\n| Neural Net    | ETSformer\ud83e\uddd1\u200d\ud83d\udd27[^19]                                                                                                              |    \u2705     |          |          |          |          | `2023 - ICLR`      |\n| Neural Net    | MICN\ud83e\uddd1\u200d\ud83d\udd27[^27]                                                                                                                   |    \u2705     |          |          |          |          | `2023 - ICLR`      |\n| Neural Net    | DLinear\ud83e\uddd1\u200d\ud83d\udd27[^17]                                                                                                                |    \u2705     |          |          |          |          | `2023 - AAAI`      |\n| Neural Net    | TiDE\ud83e\uddd1\u200d\ud83d\udd27[^28]                                                                                                                   |    \u2705     |          |          |          |          | `2023 - TMLR`      |\n| Neural Net    | SCINet\ud83e\uddd1\u200d\ud83d\udd27[^30]                                                                                                                 |    \u2705     |          |          |          |          | `2022 - NeurIPS`   |\n| Neural Net    | Nonstationary Tr.\ud83e\uddd1\u200d\ud83d\udd27[^25]                                                                                                      |    \u2705     |          |          |          |          | `2022 - NeurIPS`   |\n| Neural Net    | FiLM\ud83e\uddd1\u200d\ud83d\udd27[^22]                                                                                                                   |    \u2705     |          |          |          |          | `2022 - NeurIPS`   |\n| Neural Net    | RevIN_SCINet\ud83e\uddd1\u200d\ud83d\udd27[^31]                                                                                                           |    \u2705     |          |          |          |          | `2022 - ICLR`      |\n| Neural Net    | Pyraformer\ud83e\uddd1\u200d\ud83d\udd27[^26]                                                                                                             |    \u2705     |          |          |          |          | `2022 - ICLR`      |\n| Neural Net    | Raindrop[^5]                                                                                                                     |          |          |    \u2705     |          |          | `2022 - ICLR`      |\n| Neural Net    | FEDformer\ud83e\uddd1\u200d\ud83d\udd27[^20]                                                                                                              |    \u2705     |          |          |          |          | `2022 - ICML`      |\n| Neural Net    | Autoformer\ud83e\uddd1\u200d\ud83d\udd27[^15]                                                                                                             |    \u2705     |          |          |          |          | `2021 - NeurIPS`   |\n| Neural Net    | CSDI[^12]                                                                                                                        |    \u2705     |    \u2705     |          |          |          | `2021 - NeurIPS`   |\n| Neural Net    | Informer\ud83e\uddd1\u200d\ud83d\udd27[^21]                                                                                                               |    \u2705     |          |          |          |          | `2021 - AAAI`      |\n| Neural Net    | US-GAN[^10]                                                                                                                      |    \u2705     |          |          |          |          | `2021 - AAAI`      |\n| Neural Net    | CRLI[^6]                                                                                                                         |          |          |          |    \u2705     |          | `2021 - AAAI`      |\n| Probabilistic | BTTF[^8]                                                                                                                         |          |    \u2705     |          |          |          | `2021 - TPAMI`     |\n| Neural Net    | StemGNN\ud83e\uddd1\u200d\ud83d\udd27[^33]                                                                                                                |    \u2705     |          |          |          |          | `2020 - NeurIPS`   |\n| Neural Net    | Reformer\ud83e\uddd1\u200d\ud83d\udd27[^32]                                                                                                               |    \u2705     |          |          |          |          | `2020 - ICLR`      |\n| Neural Net    | GP-VAE[^11]                                                                                                                      |    \u2705     |          |          |          |          | `2020 - AISTATS`   |\n| Neural Net    | VaDER[^7]                                                                                                                        |          |          |          |    \u2705     |          | `2019 - GigaSci.`  |\n| Neural Net    | M-RNN[^9]                                                                                                                        |    \u2705     |          |          |          |          | `2019 - TBME`      |\n| Neural Net    | BRITS[^3]                                                                                                                        |    \u2705     |          |    \u2705     |          |          | `2018 - NeurIPS`   |\n| Neural Net    | GRU-D[^4]                                                                                                                        |    \u2705     |          |    \u2705     |          |          | `2018 - Sci. Rep.` |\n| Neural Net    | TCN\ud83e\uddd1\u200d\ud83d\udd27[^35]                                                                                                                    |    \u2705     |          |          |          |          | `2018 - arXiv`     |\n| Neural Net    | Transformer\ud83e\uddd1\u200d\ud83d\udd27[^2]                                                                                                             |    \u2705     |          |          |          |          | `2017 - NeurIPS`   |\n| Naive         | Lerp[^40]                                                                                                                        |    \u2705     |          |          |          |          |                    |\n| Naive         | LOCF/NOCB                                                                                                                        |    \u2705     |          |          |          |          |                    |\n| Naive         | Mean                                                                                                                             |    \u2705     |          |          |          |          |                    |\n| Naive         | Median                                                                                                                           |    \u2705     |          |          |          |          |                    |\n\n\ud83d\udcaf Contribute your model right now to increase your research impact! PyPOTS downloads are increasing rapidly\n(**[300K+ in total and 1K+ daily on PyPI so far](https://www.pepy.tech/projects/pypots)**),\nand your work will be widely used and cited by the community.\nRefer to the [contribution guide](https://github.com/WenjieDu/PyPOTS#-contribution) to see how to include your model in\nPyPOTS.\n\n## \u2756 PyPOTS Ecosystem\n\nAt PyPOTS, things are related to coffee, which we're familiar with. Yes, this is a coffee universe!\nAs you can see, there is a coffee pot in the PyPOTS logo. And what else? Please read on ;-)\n\n<a href=\"https://github.com/WenjieDu/TSDB\">\n    <img src=\"https://pypots.com/figs/pypots_logos/TSDB/logo_FFBG.svg\" align=\"left\" width=\"140\" alt=\"TSDB logo\"/>\n</a>\n\n\ud83d\udc48 Time series datasets are taken as coffee beans at PyPOTS, and POTS datasets are incomplete coffee beans with missing\nparts that have their own meanings. To make various public time-series datasets readily available to users,\n<i>Time Series Data Beans (TSDB)</i> is created to make loading time-series datasets super easy!\nVisit [TSDB](https://github.com/WenjieDu/TSDB) right now to know more about this handy tool \ud83d\udee0, and it now supports a\ntotal of 172 open-source datasets!\n\n<a href=\"https://github.com/WenjieDu/PyGrinder\">\n    <img src=\"https://pypots.com/figs/pypots_logos/PyGrinder/logo_FFBG.svg\" align=\"right\" width=\"140\" alt=\"PyGrinder logo\"/>\n</a>\n\n\ud83d\udc49 To simulate the real-world data beans with missingness, the ecosystem library\n[PyGrinder](https://github.com/WenjieDu/PyGrinder), a toolkit helping grind your coffee beans into incomplete ones, is\ncreated. Missing patterns fall into three categories according to Robin's theory[^13]:\nMCAR (missing completely at random), MAR (missing at random), and MNAR (missing not at random).\nPyGrinder supports all of them and additional functionalities related to missingness.\nWith PyGrinder, you can introduce synthetic missing values into your datasets with a single line of code.\n\n<a href=\"https://github.com/WenjieDu/BenchPOTS\">\n    <img src=\"https://pypots.com/figs/pypots_logos/BenchPOTS/logo_FFBG.svg\" align=\"left\" width=\"140\" alt=\"BenchPOTS logo\"/>\n</a>\n\n\ud83d\udc48 To fairly evaluate the performance of PyPOTS algorithms, the benchmarking suite\n[BenchPOTS](https://github.com/WenjieDu/BenchPOTS) is created, which provides standard and unified data-preprocessing\npipelines to prepare datasets for measuring the performance of different POTS algorithms on various tasks.\n\n<a href=\"https://github.com/WenjieDu/BrewPOTS\">\n    <img src=\"https://pypots.com/figs/pypots_logos/BrewPOTS/logo_FFBG.svg\" align=\"right\" width=\"140\" alt=\"BrewPOTS logo\"/>\n</a>\n\n\ud83d\udc49 Now the beans, grinder, and pot are ready, please have a seat on the bench and let's think about how to brew us a cup\nof coffee. Tutorials are necessary! Considering the future workload, PyPOTS tutorials are released in a single repo,\nand you can find them in [BrewPOTS](https://github.com/WenjieDu/BrewPOTS).\nTake a look at it now, and learn how to brew your POTS datasets!\n\n<p align=\"center\">\n<a href=\"https://pypots.com/ecosystem/\">\n    <img src=\"https://pypots.com/figs/pypots_logos/Ecosystem/PyPOTS_Ecosystem_Pipeline.png\" width=\"95%\"/>\n</a>\n<br>\n<b> \u2615\ufe0f Welcome to the universe of PyPOTS. Enjoy it and have fun!</b>\n</p>\n\n## \u2756 Installation\n\nYou can refer to [the installation instruction](https://docs.pypots.com/en/latest/install.html) in PyPOTS documentation\nfor a guideline with more details.\n\nPyPOTS is available on both [PyPI](https://pypi.python.org/pypi/pypots)\nand [Anaconda](https://anaconda.org/conda-forge/pypots).\nYou can install PyPOTS like below as well as\n[TSDB](https://github.com/WenjieDu/TSDB),[PyGrinder](https://github.com/WenjieDu/PyGrinder),\n[BenchPOTS](https://github.com/WenjieDu/BenchPOTS), and [AI4TS](https://github.com/WenjieDu/AI4TS):\n\n``` bash\n# via pip\npip install pypots            # the first time installation\npip install pypots --upgrade  # update pypots to the latest version\n# install from the latest source code with the latest features but may be not officially released yet\npip install https://github.com/WenjieDu/PyPOTS/archive/main.zip\n\n# via conda\nconda install conda-forge::pypots  # the first time installation\nconda update  conda-forge::pypots  # update pypots to the latest version\n```\n\n## \u2756 Usage\n\nBesides [BrewPOTS](https://github.com/WenjieDu/BrewPOTS), you can also find a simple and quick-start tutorial notebook\non Google Colab\n<a href=\"https://colab.research.google.com/drive/1HEFjylEy05-r47jRy0H9jiS_WhD0UWmQ\">\n<img src=\"https://img.shields.io/badge/GoogleColab-PyPOTS_Tutorials-F9AB00?logo=googlecolab&logoColor=white\" alt=\"Colab tutorials\" align=\"center\"/>\n</a>. If you have further questions, please refer to PyPOTS documentation [docs.pypots.com](https://docs.pypots.com).\nYou can also [raise an issue](https://github.com/WenjieDu/PyPOTS/issues) or [ask in our community](#-community).\n\nWe present you a usage example of imputing missing values in time series with PyPOTS below, you can click it to view.\n\n<details open>\n<summary><b>Click here to see an example applying SAITS on PhysioNet2012 for imputation:</b></summary>\n\n``` python\n# Data preprocessing. Tedious, but PyPOTS can help.\nimport numpy as np\nfrom sklearn.preprocessing import StandardScaler\nfrom pygrinder import mcar\nfrom pypots.data import load_specific_dataset\ndata = load_specific_dataset('physionet_2012')  # PyPOTS will automatically download and extract it.\nX = data['X']\nnum_samples = len(X['RecordID'].unique())\nX = X.drop(['RecordID', 'Time'], axis = 1)\nX = StandardScaler().fit_transform(X.to_numpy())\nX = X.reshape(num_samples, 48, -1)\nX_ori = X  # keep X_ori for validation\nX = mcar(X, 0.1)  # randomly hold out 10% observed values as ground truth\ndataset = {\"X\": X}  # X for model input\nprint(X.shape)  # (11988, 48, 37), 11988 samples and each sample has 48 time steps, 37 features\n\n# Model training. This is PyPOTS showtime.\nfrom pypots.imputation import SAITS\nfrom pypots.utils.metrics import calc_mae\nsaits = SAITS(n_steps=48, n_features=37, n_layers=2, d_model=256, n_heads=4, d_k=64, d_v=64, d_ffn=128, dropout=0.1, epochs=10)\n# Here I use the whole dataset as the training set because ground truth is not visible to the model, you can also split it into train/val/test sets\nsaits.fit(dataset)  # train the model on the dataset\nimputation = saits.impute(dataset)  # impute the originally-missing values and artificially-missing values\nindicating_mask = np.isnan(X) ^ np.isnan(X_ori)  # indicating mask for imputation error calculation\nmae = calc_mae(imputation, np.nan_to_num(X_ori), indicating_mask)  # calculate mean absolute error on the ground truth (artificially-missing values)\nsaits.save(\"save_it_here/saits_physionet2012.pypots\")  # save the model for future use\nsaits.load(\"save_it_here/saits_physionet2012.pypots\")  # reload the serialized model file for following imputation or training\n```\n\n</details>\n\n## \u2756 Citing PyPOTS\n\n> [!TIP]\n> **[Updates in Jun 2024]** \ud83d\ude0e The 1st comprehensive time-seres imputation benchmark paper\n[TSI-Bench: Benchmarking Time Series Imputation](https://arxiv.org/abs/2406.12747) now is public available.\n> The code is open source in the repo [Awesome_Imputation](https://github.com/WenjieDu/Awesome_Imputation).\n> With nearly 35,000 experiments, we provide a comprehensive benchmarking study on 28 imputation methods, 3 missing\n> patterns (points, sequences, blocks),\n> various missing rates, and 8 real-world datasets.\n>\n> **[Updates in Feb 2024]** \ud83c\udf89 Our survey\n> paper [Deep Learning for Multivariate Time Series Imputation: A Survey](https://arxiv.org/abs/2402.04059) has been\n> released on arXiv.\n> We comprehensively review the literature of the state-of-the-art deep-learning imputation methods for time series,\n> provide a taxonomy for them, and discuss the challenges and future directions in this field.\n\nThe paper introducing PyPOTS is available [on arXiv](https://arxiv.org/abs/2305.18811),\nand a short version of it is accepted by the 9th SIGKDD international workshop on Mining and Learning from Time\nSeries ([MiLeTS'23](https://kdd-milets.github.io/milets2023/))).\n**Additionally**, PyPOTS has been included as a [PyTorch Ecosystem](https://pytorch.org/ecosystem/) project.\nWe are pursuing to publish it in prestigious academic venues, e.g. JMLR (track for\n[Machine Learning Open Source Software](https://www.jmlr.org/mloss/)). If you use PyPOTS in your work,\nplease cite it as below and \ud83c\udf1fstar this repository to make others notice this library. \ud83e\udd17\n\nThere are scientific research projects using PyPOTS and referencing in their papers.\nHere is [an incomplete list of them](https://scholar.google.com/scholar?as_ylo=2022&q=%E2%80%9CPyPOTS%E2%80%9D&hl=en).\n\n```bibtex\n@article{du2023pypots,\n    title = {{PyPOTS: a Python toolbox for data mining on Partially-Observed Time Series}},\n    author = {Wenjie Du},\n    journal = {arXiv preprint arXiv:2305.18811},\n    year = {2023},\n}\n```\n\nor\n> Wenjie Du.\n> PyPOTS: a Python toolbox for data mining on Partially-Observed Time Series.\n> arXiv, abs/2305.18811, 2023.\n\n## \u2756 Contribution\n\nYou're very welcome to contribute to this exciting project!\n\nBy committing your code, you'll\n\n1. make your well-established model out-of-the-box for PyPOTS users to run,\n   and help your work obtain more exposure and impact.\n   Take a look at our [inclusion criteria](https://docs.pypots.com/en/latest/faq.html#inclusion-criteria).\n   You can utilize the `template` folder in each task package (e.g.\n   [pypots/imputation/template](https://github.com/WenjieDu/PyPOTS/tree/main/pypots/imputation/template)) to quickly\n   start;\n2. become one of [PyPOTS contributors](https://github.com/WenjieDu/PyPOTS/graphs/contributors) and\n   be listed as a volunteer developer [on the PyPOTS website](https://pypots.com/about/#volunteer-developers);\n3. get mentioned in PyPOTS [release notes](https://github.com/WenjieDu/PyPOTS/releases);\n\nYou can also contribute to PyPOTS by simply staring\ud83c\udf1f this repo to help more people notice it.\nYour star is your recognition to PyPOTS, and it matters!\n\n<details open>\n<summary>\n    <b><i>\n    \ud83d\udc4f Click here to view PyPOTS stargazers and forkers.<br>\n    We're so proud to have more and more awesome users, as well as more bright \u2728stars:\n    </i></b>\n</summary>\n<a href=\"https://github.com/WenjieDu/PyPOTS/stargazers\">\n    <img alt=\"PyPOTS stargazers\" src=\"http://reporoster.com/stars/dark/WenjieDu/PyPOTS\">\n</a>\n<br>\n<a href=\"https://github.com/WenjieDu/PyPOTS/network/members\">\n    <img alt=\"PyPOTS forkers\" src=\"http://reporoster.com/forks/dark/WenjieDu/PyPOTS\">\n</a>\n</details>\n\n\ud83d\udc40 Check out a full list of our users' affiliations [on PyPOTS website here](https://pypots.com/users/)!\n\n## \u2756 Community\n\nWe care about the feedback from our users, so we're building PyPOTS community on\n\n- [Slack](https://join.slack.com/t/pypots-org/shared_invite/zt-1gq6ufwsi-p0OZdW~e9UW_IA4_f1OfxA). General discussion,\n  Q&A, and our development team are here;\n- [LinkedIn](https://www.linkedin.com/company/pypots). Official announcements and news are here;\n- [WeChat (\u5fae\u4fe1\u516c\u4f17\u53f7)](https://mp.weixin.qq.com/s/X3ukIgL1QpNH8ZEXq1YifA). We also run a group chat on WeChat,\n  and you can get the QR code from the official account after following it;\n\nIf you have any suggestions or want to contribute ideas or share time-series related papers, join us and tell.\nPyPOTS community is open, transparent, and surely friendly. Let's work together to build and improve PyPOTS!\n\n\n[//]: # (Use APA reference style below)\n[^1]: Du, W., Cote, D., & Liu, Y. (2023).\n[SAITS: Self-Attention-based Imputation for Time Series](https://doi.org/10.1016/j.eswa.2023.119619).\n*Expert systems with applications*.\n[^2]: Vaswani, A., Shazeer, N.M., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., & Polosukhin, I. (\n2017).\n[Attention is All you Need](https://papers.nips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html).\n*NeurIPS 2017*.\n[^3]: Cao, W., Wang, D., Li, J., Zhou, H., Li, L., & Li, Y. (2018).\n[BRITS: Bidirectional Recurrent Imputation for Time Series](https://papers.nips.cc/paper/2018/hash/734e6bfcd358e25ac1db0a4241b95651-Abstract.html).\n*NeurIPS 2018*.\n[^4]: Che, Z., Purushotham, S., Cho, K., Sontag, D.A., & Liu, Y. (2018).\n[Recurrent Neural Networks for Multivariate Time Series with Missing Values](https://www.nature.com/articles/s41598-018-24271-9).\n*Scientific Reports*.\n[^5]: Zhang, X., Zeman, M., Tsiligkaridis, T., & Zitnik, M. (2022).\n[Graph-Guided Network for Irregularly Sampled Multivariate Time Series](https://arxiv.org/abs/2110.05357).\n*ICLR 2022*.\n[^6]: Ma, Q., Chen, C., Li, S., & Cottrell, G. W. (2021).\n[Learning Representations for Incomplete Time Series Clustering](https://ojs.aaai.org/index.php/AAAI/article/view/17070).\n*AAAI 2021*.\n[^7]: Jong, J.D., Emon, M.A., Wu, P., Karki, R., Sood, M., Godard, P., Ahmad, A., Vrooman, H.A., Hofmann-Apitius, M., &\nFr\u00f6hlich, H. (2019).\n[Deep learning for clustering of multivariate clinical patient trajectories with missing values](https://academic.oup.com/gigascience/article/8/11/giz134/5626377).\n*GigaScience*.\n[^8]: Chen, X., & Sun, L. (2021).\n[Bayesian Temporal Factorization for Multidimensional Time Series Prediction](https://arxiv.org/abs/1910.06366).\n*IEEE transactions on pattern analysis and machine intelligence*.\n[^9]: Yoon, J., Zame, W. R., & van der Schaar, M. (2019).\n[Estimating Missing Data in Temporal Data Streams Using Multi-Directional Recurrent Neural Networks](https://ieeexplore.ieee.org/document/8485748).\n*IEEE Transactions on Biomedical Engineering*.\n[^10]: Miao, X., Wu, Y., Wang, J., Gao, Y., Mao, X., & Yin, J. (2021).\n[Generative Semi-supervised Learning for Multivariate Time Series Imputation](https://ojs.aaai.org/index.php/AAAI/article/view/17086).\n*AAAI 2021*.\n[^11]: Fortuin, V., Baranchuk, D., Raetsch, G. & Mandt, S. (2020).\n[GP-VAE: Deep Probabilistic Time Series Imputation](https://proceedings.mlr.press/v108/fortuin20a.html).\n*AISTATS 2020*.\n[^12]: Tashiro, Y., Song, J., Song, Y., & Ermon, S. (2021).\n[CSDI: Conditional Score-based Diffusion Models for Probabilistic Time Series Imputation](https://proceedings.neurips.cc/paper/2021/hash/cfe8504bda37b575c70ee1a8276f3486-Abstract.html).\n*NeurIPS 2021*.\n[^13]: Rubin, D. B. (1976).\n[Inference and missing data](https://academic.oup.com/biomet/article-abstract/63/3/581/270932).\n*Biometrika*.\n[^14]: Wu, H., Hu, T., Liu, Y., Zhou, H., Wang, J., & Long, M. (2023).\n[TimesNet: Temporal 2d-variation modeling for general time series analysis](https://openreview.net/forum?id=ju_Uqw384Oq).\n*ICLR 2023*\n[^15]: Wu, H., Xu, J., Wang, J., & Long, M. (2021).\n[Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting](https://proceedings.neurips.cc/paper/2021/hash/bcc0d400288793e8bdcd7c19a8ac0c2b-Abstract.html).\n*NeurIPS 2021*.\n[^16]: Zhang, Y., & Yan, J. (2023).\n[Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting](https://openreview.net/forum?id=vSVLM2j9eie).\n*ICLR 2023*.\n[^17]: Zeng, A., Chen, M., Zhang, L., & Xu, Q. (2023).\n[Are transformers effective for time series forecasting?](https://ojs.aaai.org/index.php/AAAI/article/view/26317).\n*AAAI 2023*\n[^18]: Nie, Y., Nguyen, N. H., Sinthong, P., & Kalagnanam, J. (2023).\n[A time series is worth 64 words: Long-term forecasting with transformers](https://openreview.net/forum?id=Jbdc0vTOcol).\n*ICLR 2023*\n[^19]: Woo, G., Liu, C., Sahoo, D., Kumar, A., & Hoi, S. (2023).\n[ETSformer: Exponential Smoothing Transformers for Time-series Forecasting](https://openreview.net/forum?id=5m_3whfo483).\n*ICLR 2023*\n[^20]: Zhou, T., Ma, Z., Wen, Q., Wang, X., Sun, L., & Jin, R. (2022).\n[FEDformer: Frequency enhanced decomposed transformer for long-term series forecasting](https://proceedings.mlr.press/v162/zhou22g.html).\n*ICML 2022*.\n[^21]: Zhou, H., Zhang, S., Peng, J., Zhang, S., Li, J., Xiong, H., & Zhang, W. (2021).\n[Informer: Beyond efficient transformer for long sequence time-series forecasting](https://ojs.aaai.org/index.php/AAAI/article/view/17325).\n*AAAI 2021*.\n[^22]: Zhou, T., Ma, Z., Wen, Q., Sun, L., Yao, T., Yin, W., & Jin, R. (2022).\n[FiLM: Frequency improved Legendre Memory Model for Long-term Time Series Forecasting](https://proceedings.neurips.cc/paper_files/paper/2022/hash/524ef58c2bd075775861234266e5e020-Abstract-Conference.html).\n*NeurIPS 2022*.\n[^23]: Yi, K., Zhang, Q., Fan, W., Wang, S., Wang, P., He, H., An, N., Lian, D., Cao, L., & Niu, Z. (2023).\n[Frequency-domain MLPs are More Effective Learners in Time Series Forecasting](https://proceedings.neurips.cc/paper_files/paper/2023/hash/f1d16af76939f476b5f040fd1398c0a3-Abstract-Conference.html).\n*NeurIPS 2023*.\n[^24]: Liu, Y., Hu, T., Zhang, H., Wu, H., Wang, S., Ma, L., & Long, M. (2024).\n[iTransformer: Inverted Transformers Are Effective for Time Series Forecasting](https://openreview.net/forum?id=JePfAI8fah).\n*ICLR 2024*.\n[^25]: Liu, Y., Wu, H., Wang, J., & Long, M. (2022).\n[Non-stationary Transformers: Exploring the Stationarity in Time Series Forecasting](https://proceedings.neurips.cc/paper_files/paper/2022/hash/4054556fcaa934b0bf76da52cf4f92cb-Abstract-Conference.html).\n*NeurIPS 2022*.\n[^26]: Liu, S., Yu, H., Liao, C., Li, J., Lin, W., Liu, A. X., & Dustdar, S. (2022).\n[Pyraformer: Low-Complexity Pyramidal Attention for Long-Range Time Series Modeling and Forecasting](https://openreview.net/forum?id=0EXmFzUn5I).\n*ICLR 2022*.\n[^27]: Wang, H., Peng, J., Huang, F., Wang, J., Chen, J., & Xiao, Y. (2023).\n[MICN: Multi-scale Local and Global Context Modeling for Long-term Series Forecasting](https://openreview.net/forum?id=zt53IDUR1U).\n*ICLR 2023*.\n[^28]: Das, A., Kong, W., Leach, A., Mathur, S., Sen, R., & Yu, R. (2023).\n[Long-term Forecasting with TiDE: Time-series Dense Encoder](https://openreview.net/forum?id=pCbC3aQB5W).\n*TMLR 2023*.\n[^29]: Liu, Y., Li, C., Wang, J., & Long, M. (2023).\n[Koopa: Learning Non-stationary Time Series Dynamics with Koopman Predictors](https://proceedings.neurips.cc/paper_files/paper/2023/hash/28b3dc0970fa4624a63278a4268de997-Abstract-Conference.html).\n*NeurIPS 2023*.\n[^30]: Liu, M., Zeng, A., Chen, M., Xu, Z., Lai, Q., Ma, L., & Xu, Q. (2022).\n[SCINet: Time Series Modeling and Forecasting with Sample Convolution and Interaction](https://proceedings.neurips.cc/paper_files/paper/2022/hash/266983d0949aed78a16fa4782237dea7-Abstract-Conference.html).\n*NeurIPS 2022*.\n[^31]: Kim, T., Kim, J., Tae, Y., Park, C., Choi, J. H., & Choo, J. (2022).\n[Reversible Instance Normalization for Accurate Time-Series Forecasting against Distribution Shift](https://openreview.net/forum?id=cGDAkQo1C0p).\n*ICLR 2022*.\n[^32]: Kitaev, N., Kaiser, \u0141., & Levskaya, A. (2020).\n[Reformer: The Efficient Transformer](https://openreview.net/forum?id=rkgNKkHtvB).\n*ICLR 2020*.\n[^33]: Cao, D., Wang, Y., Duan, J., Zhang, C., Zhu, X., Huang, C., Tong, Y., Xu, B., Bai, J., Tong, J., & Zhang, Q. (\n2020).\n[Spectral Temporal Graph Neural Network for Multivariate Time-series Forecasting](https://proceedings.neurips.cc/paper/2020/hash/cdf6581cb7aca4b7e19ef136c6e601a5-Abstract.html).\n*NeurIPS 2020*.\n[^34]: Nie, T., Qin, G., Mei, Y., & Sun, J. (2024).\n[ImputeFormer: Low Rankness-Induced Transformers for Generalizable Spatiotemporal Imputation](https://arxiv.org/abs/2312.01728).\n*KDD 2024*.\n[^35]: Bai, S., Kolter, J. Z., & Koltun, V. (2018).\n[An empirical evaluation of generic convolutional and recurrent networks for sequence modeling](https://arxiv.org/abs/1803.01271).\n*arXiv 2018*.\n[^36]: Project Gungnir, the world 1st LLM for time-series multitask modeling, will meet you soon. \ud83d\ude80 Missing values and\nvariable lengths in your datasets?\nHard to perform multitask learning with your time series? Not problems no longer. We'll open application for public beta\ntest recently ;-) Follow us, and stay tuned!\n<a href=\"https://time-series.ai\"><img src=\"http://time-series.ai/static/figs/robot.svg\" width=\"20px\">\nTime-Series.AI</a>\n[^37]: Wang, S., Wu, H., Shi, X., Hu, T., Luo, H., Ma, L., ... & ZHOU, J. (2024).\n[TimeMixer: Decomposable Multiscale Mixing for Time Series Forecasting](https://openreview.net/forum?id=7oLshfEIC2).\n*ICLR 2024*.\n[^38]: Luo, D., & Wang X. (2024).\n[ModernTCN: A Modern Pure Convolution Structure for General Time Series Analysis](https://openreview.net/forum?id=vpJMJerXHU).\n*ICLR 2024*.\n[^39]: Zhan, T., He, Y., Li, Z., & Deng, Y. (2024).\n[Time Evidence Fusion Network: Multi-source View in Long-Term Time Series Forecasting](https://arxiv.org/abs/2405.06419).\n*arXiv 2024*.\n[^40]: [Wikipedia: Linear interpolation](https://en.wikipedia.org/wiki/Linear_interpolation)\n",
    "bugtrack_url": null,
    "license": "Copyright (c) 2023-present, Wenjie Du All rights reserved.  Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met:  1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer.  2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution.  3. Neither the name of the copyright holder nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission.  THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS \"AS IS\" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. ",
    "summary": "A Python Toolbox for Machine Learning on Partially-Observed Time Series",
    "version": "0.8.1",
    "project_urls": {
        "Bug Tracker": "https://github.com/WenjieDu/PyPOTS/issues",
        "Documentation": "https://docs.pypots.com",
        "Download": "https://github.com/WenjieDu/PyPOTS/archive/main.zip",
        "Homepage": "https://pypots.com",
        "Source": "https://github.com/WenjieDu/PyPOTS"
    },
    "split_keywords": [
        "data science",
        " data mining",
        " neural networks",
        " machine learning",
        " deep learning",
        " artificial intelligence",
        " time-series analysis",
        " time series",
        " imputation",
        " interpolation",
        " classification",
        " clustering",
        " forecasting",
        " partially observed",
        " irregular sampled",
        " partially-observed time series",
        " incomplete time series",
        " missing data",
        " missing values"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "6bf8b655b0ef87ae2ca0c9391e181cff08ed4cac88b9fb22c6e14c3690f7e8f7",
                "md5": "de52e661680ab72229405561c8c9c255",
                "sha256": "05bc32a9bd7dcf3855ba6f17631de32a9e15ceca03f4bc45eed0d0716ce5737f"
            },
            "downloads": -1,
            "filename": "pypots-0.8.1-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "de52e661680ab72229405561c8c9c255",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.8",
            "size": 497620,
            "upload_time": "2024-09-26T05:57:22",
            "upload_time_iso_8601": "2024-09-26T05:57:22.583382Z",
            "url": "https://files.pythonhosted.org/packages/6b/f8/b655b0ef87ae2ca0c9391e181cff08ed4cac88b9fb22c6e14c3690f7e8f7/pypots-0.8.1-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "5aaaa5d57881ffdd959bbfb1d167bc9353076b875af5f9f67362f95e915bbec3",
                "md5": "98bc383011e6f2fedf3ee7c0691aad70",
                "sha256": "e0308699a3d5b30531bc0656c71ddc7297fdc986e1085368f9627c67a343ab82"
            },
            "downloads": -1,
            "filename": "pypots-0.8.1.tar.gz",
            "has_sig": false,
            "md5_digest": "98bc383011e6f2fedf3ee7c0691aad70",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.8",
            "size": 283585,
            "upload_time": "2024-09-26T05:57:24",
            "upload_time_iso_8601": "2024-09-26T05:57:24.524689Z",
            "url": "https://files.pythonhosted.org/packages/5a/aa/a5d57881ffdd959bbfb1d167bc9353076b875af5f9f67362f95e915bbec3/pypots-0.8.1.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-09-26 05:57:24",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "WenjieDu",
    "github_project": "PyPOTS",
    "travis_ci": false,
    "coveralls": true,
    "github_actions": true,
    "lcname": "pypots"
}
        
Elapsed time: 0.72063s