pyaf


Namepyaf JSON
Version 5.0 PyPI version JSON
download
home_pagehttps://github.com/antoinecarme/pyaf
SummaryPython Automatic Forecasting
upload_time2023-07-12 22:24:01
maintainer
docs_urlNone
authorAntoine CARME
requires_python>=3
licenseBSD 3-clause
keywords arx automatic-forecasting autoregressive benchmark cycle decomposition exogenous forecasting heroku hierarchical-forecasting horizon jupyter pandas python scikit-learn seasonal time-series transformation trend web-service
VCS
bugtrack_url
requirements pandas datetime scikit-learn matplotlib pydot flask statsmodels
Travis-CI
coveralls test coverage No coveralls.
            
PyAF (Python Automatic Forecasting)
===================================

[![Github/CI/CD](https://github.com/antoinecarme/pyaf/actions/workflows/python-app.yml/badge.svg)](https://github.com/antoinecarme/pyaf/actions/workflows/python-app.yml)

PyAF is an Open Source Python library for Automatic Forecasting built on top of
popular data science python modules: NumPy, SciPy, Pandas and scikit-learn.

PyAF works as an automated process for predicting future values of a signal
using a machine learning approach. It provides a set of features that is
comparable to some popular commercial automatic forecasting products.

PyAF has been developed, tested and benchmarked using a **python 3.x** version.

PyAF is distributed under the [3-Clause BSD license](https://tldrlegal.com/license/bsd-3-clause-license-%28revised%29).


Demo 
----
```Python
import numpy as np, pandas as pd
import pyaf.ForecastEngine as autof

if __name__ == '__main__':
   # generate a daily signal covering one year 2016 in a pandas dataframe
   N = 360
   df_train = pd.DataFrame({"Date": pd.date_range(start="2016-01-25", periods=N, freq='D'),
   	                    "Signal": (np.arange(N)//40 + np.arange(N) % 21 + np.random.randn(N))})
				  
   # create a forecast engine, the main object handling all the operations
   lEngine = autof.cForecastEngine()

   # get the best time series model for predicting one week
   lEngine.train(iInputDS=df_train, iTime='Date', iSignal='Signal', iHorizon=7);
   lEngine.getModelInfo() # => relative error 7% (MAPE)

   # predict one week
   df_forecast = lEngine.forecast(iInputDS=df_train, iHorizon=7)
   # list the columns of the forecast dataset
   print(df_forecast.columns)

   # print the real forecasts
   # Future dates : ['2017-01-19T00:00:00.000000000' '2017-01-20T00:00:00.000000000' '2017-01-21T00:00:00.000000000' '2017-01-22T00:00:00.000000000' '2017-01-23T00:00:00.000000000' '2017-01-24T00:00:00.000000000' '2017-01-25T00:00:00.000000000']
   print(df_forecast['Date'].tail(7).values)

   # signal forecast : [ 9.74934646  10.04419761  12.15136455  12.20369717  14.09607727 15.68086323  16.22296559]
   print(df_forecast['Signal_Forecast'].tail(7).values)
```
[also availabe as a jupyter notebook](docs/sample_code.ipynb)


Features
--------
PyAF allows forecasting a time series (or a signal) for future values in a fully automated
way. To build forecasts, PyAF allows using time information (by identifying **long-term** evolution and **periodic patterns**), analyzes the **past of the signal**, exploits **exogenous data** (user-provided time series that may be correlated with the signal) as well as the **hierarchical structure** of the signal (by aggregating spatial components forecasts, for example). 

PyAF **uses [Pandas](http://pandas.pydata.org/) as a data access layer**. It consumes data coming from a pandas data-
frame (with time and signal columns), builds a time series model, and outputs
the forecasts in a pandas data-frame. Pandas is an excellent data access layer,
it allows reading/writing a huge set of file formats, accessing various data
sources (databases) and has an extensive set of algorithms to handle data-
frames (aggregation, statistics, linear algebra, plotting, etc.).


PyAF statistical time series models are built/estimated/trained using [scikit-learn](http://scikit-learn.org).


The following features are available :
   1. **Training a model** to forecast a time series (given in a pandas data-frame
      with time and signal columns).
        * PyAF uses a **machine learning approach** (the signal is cut into estimation
      and validation parts, respectively, 80% and 20% of the signal).
        * A [time-series cross-validation](https://github.com/antoinecarme/pyaf/issues/105) can also be used.
   2. Forecasting a time series model on a given **horizon** (forecast result is
      also a pandas data-frame) and providing **prediction/confidence intervals** for
      the forecasts.
   3. Generic training features
         * [Signal decomposition](http://en.wikipedia.org/wiki/Decomposition_of_time_series) as the sum of a trend, periodic and AR components.
         * PyAF works as a competition between a **comprehensive set of possible signal 
      transformations and linear decompositions**. For each transformed
      signal, a set of possible trends, periodic components and AR models is
      generated and all the possible combinations are estimated. The best
      decomposition in terms of performance is kept to forecast the signal (the
      performance is computed on a part of the signal that was not used for the
      estimation).
         * **Signal transformation** is supported before **signal decompositions**. Four
      transformations are supported by default. Other transformations are
      available (Box-Cox, etc.).
         * All models are estimated using **standard procedures and state-of-the-art
      time series modeling**. For example, trend regressions and AR/ARX models
      are estimated using scikit-learn linear regression models.
      * Standard performance measures are used (L1, RMSE, MAPE, MedAE, LnQ, etc.)
   4. PyAF analyzes the **time variable** and infers the frequency from the data.
      * Natural time frequencies are supported: Minute, Hour, Day, Week and Month.
      * Strange frequencies like every 3.2 days or every 17 minutes are supported if data are recorded accordingly (every other Monday => two weeks frequency).
      * The frequency is computed as the mean duration between consecutive observations by default (as a [pandas DateOffset](https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.tseries.offsets.DateOffset.html)).
      * The frequency is used to generate values for future dates automatically.
      * PyAF does its best when dates are not regularly observed. Time frequency is approximate in this case.
      * Real/Integer valued (fake) dates are also supported and handled in a similar way.
   4. **Exogenous Data Support**
        * Exogenous data can be provided to improve the forecasts. These are
      expected to be **stored in an external data-frame** (this data-frame will be
      merged with the training data-frame).
        * Exogenous data are integrated into the modeling process through their **past values**
      ([ARX models](http://en.wikipedia.org/wiki/Autoregressive%E2%80%93moving-average_model)).
        * **Exogenous variables can be of any type** (numeric, string, date or
      object).
        * Exogenous variables are **dummified** for the non-numeric types, and
      **standardized** for the numeric types.
   5. PyAF implements **Hierarchical Forecasting**. It follows the excellent approach used in [Rob J
      Hyndman and George Athanasopoulos book](http://www.otexts.org/fpp/9/4). Thanks @robjhyndman
        * **Hierarchies** and **grouped time series** are supported.
        * **Bottom-Up**, **Top-Down** (using proportions), **Middle-Out** and **Optimal Combinations** are
      implemented.
   6. The modeling process is **customizable** and has a huge set of **options**. The
      default values of these options should however be OK to produce a reasonable quality model in a limited amount of time (a few minutes).
      * These options give access to a full set of [signal transformations](https://github.com/antoinecarme/pyaf/blob/ba09233db42d43b9aa16b6151f00794193401841/TS/Options.py#L18) and [AR-like models](https://github.com/antoinecarme/pyaf/blob/ba09233db42d43b9aa16b6151f00794193401841/TS/Options.py#L37) that are not enabled by default.
      * Gives rise to Logit, Fisher transformations as well as XGBoost, Support Vector Regressions and Croston intermittent models, LGBM, among others.
      * By default, PyAF uses a **fast mode** that activates many popular models. It is also possible to activate a **slow mode**, in which PyAF explores all possible models.
      * Specific models and features can be customized.
   7. A **benchmarking process** is in place (using M1, M2, M3 competitions, NN3,
      NN5 forecasting competitions).
         * This process will be used to control the quality of modeling changes introduced in the future versions of PyAF. A  [related  github issue](https://github.com/antoinecarme/pyaf/issues/45) is created. 
         * Benchmarks data/reports are saved in a separate [github repository](https://github.com/antoinecarme/PyAF_Benchmarks). 
         * Sample [benchmark report](https://github.com/antoinecarme/PyAF_Benchmarks/blob/master/reporting/data/M1_COMP_debrief.csv) with 1001 datasets from the M1 Forecasting Competition.
   8. Basic **plotting** functions using matplotlib with standard time series and
      forecasts plots.
   9. **Software Quality** Highlights
      * An **object-oriented** approach is used for the system design. Separation of
      concerns is the key factor here.
      * **Fully written in Python** with NumPy, SciPy, Pandas and scikit-learn
      objects. Tries to be **column-based** everywhere for performance reasons (respecting some modeling time and memory constraints).
      * Internally using a fit/predict pattern, inspired by scikit-learn, to estimate/forecast the different signal components (trends, cycles and AR models).
      * A **test-driven approach (TDD)** is used. Test scripts are available in the [tests](tests)
      directory, one directory for each feature.
      * **TDD** implies that even the most recent features have some sample scripts in this [directory](tests). Want to know how to use cross-validation with PyAF? Here are [some scripts](tests/cross_validation).  
      * Some **[jupyter notebooks](docs)** are available for demo purposes with standard time series and forecasts plots.
      * Very **simple API** for training and forecasting.
   10. A basic **RESTful Web Service** (Flask) is available.
       * This service allows building a time series model, forecasting future data and some standard plots by providing a minimal specification of the signal in the JSON request body (at least a link to a csv file containing the data).
       * See [this doc](WS/README.md) and the [related github issue](https://github.com/antoinecarme/pyaf/issues/20) for more details.

PyAF is a work in progress. The set of features is evolving. Your feature
requests, comments, help, hints are very welcome.


Installation
------------

PyAF has been developed, tested and used on a python 3.x version. 

It can be installed from PyPI for the latest official release:
   
	pip install pyaf
   
The development version is also available by executing:

	pip install scipy pandas scikit-learn matplotlib pydot xgboost statsmodels
	pip install --upgrade git+git://github.com/antoinecarme/pyaf.git


Development
-----------

Code contributions are welcome. Bug reports, request for new features and
documentation, tests are welcome. Please use the GitHub platform for these tasks.

You can check the latest sources of PyAF from GitHub with the command::

	git clone http://github.com/antoinecarme/pyaf.git


Project history
-----------

This project was started in summer 2016 as a POC to check the feasibility of an
automatic forecasting tool based only on Python available data science software
(NumPy, SciPy, Pandas, scikit-learn, etc.).

See the [AUTHORS.rst](AUTHORS.rst) file for a complete list of contributors.

Help and Support
----------------

PyAF is currently maintained by the original developer. PyAF support will be
provided when possible and even if you are not creating an issue, you are encouraged to follow [these guidelines](ISSUE_TEMPLATE.md).

Bug reports, improvement requests, documentation, hints and test scripts are
welcome. Please use the GitHub platform for these tasks.

Please don't ask too much about new features. PyAF is only about forecasting (the last F). To keep PyAF design simple and flexible, we avoid [Feature Creep](https://en.wikipedia.org/wiki/Feature_creep).

For your commercial forecasting projects, please consider using the services of a forecasting expert near you (be it an R or a Python expert). 

Documentation
----------------

An [introductory notebook](https://github.com/antoinecarme/pyaf/blob/master/docs/PyAF_Introduction.ipynb) to the time series forecasting with PyAF is available here. It contains some real-world examples and use cases.

A specific notebook describing the use of exogenous data is [available here](https://github.com/antoinecarme/pyaf/blob/master/docs/PyAF_Exogenous.ipynb).

Notebooks describing an example of hierarchical forecasting models are available for [Signal Hierarchies](https://github.com/antoinecarme/pyaf/blob/master/docs/PyAF_Hierarchical_FrenchWineExportation.ipynb) and for [Grouped Signals](https://github.com/antoinecarme/pyaf/blob/master/docs/PyAF_GroupedSignals_FrenchWineExportation.ipynb).

The python code is not yet fully documented. This is a top priority (TODO). 

Communication
----------------

Comments, appreciations, remarks, etc .... are welcome. Your feedback is
welcome if you use this library in a project or a publication.

            

Raw data

            {
    "_id": null,
    "home_page": "https://github.com/antoinecarme/pyaf",
    "name": "pyaf",
    "maintainer": "",
    "docs_url": null,
    "requires_python": ">=3",
    "maintainer_email": "",
    "keywords": "arx automatic-forecasting autoregressive benchmark cycle decomposition exogenous forecasting heroku hierarchical-forecasting horizon jupyter pandas python scikit-learn seasonal time-series transformation trend web-service",
    "author": "Antoine CARME",
    "author_email": "antoine.carme@outlook.com",
    "download_url": "",
    "platform": null,
    "description": "\nPyAF (Python Automatic Forecasting)\n===================================\n\n[![Github/CI/CD](https://github.com/antoinecarme/pyaf/actions/workflows/python-app.yml/badge.svg)](https://github.com/antoinecarme/pyaf/actions/workflows/python-app.yml)\n\nPyAF is an Open Source Python library for Automatic Forecasting built on top of\npopular data science python modules: NumPy, SciPy, Pandas and scikit-learn.\n\nPyAF works as an automated process for predicting future values of a signal\nusing a machine learning approach. It provides a set of features that is\ncomparable to some popular commercial automatic forecasting products.\n\nPyAF has been developed, tested and benchmarked using a **python 3.x** version.\n\nPyAF is distributed under the [3-Clause BSD license](https://tldrlegal.com/license/bsd-3-clause-license-%28revised%29).\n\n\nDemo \n----\n```Python\nimport numpy as np, pandas as pd\nimport pyaf.ForecastEngine as autof\n\nif __name__ == '__main__':\n   # generate a daily signal covering one year 2016 in a pandas dataframe\n   N = 360\n   df_train = pd.DataFrame({\"Date\": pd.date_range(start=\"2016-01-25\", periods=N, freq='D'),\n   \t                    \"Signal\": (np.arange(N)//40 + np.arange(N) % 21 + np.random.randn(N))})\n\t\t\t\t  \n   # create a forecast engine, the main object handling all the operations\n   lEngine = autof.cForecastEngine()\n\n   # get the best time series model for predicting one week\n   lEngine.train(iInputDS=df_train, iTime='Date', iSignal='Signal', iHorizon=7);\n   lEngine.getModelInfo() # => relative error 7% (MAPE)\n\n   # predict one week\n   df_forecast = lEngine.forecast(iInputDS=df_train, iHorizon=7)\n   # list the columns of the forecast dataset\n   print(df_forecast.columns)\n\n   # print the real forecasts\n   # Future dates : ['2017-01-19T00:00:00.000000000' '2017-01-20T00:00:00.000000000' '2017-01-21T00:00:00.000000000' '2017-01-22T00:00:00.000000000' '2017-01-23T00:00:00.000000000' '2017-01-24T00:00:00.000000000' '2017-01-25T00:00:00.000000000']\n   print(df_forecast['Date'].tail(7).values)\n\n   # signal forecast : [ 9.74934646  10.04419761  12.15136455  12.20369717  14.09607727 15.68086323  16.22296559]\n   print(df_forecast['Signal_Forecast'].tail(7).values)\n```\n[also availabe as a jupyter notebook](docs/sample_code.ipynb)\n\n\nFeatures\n--------\nPyAF allows forecasting a time series (or a signal) for future values in a fully automated\nway. To build forecasts, PyAF allows using time information (by identifying **long-term** evolution and **periodic patterns**), analyzes the **past of the signal**, exploits **exogenous data** (user-provided time series that may be correlated with the signal) as well as the **hierarchical structure** of the signal (by aggregating spatial components forecasts, for example). \n\nPyAF **uses [Pandas](http://pandas.pydata.org/) as a data access layer**. It consumes data coming from a pandas data-\nframe (with time and signal columns), builds a time series model, and outputs\nthe forecasts in a pandas data-frame. Pandas is an excellent data access layer,\nit allows reading/writing a huge set of file formats, accessing various data\nsources (databases) and has an extensive set of algorithms to handle data-\nframes (aggregation, statistics, linear algebra, plotting, etc.).\n\n\nPyAF statistical time series models are built/estimated/trained using [scikit-learn](http://scikit-learn.org).\n\n\nThe following features are available :\n   1. **Training a model** to forecast a time series (given in a pandas data-frame\n      with time and signal columns).\n        * PyAF uses a **machine learning approach** (the signal is cut into estimation\n      and validation parts, respectively, 80% and 20% of the signal).\n        * A [time-series cross-validation](https://github.com/antoinecarme/pyaf/issues/105) can also be used.\n   2. Forecasting a time series model on a given **horizon** (forecast result is\n      also a pandas data-frame) and providing **prediction/confidence intervals** for\n      the forecasts.\n   3. Generic training features\n         * [Signal decomposition](http://en.wikipedia.org/wiki/Decomposition_of_time_series) as the sum of a trend, periodic and AR components.\n         * PyAF works as a competition between a **comprehensive set of possible signal \n      transformations and linear decompositions**. For each transformed\n      signal, a set of possible trends, periodic components and AR models is\n      generated and all the possible combinations are estimated. The best\n      decomposition in terms of performance is kept to forecast the signal (the\n      performance is computed on a part of the signal that was not used for the\n      estimation).\n         * **Signal transformation** is supported before **signal decompositions**. Four\n      transformations are supported by default. Other transformations are\n      available (Box-Cox, etc.).\n         * All models are estimated using **standard procedures and state-of-the-art\n      time series modeling**. For example, trend regressions and AR/ARX models\n      are estimated using scikit-learn linear regression models.\n      * Standard performance measures are used (L1, RMSE, MAPE, MedAE, LnQ, etc.)\n   4. PyAF analyzes the **time variable** and infers the frequency from the data.\n      * Natural time frequencies are supported: Minute, Hour, Day, Week and Month.\n      * Strange frequencies like every 3.2 days or every 17 minutes are supported if data are recorded accordingly (every other Monday => two weeks frequency).\n      * The frequency is computed as the mean duration between consecutive observations by default (as a [pandas DateOffset](https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.tseries.offsets.DateOffset.html)).\n      * The frequency is used to generate values for future dates automatically.\n      * PyAF does its best when dates are not regularly observed. Time frequency is approximate in this case.\n      * Real/Integer valued (fake) dates are also supported and handled in a similar way.\n   4. **Exogenous Data Support**\n        * Exogenous data can be provided to improve the forecasts. These are\n      expected to be **stored in an external data-frame** (this data-frame will be\n      merged with the training data-frame).\n        * Exogenous data are integrated into the modeling process through their **past values**\n      ([ARX models](http://en.wikipedia.org/wiki/Autoregressive%E2%80%93moving-average_model)).\n        * **Exogenous variables can be of any type** (numeric, string, date or\n      object).\n        * Exogenous variables are **dummified** for the non-numeric types, and\n      **standardized** for the numeric types.\n   5. PyAF implements **Hierarchical Forecasting**. It follows the excellent approach used in [Rob J\n      Hyndman and George Athanasopoulos book](http://www.otexts.org/fpp/9/4). Thanks @robjhyndman\n        * **Hierarchies** and **grouped time series** are supported.\n        * **Bottom-Up**, **Top-Down** (using proportions), **Middle-Out** and **Optimal Combinations** are\n      implemented.\n   6. The modeling process is **customizable** and has a huge set of **options**. The\n      default values of these options should however be OK to produce a reasonable quality model in a limited amount of time (a few minutes).\n      * These options give access to a full set of [signal transformations](https://github.com/antoinecarme/pyaf/blob/ba09233db42d43b9aa16b6151f00794193401841/TS/Options.py#L18) and [AR-like models](https://github.com/antoinecarme/pyaf/blob/ba09233db42d43b9aa16b6151f00794193401841/TS/Options.py#L37) that are not enabled by default.\n      * Gives rise to Logit, Fisher transformations as well as XGBoost, Support Vector Regressions and Croston intermittent models, LGBM, among others.\n      * By default, PyAF uses a **fast mode** that activates many popular models. It is also possible to activate a **slow mode**, in which PyAF explores all possible models.\n      * Specific models and features can be customized.\n   7. A **benchmarking process** is in place (using M1, M2, M3 competitions, NN3,\n      NN5 forecasting competitions).\n         * This process will be used to control the quality of modeling changes introduced in the future versions of PyAF. A  [related  github issue](https://github.com/antoinecarme/pyaf/issues/45) is created. \n         * Benchmarks data/reports are saved in a separate [github repository](https://github.com/antoinecarme/PyAF_Benchmarks). \n         * Sample [benchmark report](https://github.com/antoinecarme/PyAF_Benchmarks/blob/master/reporting/data/M1_COMP_debrief.csv) with 1001 datasets from the M1 Forecasting Competition.\n   8. Basic **plotting** functions using matplotlib with standard time series and\n      forecasts plots.\n   9. **Software Quality** Highlights\n      * An **object-oriented** approach is used for the system design. Separation of\n      concerns is the key factor here.\n      * **Fully written in Python** with NumPy, SciPy, Pandas and scikit-learn\n      objects. Tries to be **column-based** everywhere for performance reasons (respecting some modeling time and memory constraints).\n      * Internally using a fit/predict pattern, inspired by scikit-learn, to estimate/forecast the different signal components (trends, cycles and AR models).\n      * A **test-driven approach (TDD)** is used. Test scripts are available in the [tests](tests)\n      directory, one directory for each feature.\n      * **TDD** implies that even the most recent features have some sample scripts in this [directory](tests). Want to know how to use cross-validation with PyAF? Here are [some scripts](tests/cross_validation).  \n      * Some **[jupyter notebooks](docs)** are available for demo purposes with standard time series and forecasts plots.\n      * Very **simple API** for training and forecasting.\n   10. A basic **RESTful Web Service** (Flask) is available.\n       * This service allows building a time series model, forecasting future data and some standard plots by providing a minimal specification of the signal in the JSON request body (at least a link to a csv file containing the data).\n       * See [this doc](WS/README.md) and the [related github issue](https://github.com/antoinecarme/pyaf/issues/20) for more details.\n\nPyAF is a work in progress. The set of features is evolving. Your feature\nrequests, comments, help, hints are very welcome.\n\n\nInstallation\n------------\n\nPyAF has been developed, tested and used on a python 3.x version. \n\nIt can be installed from PyPI for the latest official release:\n   \n\tpip install pyaf\n   \nThe development version is also available by executing:\n\n\tpip install scipy pandas scikit-learn matplotlib pydot xgboost statsmodels\n\tpip install --upgrade git+git://github.com/antoinecarme/pyaf.git\n\n\nDevelopment\n-----------\n\nCode contributions are welcome. Bug reports, request for new features and\ndocumentation, tests are welcome. Please use the GitHub platform for these tasks.\n\nYou can check the latest sources of PyAF from GitHub with the command::\n\n\tgit clone http://github.com/antoinecarme/pyaf.git\n\n\nProject history\n-----------\n\nThis project was started in summer 2016 as a POC to check the feasibility of an\nautomatic forecasting tool based only on Python available data science software\n(NumPy, SciPy, Pandas, scikit-learn, etc.).\n\nSee the [AUTHORS.rst](AUTHORS.rst) file for a complete list of contributors.\n\nHelp and Support\n----------------\n\nPyAF is currently maintained by the original developer. PyAF support will be\nprovided when possible and even if you are not creating an issue, you are encouraged to follow [these guidelines](ISSUE_TEMPLATE.md).\n\nBug reports, improvement requests, documentation, hints and test scripts are\nwelcome. Please use the GitHub platform for these tasks.\n\nPlease don't ask too much about new features. PyAF is only about forecasting (the last F). To keep PyAF design simple and flexible, we avoid [Feature Creep](https://en.wikipedia.org/wiki/Feature_creep).\n\nFor your commercial forecasting projects, please consider using the services of a forecasting expert near you (be it an R or a Python expert). \n\nDocumentation\n----------------\n\nAn [introductory notebook](https://github.com/antoinecarme/pyaf/blob/master/docs/PyAF_Introduction.ipynb) to the time series forecasting with PyAF is available here. It contains some real-world examples and use cases.\n\nA specific notebook describing the use of exogenous data is [available here](https://github.com/antoinecarme/pyaf/blob/master/docs/PyAF_Exogenous.ipynb).\n\nNotebooks describing an example of hierarchical forecasting models are available for [Signal Hierarchies](https://github.com/antoinecarme/pyaf/blob/master/docs/PyAF_Hierarchical_FrenchWineExportation.ipynb) and for [Grouped Signals](https://github.com/antoinecarme/pyaf/blob/master/docs/PyAF_GroupedSignals_FrenchWineExportation.ipynb).\n\nThe python code is not yet fully documented. This is a top priority (TODO). \n\nCommunication\n----------------\n\nComments, appreciations, remarks, etc .... are welcome. Your feedback is\nwelcome if you use this library in a project or a publication.\n",
    "bugtrack_url": null,
    "license": "BSD 3-clause",
    "summary": "Python Automatic Forecasting",
    "version": "5.0",
    "project_urls": {
        "Homepage": "https://github.com/antoinecarme/pyaf"
    },
    "split_keywords": [
        "arx",
        "automatic-forecasting",
        "autoregressive",
        "benchmark",
        "cycle",
        "decomposition",
        "exogenous",
        "forecasting",
        "heroku",
        "hierarchical-forecasting",
        "horizon",
        "jupyter",
        "pandas",
        "python",
        "scikit-learn",
        "seasonal",
        "time-series",
        "transformation",
        "trend",
        "web-service"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "d735b798eab23e24d570d481cf97b3633f51a44569e19bdfc4022f065b944cea",
                "md5": "a190cf3c0a065c012de56d36426cde4f",
                "sha256": "8122dca416b8c8662fe6d68e94f42de9b831cacc1bcc8ba9e5c5631792828af0"
            },
            "downloads": -1,
            "filename": "pyaf-5.0-py2.py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "a190cf3c0a065c012de56d36426cde4f",
            "packagetype": "bdist_wheel",
            "python_version": "py2.py3",
            "requires_python": ">=3",
            "size": 130730,
            "upload_time": "2023-07-12T22:24:01",
            "upload_time_iso_8601": "2023-07-12T22:24:01.310004Z",
            "url": "https://files.pythonhosted.org/packages/d7/35/b798eab23e24d570d481cf97b3633f51a44569e19bdfc4022f065b944cea/pyaf-5.0-py2.py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2023-07-12 22:24:01",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "antoinecarme",
    "github_project": "pyaf",
    "travis_ci": true,
    "coveralls": false,
    "github_actions": true,
    "circle": true,
    "requirements": [
        {
            "name": "pandas",
            "specs": []
        },
        {
            "name": "datetime",
            "specs": []
        },
        {
            "name": "scikit-learn",
            "specs": []
        },
        {
            "name": "matplotlib",
            "specs": []
        },
        {
            "name": "pydot",
            "specs": []
        },
        {
            "name": "flask",
            "specs": []
        },
        {
            "name": "statsmodels",
            "specs": []
        }
    ],
    "lcname": "pyaf"
}
        
Elapsed time: 0.08683s