returnn


Namereturnn JSON
Version 1.20240327.165809 PyPI version JSON
download
home_pagehttps://github.com/rwth-i6/returnn/
SummaryThe RWTH extensible training framework for universal recurrent neural networks
upload_time2024-03-27 16:14:30
maintainerNone
docs_urlNone
authorAlbert Zeyer
requires_pythonNone
licenseRETURNN license
keywords
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            ==================
Welcome to RETURNN
==================

`GitHub repository <https://github.com/rwth-i6/returnn>`__.
`RETURNN paper 2016 <https://arxiv.org/abs/1608.00895>`_,
`RETURNN paper 2018 <https://arxiv.org/abs/1805.05225>`_.

RETURNN - RWTH extensible training framework for universal recurrent neural networks,
is a Theano/TensorFlow-based implementation of modern recurrent neural network architectures.
It is optimized for fast and reliable training of recurrent neural networks in a multi-GPU environment.

The high-level features and goals of RETURNN are:

* **Simplicity**

  * Writing config / code is simple & straight-forward (setting up experiment, defining model)
  * Debugging in case of problems is simple
  * Reading config / code is simple (defined model, training, decoding all becomes clear)

* **Flexibility**

  * Allow for many different kinds of experiments / models

* **Efficiency**

  * Training speed
  * Decoding speed

All items are important for research, decoding speed is esp. important for production.

See our `Interspeech 2020 tutorial "Efficient and Flexible Implementation of Machine Learning for ASR and MT" video <https://www.youtube.com/watch?v=wPKdYqSOlAY>`__
(`slides <https://www-i6.informatik.rwth-aachen.de/publications/download/1154/Zeyer--2020.pdf>`__)
with an introduction of the core concepts.

More specific features include:

- Mini-batch training of feed-forward neural networks
- Sequence-chunking based batch training for recurrent neural networks
- Long short-term memory recurrent neural networks
  including our own fast CUDA kernel
- Multidimensional LSTM (GPU only, there is no CPU version)
- Memory management for large data sets
- Work distribution across multiple devices
- Flexible and fast architecture which allows all kinds of encoder-attention-decoder models

See `documentation <https://returnn.readthedocs.io/>`__.
See `basic usage <https://returnn.readthedocs.io/en/latest/basic_usage.html>`__
and `technological overview <https://returnn.readthedocs.io/en/latest/tech_overview.html>`__.

`Here is the video recording of a RETURNN overview talk <https://www-i6.informatik.rwth-aachen.de/web/Software/returnn/downloads/workshop-2019-01-29/01.recording.cut.mp4>`_
(`slides <https://www-i6.informatik.rwth-aachen.de/web/Software/returnn/downloads/workshop-2019-01-29/01.returnn-overview.session1.handout.v1.pdf>`__,
`exercise sheet <https://www-i6.informatik.rwth-aachen.de/web/Software/returnn/downloads/workshop-2019-01-29/01.exercise_sheet.pdf>`__;
hosted by eBay).

There are `many example demos <https://github.com/rwth-i6/returnn/blob/master/demos/>`_
which work on artificially generated data,
i.e. they should work as-is.

There are `some real-world examples <https://github.com/rwth-i6/returnn-experiments>`_
such as setups for speech recognition on the Switchboard or LibriSpeech corpus.

Some benchmark setups against other frameworks
can be found `here <https://github.com/rwth-i6/returnn-benchmarks>`_.
The results are in the `RETURNN paper 2016 <https://arxiv.org/abs/1608.00895>`_.
Performance benchmarks of our LSTM kernel vs CuDNN and other TensorFlow kernels
are in `TensorFlow LSTM benchmark <https://returnn.readthedocs.io/en/latest/tf_lstm_benchmark.html>`__.

There is also `a wiki <https://github.com/rwth-i6/returnn/wiki>`_.
Questions can also be asked on
`StackOverflow using the RETURNN tag <https://stackoverflow.com/questions/tagged/returnn>`_.

.. image:: https://github.com/rwth-i6/returnn/workflows/CI/badge.svg
    :target: https://github.com/rwth-i6/returnn/actions

            

Raw data

            {
    "_id": null,
    "home_page": "https://github.com/rwth-i6/returnn/",
    "name": "returnn",
    "maintainer": null,
    "docs_url": null,
    "requires_python": null,
    "maintainer_email": null,
    "keywords": null,
    "author": "Albert Zeyer",
    "author_email": "albzey@gmail.com",
    "download_url": "https://files.pythonhosted.org/packages/2a/5e/3f7de6a9865c5aeaa60a739883127a4576e685352374f76ebd03a4a46485/returnn-1.20240327.165809.tar.gz",
    "platform": null,
    "description": "==================\nWelcome to RETURNN\n==================\n\n`GitHub repository <https://github.com/rwth-i6/returnn>`__.\n`RETURNN paper 2016 <https://arxiv.org/abs/1608.00895>`_,\n`RETURNN paper 2018 <https://arxiv.org/abs/1805.05225>`_.\n\nRETURNN - RWTH extensible training framework for universal recurrent neural networks,\nis a Theano/TensorFlow-based implementation of modern recurrent neural network architectures.\nIt is optimized for fast and reliable training of recurrent neural networks in a multi-GPU environment.\n\nThe high-level features and goals of RETURNN are:\n\n* **Simplicity**\n\n  * Writing config / code is simple & straight-forward (setting up experiment, defining model)\n  * Debugging in case of problems is simple\n  * Reading config / code is simple (defined model, training, decoding all becomes clear)\n\n* **Flexibility**\n\n  * Allow for many different kinds of experiments / models\n\n* **Efficiency**\n\n  * Training speed\n  * Decoding speed\n\nAll items are important for research, decoding speed is esp. important for production.\n\nSee our `Interspeech 2020 tutorial \"Efficient and Flexible Implementation of Machine Learning for ASR and MT\" video <https://www.youtube.com/watch?v=wPKdYqSOlAY>`__\n(`slides <https://www-i6.informatik.rwth-aachen.de/publications/download/1154/Zeyer--2020.pdf>`__)\nwith an introduction of the core concepts.\n\nMore specific features include:\n\n- Mini-batch training of feed-forward neural networks\n- Sequence-chunking based batch training for recurrent neural networks\n- Long short-term memory recurrent neural networks\n  including our own fast CUDA kernel\n- Multidimensional LSTM (GPU only, there is no CPU version)\n- Memory management for large data sets\n- Work distribution across multiple devices\n- Flexible and fast architecture which allows all kinds of encoder-attention-decoder models\n\nSee `documentation <https://returnn.readthedocs.io/>`__.\nSee `basic usage <https://returnn.readthedocs.io/en/latest/basic_usage.html>`__\nand `technological overview <https://returnn.readthedocs.io/en/latest/tech_overview.html>`__.\n\n`Here is the video recording of a RETURNN overview talk <https://www-i6.informatik.rwth-aachen.de/web/Software/returnn/downloads/workshop-2019-01-29/01.recording.cut.mp4>`_\n(`slides <https://www-i6.informatik.rwth-aachen.de/web/Software/returnn/downloads/workshop-2019-01-29/01.returnn-overview.session1.handout.v1.pdf>`__,\n`exercise sheet <https://www-i6.informatik.rwth-aachen.de/web/Software/returnn/downloads/workshop-2019-01-29/01.exercise_sheet.pdf>`__;\nhosted by eBay).\n\nThere are `many example demos <https://github.com/rwth-i6/returnn/blob/master/demos/>`_\nwhich work on artificially generated data,\ni.e. they should work as-is.\n\nThere are `some real-world examples <https://github.com/rwth-i6/returnn-experiments>`_\nsuch as setups for speech recognition on the Switchboard or LibriSpeech corpus.\n\nSome benchmark setups against other frameworks\ncan be found `here <https://github.com/rwth-i6/returnn-benchmarks>`_.\nThe results are in the `RETURNN paper 2016 <https://arxiv.org/abs/1608.00895>`_.\nPerformance benchmarks of our LSTM kernel vs CuDNN and other TensorFlow kernels\nare in `TensorFlow LSTM benchmark <https://returnn.readthedocs.io/en/latest/tf_lstm_benchmark.html>`__.\n\nThere is also `a wiki <https://github.com/rwth-i6/returnn/wiki>`_.\nQuestions can also be asked on\n`StackOverflow using the RETURNN tag <https://stackoverflow.com/questions/tagged/returnn>`_.\n\n.. image:: https://github.com/rwth-i6/returnn/workflows/CI/badge.svg\n    :target: https://github.com/rwth-i6/returnn/actions\n",
    "bugtrack_url": null,
    "license": "RETURNN license",
    "summary": "The RWTH extensible training framework for universal recurrent neural networks",
    "version": "1.20240327.165809",
    "project_urls": {
        "Homepage": "https://github.com/rwth-i6/returnn/"
    },
    "split_keywords": [],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "2a5e3f7de6a9865c5aeaa60a739883127a4576e685352374f76ebd03a4a46485",
                "md5": "349bb97592876ed87cb5dcbe3a769058",
                "sha256": "b55cdf0cfea74b261ebe989317eb87fc02bd9448a3b6e8e47dbe7c8118174852"
            },
            "downloads": -1,
            "filename": "returnn-1.20240327.165809.tar.gz",
            "has_sig": false,
            "md5_digest": "349bb97592876ed87cb5dcbe3a769058",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": null,
            "size": 2164139,
            "upload_time": "2024-03-27T16:14:30",
            "upload_time_iso_8601": "2024-03-27T16:14:30.753818Z",
            "url": "https://files.pythonhosted.org/packages/2a/5e/3f7de6a9865c5aeaa60a739883127a4576e685352374f76ebd03a4a46485/returnn-1.20240327.165809.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-03-27 16:14:30",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "rwth-i6",
    "github_project": "returnn",
    "travis_ci": false,
    "coveralls": false,
    "github_actions": true,
    "requirements": [],
    "lcname": "returnn"
}
        
Elapsed time: 0.24268s