twm-faust


Nametwm-faust JSON
Version 1.14.3 PyPI version JSON
download
home_pagehttp://faust.readthedocs.io/
SummaryPython Stream processing.
upload_time2024-03-06 03:44:11
maintainer
docs_urlNone
authorRobinhood Markets, Inc.
requires_python>=3.6.0
licenseBSD 3-Clause
keywords stream processing asyncio distributed queue kafka
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI
coveralls test coverage
            .. XXX Need to change this image to readthedocs before release

.. image:: https://raw.githubusercontent.com/robinhood/faust/8ee5e209322d9edf5bdb79b992ef986be2de4bb4/artwork/banner-alt1.png

===========================
 Python Stream Processing
===========================

Note: This project is a fork of the original **Faust** stream processing library.

|build-status| |coverage| |license| |wheel| |pyversion| |pyimp|

:Version: 1.14.3
:Web: http://faust.readthedocs.io/
:Download: http://pypi.org/project/faust
:Source: http://github.com/robinhood/faust
:Keywords: distributed, stream, async, processing, data, queue, state management


.. sourcecode:: python

    # Python Streams
    # Forever scalable event processing & in-memory durable K/V store;
    # as a library w/ asyncio & static typing.
    import faust

**Faust** is a stream processing library, porting the ideas from
`Kafka Streams`_ to Python.

It is used at `Robinhood`_ to build high performance distributed systems
and real-time data pipelines that process billions of events every day.

Faust provides both *stream processing* and *event processing*,
sharing similarity with tools such as
`Kafka Streams`_, `Apache Spark`_/`Storm`_/`Samza`_/`Flink`_,

It does not use a DSL, it's just Python!
This means you can use all your favorite Python libraries
when stream processing: NumPy, PyTorch, Pandas, NLTK, Django,
Flask, SQLAlchemy, ++

Faust requires Python 3.6 or later for the new `async/await`_ syntax,
and variable type annotations.

Here's an example processing a stream of incoming orders:

.. sourcecode:: python

    app = faust.App('myapp', broker='kafka://localhost')

    # Models describe how messages are serialized:
    # {"account_id": "3fae-...", "amount": 3}
    class Order(faust.Record):
        account_id: str
        amount: int

    @app.agent(value_type=Order)
    async def order(orders):
        async for order in orders:
            # process infinite stream of orders.
            print(f'Order for {order.account_id}: {order.amount}')

The Agent decorator defines a "stream processor" that essentially
consumes from a Kafka topic and does something for every event it receives.

The agent is an ``async def`` function, so can also perform
other operations asynchronously, such as web requests.

This system can persist state, acting like a database.
Tables are named distributed key/value stores you can use
as regular Python dictionaries.

Tables are stored locally on each machine using a super fast
embedded database written in C++, called `RocksDB`_.

Tables can also store aggregate counts that are optionally "windowed"
so you can keep track
of "number of clicks from the last day," or
"number of clicks in the last hour." for example. Like `Kafka Streams`_,
we support tumbling, hopping and sliding windows of time, and old windows
can be expired to stop data from filling up.

For reliability we use a Kafka topic as "write-ahead-log".
Whenever a key is changed we publish to the changelog.
Standby nodes consume from this changelog to keep an exact replica
of the data and enables instant recovery should any of the nodes fail.

To the user a table is just a dictionary, but data is persisted between
restarts and replicated across nodes so on failover other nodes can take over
automatically.

You can count page views by URL:

.. sourcecode:: python

    # data sent to 'clicks' topic sharded by URL key.
    # e.g. key="http://example.com" value="1"
    click_topic = app.topic('clicks', key_type=str, value_type=int)

    # default value for missing URL will be 0 with `default=int`
    counts = app.Table('click_counts', default=int)

    @app.agent(click_topic)
    async def count_click(clicks):
        async for url, count in clicks.items():
            counts[url] += count

The data sent to the Kafka topic is partitioned, which means
the clicks will be sharded by URL in such a way that every count
for the same URL will be delivered to the same Faust worker instance.


Faust supports any type of stream data: bytes, Unicode and serialized
structures, but also comes with "Models" that use modern Python
syntax to describe how keys and values in streams are serialized:

.. sourcecode:: python

    # Order is a json serialized dictionary,
    # having these fields:

    class Order(faust.Record):
        account_id: str
        product_id: str
        price: float
        quantity: float = 1.0

    orders_topic = app.topic('orders', key_type=str, value_type=Order)

    @app.agent(orders_topic)
    async def process_order(orders):
        async for order in orders:
            # process each order using regular Python
            total_price = order.price * order.quantity
            await send_order_received_email(order.account_id, order)

Faust is statically typed, using the ``mypy`` type checker,
so you can take advantage of static types when writing applications.

The Faust source code is small, well organized, and serves as a good
resource for learning the implementation of `Kafka Streams`_.

**Learn more about Faust in the** `introduction`_ **introduction page**
    to read more about Faust, system requirements, installation instructions,
    community resources, and more.

**or go directly to the** `quickstart`_ **tutorial**
    to see Faust in action by programming a streaming application.

**then explore the** `User Guide`_
    for in-depth information organized by topic.

.. _`Robinhood`: http://robinhood.com
.. _`async/await`:
    https://medium.freecodecamp.org/a-guide-to-asynchronous-programming-in-python-with-asyncio-232e2afa44f6
.. _`Kafka Streams`: https://kafka.apache.org/documentation/streams
.. _`Apache Spark`: http://spark.apache.org
.. _`Storm`: http://storm.apache.org
.. _`Samza`: http://samza.apache.org
.. _`Flink`: http://flink.apache.org
.. _`RocksDB`: http://rocksdb.org

.. _`introduction`: http://faust.readthedocs.io/en/latest/introduction.html

.. _`quickstart`: http://faust.readthedocs.io/en/latest/playbooks/quickstart.html

.. _`User Guide`: http://faust.readthedocs.io/en/latest/userguide/index.html

Faust is...
===========

**Simple**
    Faust is extremely easy to use. To get started using other stream processing
    solutions you have complicated hello-world projects, and
    infrastructure requirements.  Faust only requires Kafka,
    the rest is just Python, so If you know Python you can already use Faust to do
    stream processing, and it can integrate with just about anything.

    Here's one of the easier applications you can make::

        import faust

        class Greeting(faust.Record):
            from_name: str
            to_name: str

        app = faust.App('hello-app', broker='kafka://localhost')
        topic = app.topic('hello-topic', value_type=Greeting)

        @app.agent(topic)
        async def hello(greetings):
            async for greeting in greetings:
                print(f'Hello from {greeting.from_name} to {greeting.to_name}')

        @app.timer(interval=1.0)
        async def example_sender(app):
            await hello.send(
                value=Greeting(from_name='Faust', to_name='you'),
            )

        if __name__ == '__main__':
            app.main()

    You're probably a bit intimidated by the `async` and `await` keywords,
    but you don't have to know how ``asyncio`` works to use
    Faust: just mimic the examples, and you'll be fine.

    The example application starts two tasks: one is processing a stream,
    the other is a background thread sending events to that stream.
    In a real-life application, your system will publish
    events to Kafka topics that your processors can consume from,
    and the background thread is only needed to feed data into our
    example.

**Highly Available**
    Faust is highly available and can survive network problems and server
    crashes.  In the case of node failure, it can automatically recover,
    and tables have standby nodes that will take over.

**Distributed**
    Start more instances of your application as needed.

**Fast**
    A single-core Faust worker instance can already process tens of thousands
    of events every second, and we are reasonably confident that throughput will
    increase once we can support a more optimized Kafka client.

**Flexible**
    Faust is just Python, and a stream is an infinite asynchronous iterator.
    If you know how to use Python, you already know how to use Faust,
    and it works with your favorite Python libraries like Django, Flask,
    SQLAlchemy, NTLK, NumPy, SciPy, TensorFlow, etc.

Installation
============

You can install Faust either via the Python Package Index (PyPI)
or from source.

To install using `pip`:

.. sourcecode:: console

    $ pip install -U faust

Bundles
-------

Faust also defines a group of ``setuptools`` extensions that can be used
to install Faust and the dependencies for a given feature.

You can specify these in your requirements or on the ``pip``
command-line by using brackets. Separate multiple bundles using the comma:

.. sourcecode:: console

    $ pip install "faust[rocksdb]"

    $ pip install "faust[rocksdb,uvloop,fast,redis]"

The following bundles are available:

Stores
~~~~~~

:``faust[rocksdb]``:
    for using `RocksDB`_ for storing Faust table state.

    **Recommended in production.**


Caching
~~~~~~~

:``faust[redis]``:
    for using `Redis_` as a simple caching backend (Memcached-style).

Codecs
~~~~~~

:``faust[yaml]``:
    for using YAML and the ``PyYAML`` library in streams.

Optimization
~~~~~~~~~~~~

:``faust[fast]``:
    for installing all the available C speedup extensions to Faust core.

Sensors
~~~~~~~

:``faust[datadog]``:
    for using the Datadog Faust monitor.

:``faust[statsd]``:
    for using the Statsd Faust monitor.

Event Loops
~~~~~~~~~~~

:``faust[uvloop]``:
    for using Faust with ``uvloop``.

:``faust[eventlet]``:
    for using Faust with ``eventlet``

Debugging
~~~~~~~~~

:``faust[debug]``:
    for using ``aiomonitor`` to connect and debug a running Faust worker.

:``faust[setproctitle]``:
    when the ``setproctitle`` module is installed the Faust worker will
    use it to set a nicer process name in ``ps``/``top`` listings.
    Also installed with the ``fast`` and ``debug`` bundles.

Downloading and installing from source
--------------------------------------

Download the latest version of Faust from
http://pypi.org/project/faust

You can install it by doing:

.. sourcecode:: console

    $ tar xvfz faust-0.0.0.tar.gz
    $ cd faust-0.0.0
    $ python setup.py build
    # python setup.py install

The last command must be executed as a privileged user if
you are not currently using a virtualenv.

Using the development version
-----------------------------

With pip
~~~~~~~~

You can install the latest snapshot of Faust using the following
``pip`` command:

.. sourcecode:: console

    $ pip install https://github.com/robinhood/faust/zipball/master#egg=faust

FAQ
===

Can I use Faust with Django/Flask/etc.?
---------------------------------------

Yes! Use ``eventlet`` as a bridge to integrate with ``asyncio``.


Using ``eventlet``
~~~~~~~~~~~~~~~~~~~~~~

This approach works with any blocking Python library that can work with
``eventlet``.

Using ``eventlet`` requires you to install the ``aioeventlet`` module,
and you can install this as a bundle along with Faust:

.. sourcecode:: console

    $ pip install -U faust[eventlet]

Then to actually use eventlet as the event loop you have to either
use the ``-L <faust --loop>`` argument to the ``faust`` program:

.. sourcecode:: console

    $ faust -L eventlet -A myproj worker -l info

or add ``import mode.loop.eventlet`` at the top of your entry point script:

.. sourcecode:: python

    #!/usr/bin/env python3
    import mode.loop.eventlet  # noqa

.. warning::

    It's very important this is at the very top of the module,
    and that it executes before you import libraries.

Can I use Faust with Tornado?
-----------------------------

Yes! Use the ``tornado.platform.asyncio`` bridge:
http://www.tornadoweb.org/en/stable/asyncio.html

Can I use Faust with Twisted?
-----------------------------

Yes! Use the ``asyncio`` reactor implementation:
https://twistedmatrix.com/documents/17.1.0/api/twisted.internet.asyncioreactor.html


Will you support Python 2.7 or Python 3.5?
------------------------------------------

No. Faust requires Python 3.6 or later, since it heavily uses features that were
introduced in Python 3.6 (`async`, `await`, variable type annotations).


I get a maximum number of open files exceeded error by RocksDB when running a Faust app locally. How can I fix this?
--------------------------------------------------------------------------------------------------------------------

You may need to increase the limit for the maximum number of open files. The
following post explains how to do so on OS X:
https://blog.dekstroza.io/ulimit-shenanigans-on-osx-el-capitan/


What kafka versions faust supports?
---------------------------------------

Faust supports kafka with version >= 0.10.

Getting Help
============

Slack
-----

For discussions about the usage, development, and future of Faust,
please join the fauststream Slack.

* https://fauststream.slack.com
* Sign-up: https://join.slack.com/t/fauststream/shared_invite/enQtNDEzMTIyMTUyNzU2LTIyMjNjY2M2YzA2OWFhMDlmMzVkODk3YTBlYThlYmZiNTUwZDJlYWZiZTdkN2Q4ZGU4NWM4YWMyNTM5MGQ5OTg

Resources
=========

Bug tracker
-----------

If you have any suggestions, bug reports, or annoyances please report them
to our issue tracker at https://github.com/robinhood/faust/issues/

License
=======

This software is licensed under the `New BSD License`. See the ``LICENSE``
file in the top distribution directory for the full license text.

.. # vim: syntax=rst expandtab tabstop=4 shiftwidth=4 shiftround

Contributing
============

Development of `Faust` happens at GitHub: https://github.com/robinhood/faust

You're highly encouraged to participate in the development
of `Faust`.

Be sure to also read the `Contributing to Faust`_ section in the
documentation.

.. _`Contributing to Faust`:
    http://faust.readthedocs.io/en/latest/contributing.html

Code of Conduct
===============

Everyone interacting in the project's code bases, issue trackers, chat rooms,
and mailing lists is expected to follow the Faust Code of Conduct.

As contributors and maintainers of these projects, and in the interest of fostering
an open and welcoming community, we pledge to respect all people who contribute
through reporting issues, posting feature requests, updating documentation,
submitting pull requests or patches, and other activities.

We are committed to making participation in these projects a harassment-free
experience for everyone, regardless of level of experience, gender,
gender identity and expression, sexual orientation, disability,
personal appearance, body size, race, ethnicity, age,
religion, or nationality.

Examples of unacceptable behavior by participants include:

* The use of sexualized language or imagery
* Personal attacks
* Trolling or insulting/derogatory comments
* Public or private harassment
* Publishing other's private information, such as physical
  or electronic addresses, without explicit permission
* Other unethical or unprofessional conduct.

Project maintainers have the right and responsibility to remove, edit, or reject
comments, commits, code, wiki edits, issues, and other contributions that are
not aligned to this Code of Conduct. By adopting this Code of Conduct,
project maintainers commit themselves to fairly and consistently applying
these principles to every aspect of managing this project. Project maintainers
who do not follow or enforce the Code of Conduct may be permanently removed from
the project team.

This code of conduct applies both within project spaces and in public spaces
when an individual is representing the project or its community.

Instances of abusive, harassing, or otherwise unacceptable behavior may be
reported by opening an issue or contacting one or more of the project maintainers.

This Code of Conduct is adapted from the Contributor Covenant,
version 1.2.0 available at http://contributor-covenant.org/version/1/2/0/.

.. |build-status| image:: https://secure.travis-ci.org/robinhood/faust.png?branch=master
    :alt: Build status
    :target: https://travis-ci.org/robinhood/faust

.. |coverage| image:: https://codecov.io/github/robinhood/faust/coverage.svg?branch=master
    :target: https://codecov.io/github/robinhood/faust?branch=master

.. |license| image:: https://img.shields.io/pypi/l/faust.svg
    :alt: BSD License
    :target: https://opensource.org/licenses/BSD-3-Clause

.. |wheel| image:: https://img.shields.io/pypi/wheel/faust.svg
    :alt: faust can be installed via wheel
    :target: http://pypi.org/project/faust/

.. |pyversion| image:: https://img.shields.io/pypi/pyversions/faust.svg
    :alt: Supported Python versions.
    :target: http://pypi.org/project/faust/

.. |pyimp| image:: https://img.shields.io/pypi/implementation/faust.svg
    :alt: Support Python implementations.
    :target: http://pypi.org/project/faust/

            

Raw data

            {
    "_id": null,
    "home_page": "http://faust.readthedocs.io/",
    "name": "twm-faust",
    "maintainer": "",
    "docs_url": null,
    "requires_python": ">=3.6.0",
    "maintainer_email": "",
    "keywords": "stream,processing,asyncio,distributed,queue,kafka",
    "author": "Robinhood Markets, Inc.",
    "author_email": "contact@fauststream.com",
    "download_url": "https://files.pythonhosted.org/packages/15/a0/e5b9467df71ec1718e5a981daa0715e52a082522e03b11841743bef9ce05/twm-faust-1.14.3.tar.gz",
    "platform": "any",
    "description": ".. XXX Need to change this image to readthedocs before release\n\n.. image:: https://raw.githubusercontent.com/robinhood/faust/8ee5e209322d9edf5bdb79b992ef986be2de4bb4/artwork/banner-alt1.png\n\n===========================\n Python Stream Processing\n===========================\n\nNote: This project is a fork of the original **Faust** stream processing library.\n\n|build-status| |coverage| |license| |wheel| |pyversion| |pyimp|\n\n:Version: 1.14.3\n:Web: http://faust.readthedocs.io/\n:Download: http://pypi.org/project/faust\n:Source: http://github.com/robinhood/faust\n:Keywords: distributed, stream, async, processing, data, queue, state management\n\n\n.. sourcecode:: python\n\n    # Python Streams\n    # Forever scalable event processing & in-memory durable K/V store;\n    # as a library w/ asyncio & static typing.\n    import faust\n\n**Faust** is a stream processing library, porting the ideas from\n`Kafka Streams`_ to Python.\n\nIt is used at `Robinhood`_ to build high performance distributed systems\nand real-time data pipelines that process billions of events every day.\n\nFaust provides both *stream processing* and *event processing*,\nsharing similarity with tools such as\n`Kafka Streams`_, `Apache Spark`_/`Storm`_/`Samza`_/`Flink`_,\n\nIt does not use a DSL, it's just Python!\nThis means you can use all your favorite Python libraries\nwhen stream processing: NumPy, PyTorch, Pandas, NLTK, Django,\nFlask, SQLAlchemy, ++\n\nFaust requires Python 3.6 or later for the new `async/await`_ syntax,\nand variable type annotations.\n\nHere's an example processing a stream of incoming orders:\n\n.. sourcecode:: python\n\n    app = faust.App('myapp', broker='kafka://localhost')\n\n    # Models describe how messages are serialized:\n    # {\"account_id\": \"3fae-...\", \"amount\": 3}\n    class Order(faust.Record):\n        account_id: str\n        amount: int\n\n    @app.agent(value_type=Order)\n    async def order(orders):\n        async for order in orders:\n            # process infinite stream of orders.\n            print(f'Order for {order.account_id}: {order.amount}')\n\nThe Agent decorator defines a \"stream processor\" that essentially\nconsumes from a Kafka topic and does something for every event it receives.\n\nThe agent is an ``async def`` function, so can also perform\nother operations asynchronously, such as web requests.\n\nThis system can persist state, acting like a database.\nTables are named distributed key/value stores you can use\nas regular Python dictionaries.\n\nTables are stored locally on each machine using a super fast\nembedded database written in C++, called `RocksDB`_.\n\nTables can also store aggregate counts that are optionally \"windowed\"\nso you can keep track\nof \"number of clicks from the last day,\" or\n\"number of clicks in the last hour.\" for example. Like `Kafka Streams`_,\nwe support tumbling, hopping and sliding windows of time, and old windows\ncan be expired to stop data from filling up.\n\nFor reliability we use a Kafka topic as \"write-ahead-log\".\nWhenever a key is changed we publish to the changelog.\nStandby nodes consume from this changelog to keep an exact replica\nof the data and enables instant recovery should any of the nodes fail.\n\nTo the user a table is just a dictionary, but data is persisted between\nrestarts and replicated across nodes so on failover other nodes can take over\nautomatically.\n\nYou can count page views by URL:\n\n.. sourcecode:: python\n\n    # data sent to 'clicks' topic sharded by URL key.\n    # e.g. key=\"http://example.com\" value=\"1\"\n    click_topic = app.topic('clicks', key_type=str, value_type=int)\n\n    # default value for missing URL will be 0 with `default=int`\n    counts = app.Table('click_counts', default=int)\n\n    @app.agent(click_topic)\n    async def count_click(clicks):\n        async for url, count in clicks.items():\n            counts[url] += count\n\nThe data sent to the Kafka topic is partitioned, which means\nthe clicks will be sharded by URL in such a way that every count\nfor the same URL will be delivered to the same Faust worker instance.\n\n\nFaust supports any type of stream data: bytes, Unicode and serialized\nstructures, but also comes with \"Models\" that use modern Python\nsyntax to describe how keys and values in streams are serialized:\n\n.. sourcecode:: python\n\n    # Order is a json serialized dictionary,\n    # having these fields:\n\n    class Order(faust.Record):\n        account_id: str\n        product_id: str\n        price: float\n        quantity: float = 1.0\n\n    orders_topic = app.topic('orders', key_type=str, value_type=Order)\n\n    @app.agent(orders_topic)\n    async def process_order(orders):\n        async for order in orders:\n            # process each order using regular Python\n            total_price = order.price * order.quantity\n            await send_order_received_email(order.account_id, order)\n\nFaust is statically typed, using the ``mypy`` type checker,\nso you can take advantage of static types when writing applications.\n\nThe Faust source code is small, well organized, and serves as a good\nresource for learning the implementation of `Kafka Streams`_.\n\n**Learn more about Faust in the** `introduction`_ **introduction page**\n    to read more about Faust, system requirements, installation instructions,\n    community resources, and more.\n\n**or go directly to the** `quickstart`_ **tutorial**\n    to see Faust in action by programming a streaming application.\n\n**then explore the** `User Guide`_\n    for in-depth information organized by topic.\n\n.. _`Robinhood`: http://robinhood.com\n.. _`async/await`:\n    https://medium.freecodecamp.org/a-guide-to-asynchronous-programming-in-python-with-asyncio-232e2afa44f6\n.. _`Kafka Streams`: https://kafka.apache.org/documentation/streams\n.. _`Apache Spark`: http://spark.apache.org\n.. _`Storm`: http://storm.apache.org\n.. _`Samza`: http://samza.apache.org\n.. _`Flink`: http://flink.apache.org\n.. _`RocksDB`: http://rocksdb.org\n\n.. _`introduction`: http://faust.readthedocs.io/en/latest/introduction.html\n\n.. _`quickstart`: http://faust.readthedocs.io/en/latest/playbooks/quickstart.html\n\n.. _`User Guide`: http://faust.readthedocs.io/en/latest/userguide/index.html\n\nFaust is...\n===========\n\n**Simple**\n    Faust is extremely easy to use. To get started using other stream processing\n    solutions you have complicated hello-world projects, and\n    infrastructure requirements.  Faust only requires Kafka,\n    the rest is just Python, so If you know Python you can already use Faust to do\n    stream processing, and it can integrate with just about anything.\n\n    Here's one of the easier applications you can make::\n\n        import faust\n\n        class Greeting(faust.Record):\n            from_name: str\n            to_name: str\n\n        app = faust.App('hello-app', broker='kafka://localhost')\n        topic = app.topic('hello-topic', value_type=Greeting)\n\n        @app.agent(topic)\n        async def hello(greetings):\n            async for greeting in greetings:\n                print(f'Hello from {greeting.from_name} to {greeting.to_name}')\n\n        @app.timer(interval=1.0)\n        async def example_sender(app):\n            await hello.send(\n                value=Greeting(from_name='Faust', to_name='you'),\n            )\n\n        if __name__ == '__main__':\n            app.main()\n\n    You're probably a bit intimidated by the `async` and `await` keywords,\n    but you don't have to know how ``asyncio`` works to use\n    Faust: just mimic the examples, and you'll be fine.\n\n    The example application starts two tasks: one is processing a stream,\n    the other is a background thread sending events to that stream.\n    In a real-life application, your system will publish\n    events to Kafka topics that your processors can consume from,\n    and the background thread is only needed to feed data into our\n    example.\n\n**Highly Available**\n    Faust is highly available and can survive network problems and server\n    crashes.  In the case of node failure, it can automatically recover,\n    and tables have standby nodes that will take over.\n\n**Distributed**\n    Start more instances of your application as needed.\n\n**Fast**\n    A single-core Faust worker instance can already process tens of thousands\n    of events every second, and we are reasonably confident that throughput will\n    increase once we can support a more optimized Kafka client.\n\n**Flexible**\n    Faust is just Python, and a stream is an infinite asynchronous iterator.\n    If you know how to use Python, you already know how to use Faust,\n    and it works with your favorite Python libraries like Django, Flask,\n    SQLAlchemy, NTLK, NumPy, SciPy, TensorFlow, etc.\n\nInstallation\n============\n\nYou can install Faust either via the Python Package Index (PyPI)\nor from source.\n\nTo install using `pip`:\n\n.. sourcecode:: console\n\n    $ pip install -U faust\n\nBundles\n-------\n\nFaust also defines a group of ``setuptools`` extensions that can be used\nto install Faust and the dependencies for a given feature.\n\nYou can specify these in your requirements or on the ``pip``\ncommand-line by using brackets. Separate multiple bundles using the comma:\n\n.. sourcecode:: console\n\n    $ pip install \"faust[rocksdb]\"\n\n    $ pip install \"faust[rocksdb,uvloop,fast,redis]\"\n\nThe following bundles are available:\n\nStores\n~~~~~~\n\n:``faust[rocksdb]``:\n    for using `RocksDB`_ for storing Faust table state.\n\n    **Recommended in production.**\n\n\nCaching\n~~~~~~~\n\n:``faust[redis]``:\n    for using `Redis_` as a simple caching backend (Memcached-style).\n\nCodecs\n~~~~~~\n\n:``faust[yaml]``:\n    for using YAML and the ``PyYAML`` library in streams.\n\nOptimization\n~~~~~~~~~~~~\n\n:``faust[fast]``:\n    for installing all the available C speedup extensions to Faust core.\n\nSensors\n~~~~~~~\n\n:``faust[datadog]``:\n    for using the Datadog Faust monitor.\n\n:``faust[statsd]``:\n    for using the Statsd Faust monitor.\n\nEvent Loops\n~~~~~~~~~~~\n\n:``faust[uvloop]``:\n    for using Faust with ``uvloop``.\n\n:``faust[eventlet]``:\n    for using Faust with ``eventlet``\n\nDebugging\n~~~~~~~~~\n\n:``faust[debug]``:\n    for using ``aiomonitor`` to connect and debug a running Faust worker.\n\n:``faust[setproctitle]``:\n    when the ``setproctitle`` module is installed the Faust worker will\n    use it to set a nicer process name in ``ps``/``top`` listings.\n    Also installed with the ``fast`` and ``debug`` bundles.\n\nDownloading and installing from source\n--------------------------------------\n\nDownload the latest version of Faust from\nhttp://pypi.org/project/faust\n\nYou can install it by doing:\n\n.. sourcecode:: console\n\n    $ tar xvfz faust-0.0.0.tar.gz\n    $ cd faust-0.0.0\n    $ python setup.py build\n    # python setup.py install\n\nThe last command must be executed as a privileged user if\nyou are not currently using a virtualenv.\n\nUsing the development version\n-----------------------------\n\nWith pip\n~~~~~~~~\n\nYou can install the latest snapshot of Faust using the following\n``pip`` command:\n\n.. sourcecode:: console\n\n    $ pip install https://github.com/robinhood/faust/zipball/master#egg=faust\n\nFAQ\n===\n\nCan I use Faust with Django/Flask/etc.?\n---------------------------------------\n\nYes! Use ``eventlet`` as a bridge to integrate with ``asyncio``.\n\n\nUsing ``eventlet``\n~~~~~~~~~~~~~~~~~~~~~~\n\nThis approach works with any blocking Python library that can work with\n``eventlet``.\n\nUsing ``eventlet`` requires you to install the ``aioeventlet`` module,\nand you can install this as a bundle along with Faust:\n\n.. sourcecode:: console\n\n    $ pip install -U faust[eventlet]\n\nThen to actually use eventlet as the event loop you have to either\nuse the ``-L <faust --loop>`` argument to the ``faust`` program:\n\n.. sourcecode:: console\n\n    $ faust -L eventlet -A myproj worker -l info\n\nor add ``import mode.loop.eventlet`` at the top of your entry point script:\n\n.. sourcecode:: python\n\n    #!/usr/bin/env python3\n    import mode.loop.eventlet  # noqa\n\n.. warning::\n\n    It's very important this is at the very top of the module,\n    and that it executes before you import libraries.\n\nCan I use Faust with Tornado?\n-----------------------------\n\nYes! Use the ``tornado.platform.asyncio`` bridge:\nhttp://www.tornadoweb.org/en/stable/asyncio.html\n\nCan I use Faust with Twisted?\n-----------------------------\n\nYes! Use the ``asyncio`` reactor implementation:\nhttps://twistedmatrix.com/documents/17.1.0/api/twisted.internet.asyncioreactor.html\n\n\nWill you support Python 2.7 or Python 3.5?\n------------------------------------------\n\nNo. Faust requires Python 3.6 or later, since it heavily uses features that were\nintroduced in Python 3.6 (`async`, `await`, variable type annotations).\n\n\nI get a maximum number of open files exceeded error by RocksDB when running a Faust app locally. How can I fix this?\n--------------------------------------------------------------------------------------------------------------------\n\nYou may need to increase the limit for the maximum number of open files. The\nfollowing post explains how to do so on OS X:\nhttps://blog.dekstroza.io/ulimit-shenanigans-on-osx-el-capitan/\n\n\nWhat kafka versions faust supports?\n---------------------------------------\n\nFaust supports kafka with version >= 0.10.\n\nGetting Help\n============\n\nSlack\n-----\n\nFor discussions about the usage, development, and future of Faust,\nplease join the fauststream Slack.\n\n* https://fauststream.slack.com\n* Sign-up: https://join.slack.com/t/fauststream/shared_invite/enQtNDEzMTIyMTUyNzU2LTIyMjNjY2M2YzA2OWFhMDlmMzVkODk3YTBlYThlYmZiNTUwZDJlYWZiZTdkN2Q4ZGU4NWM4YWMyNTM5MGQ5OTg\n\nResources\n=========\n\nBug tracker\n-----------\n\nIf you have any suggestions, bug reports, or annoyances please report them\nto our issue tracker at https://github.com/robinhood/faust/issues/\n\nLicense\n=======\n\nThis software is licensed under the `New BSD License`. See the ``LICENSE``\nfile in the top distribution directory for the full license text.\n\n.. # vim: syntax=rst expandtab tabstop=4 shiftwidth=4 shiftround\n\nContributing\n============\n\nDevelopment of `Faust` happens at GitHub: https://github.com/robinhood/faust\n\nYou're highly encouraged to participate in the development\nof `Faust`.\n\nBe sure to also read the `Contributing to Faust`_ section in the\ndocumentation.\n\n.. _`Contributing to Faust`:\n    http://faust.readthedocs.io/en/latest/contributing.html\n\nCode of Conduct\n===============\n\nEveryone interacting in the project's code bases, issue trackers, chat rooms,\nand mailing lists is expected to follow the Faust Code of Conduct.\n\nAs contributors and maintainers of these projects, and in the interest of fostering\nan open and welcoming community, we pledge to respect all people who contribute\nthrough reporting issues, posting feature requests, updating documentation,\nsubmitting pull requests or patches, and other activities.\n\nWe are committed to making participation in these projects a harassment-free\nexperience for everyone, regardless of level of experience, gender,\ngender identity and expression, sexual orientation, disability,\npersonal appearance, body size, race, ethnicity, age,\nreligion, or nationality.\n\nExamples of unacceptable behavior by participants include:\n\n* The use of sexualized language or imagery\n* Personal attacks\n* Trolling or insulting/derogatory comments\n* Public or private harassment\n* Publishing other's private information, such as physical\n  or electronic addresses, without explicit permission\n* Other unethical or unprofessional conduct.\n\nProject maintainers have the right and responsibility to remove, edit, or reject\ncomments, commits, code, wiki edits, issues, and other contributions that are\nnot aligned to this Code of Conduct. By adopting this Code of Conduct,\nproject maintainers commit themselves to fairly and consistently applying\nthese principles to every aspect of managing this project. Project maintainers\nwho do not follow or enforce the Code of Conduct may be permanently removed from\nthe project team.\n\nThis code of conduct applies both within project spaces and in public spaces\nwhen an individual is representing the project or its community.\n\nInstances of abusive, harassing, or otherwise unacceptable behavior may be\nreported by opening an issue or contacting one or more of the project maintainers.\n\nThis Code of Conduct is adapted from the Contributor Covenant,\nversion 1.2.0 available at http://contributor-covenant.org/version/1/2/0/.\n\n.. |build-status| image:: https://secure.travis-ci.org/robinhood/faust.png?branch=master\n    :alt: Build status\n    :target: https://travis-ci.org/robinhood/faust\n\n.. |coverage| image:: https://codecov.io/github/robinhood/faust/coverage.svg?branch=master\n    :target: https://codecov.io/github/robinhood/faust?branch=master\n\n.. |license| image:: https://img.shields.io/pypi/l/faust.svg\n    :alt: BSD License\n    :target: https://opensource.org/licenses/BSD-3-Clause\n\n.. |wheel| image:: https://img.shields.io/pypi/wheel/faust.svg\n    :alt: faust can be installed via wheel\n    :target: http://pypi.org/project/faust/\n\n.. |pyversion| image:: https://img.shields.io/pypi/pyversions/faust.svg\n    :alt: Supported Python versions.\n    :target: http://pypi.org/project/faust/\n\n.. |pyimp| image:: https://img.shields.io/pypi/implementation/faust.svg\n    :alt: Support Python implementations.\n    :target: http://pypi.org/project/faust/\n",
    "bugtrack_url": null,
    "license": "BSD 3-Clause",
    "summary": "Python Stream processing.",
    "version": "1.14.3",
    "project_urls": {
        "Bug Reports": "https://github.com/robinhood/faust/issues",
        "Documentation": "https://faust.readthedocs.io/",
        "Homepage": "http://faust.readthedocs.io/",
        "Source": "https://github.com/robinhood/faust"
    },
    "split_keywords": [
        "stream",
        "processing",
        "asyncio",
        "distributed",
        "queue",
        "kafka"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "db7e819311ceed69f954bc696a3408f3b7242f9591cb1792779f63e5db820367",
                "md5": "f83b6bf1dd4431929a84a13e90307766",
                "sha256": "72e92908f7d5cd39c4d7d7999f807d8387cdead9c83245d06a0811505b755189"
            },
            "downloads": -1,
            "filename": "twm_faust-1.14.3-py2.py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "f83b6bf1dd4431929a84a13e90307766",
            "packagetype": "bdist_wheel",
            "python_version": "py2.py3",
            "requires_python": ">=3.6.0",
            "size": 334929,
            "upload_time": "2024-03-06T03:44:09",
            "upload_time_iso_8601": "2024-03-06T03:44:09.164150Z",
            "url": "https://files.pythonhosted.org/packages/db/7e/819311ceed69f954bc696a3408f3b7242f9591cb1792779f63e5db820367/twm_faust-1.14.3-py2.py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "15a0e5b9467df71ec1718e5a981daa0715e52a082522e03b11841743bef9ce05",
                "md5": "b6a2b693e791ce51bf76458c8a4a5312",
                "sha256": "6c85be3809c5e7dbb4d646f46329762f13a8d4336905cdf4657a4b6fc88ee9c4"
            },
            "downloads": -1,
            "filename": "twm-faust-1.14.3.tar.gz",
            "has_sig": false,
            "md5_digest": "b6a2b693e791ce51bf76458c8a4a5312",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.6.0",
            "size": 664455,
            "upload_time": "2024-03-06T03:44:11",
            "upload_time_iso_8601": "2024-03-06T03:44:11.341826Z",
            "url": "https://files.pythonhosted.org/packages/15/a0/e5b9467df71ec1718e5a981daa0715e52a082522e03b11841743bef9ce05/twm-faust-1.14.3.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-03-06 03:44:11",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "robinhood",
    "github_project": "faust",
    "travis_ci": true,
    "coveralls": true,
    "github_actions": false,
    "tox": true,
    "lcname": "twm-faust"
}
        
Elapsed time: 0.26591s