adoc-airflow-plugin


Nameadoc-airflow-plugin JSON
Version 2.0.0 PyPI version JSON
download
home_pageNone
SummaryAcceldata Airflow Listener Plugin
upload_time2024-10-03 11:17:58
maintainerNone
docs_urlNone
authoracceldata
requires_python>=3.8
licenseNone
keywords acceldata
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            ## Overview
The Acceldata Listener plugin integrates Airflow DAGs for automatic observation in ADOC.

## Features
The plugin performs the following actions without requiring additional code in your Airflow DAG, unless you disable instrumentation through environment variables.

- **When the DAG starts:**
  - It creates the pipeline if it does not already exist in ADOC.
  - It creates a new pipeline run in ADOC.

- **When a TaskInstance starts:**
  - It creates jobs in ADOC for each of the Airflow operators used in the task.
  - It constructs job input nodes based on the upstream tasks.
  - It creates a span and associates it with the jobs.
  - It emits span events with metadata.

- **When a TaskInstance is completed:**
  - It emits span events with metadata.
  - It ends the spans with either success or failure.

- **When the DAG is completed:**
  - It updates the pipeline run with success or failure in ADOC.

## Prerequisites
Ensure the following applications are installed on your system:
- Python V3.8.0 and above ([Download Python](https://www.python.org/downloads/))
- Airflow V2.5.0 and above ([Apache Airflow](https://airflow.apache.org/docs/apache-airflow/stable/installation.html))

API keys are essential for authentication when making calls to ADOC. You can generate API keys in the ADOC UI's Admin Central by visiting the [API Keys](https://docs.acceldata.io/documentation/api-keys) section.

## Configuration

### Plugin Environment Variables
The `adoc_airflow_plugin` uses the `acceldata-sdk` to push data to the ADOC backend.

**Mandatory Environment Variables:**
The ADOC client requires the following environment variables:
- `TORCH_CATALOG_URL`: The URL of your ADOC Server instance.
- `TORCH_ACCESS_KEY`: The API access key generated from the ADOC UI.
- `TORCH_SECRET_KEY`: The API secret key generated from the ADOC UI.

**Optional Environment Variables:**
By default, all DAGs are observed. However, the following environment variables can be set to modify this behavior.

**Note:** The variables for ignoring or observing DAGs are mutually exclusive.

- If the following environment variables match specific DAG IDs, those DAGs will be ignored from observation, while all other DAGs will still be observed:
  - `DAGIDS_TO_IGNORE`: Comma-separated list of DAG IDs to ignore.
  - `DAGIDS_REGEX_TO_IGNORE`: Regular expression pattern for DAG IDs to ignore.

- If the following environment variables match specific DAG IDs, only those DAGs will be observed, and all others will be ignored:
  - `DAGIDS_TO_OBSERVE`: Comma-separated list of DAG IDs to observe.
  - `DAGIDS_REGEX_TO_OBSERVE`: Regular expression pattern for DAG IDs to observe.

            

Raw data

            {
    "_id": null,
    "home_page": null,
    "name": "adoc-airflow-plugin",
    "maintainer": null,
    "docs_url": null,
    "requires_python": ">=3.8",
    "maintainer_email": null,
    "keywords": "acceldata",
    "author": "acceldata",
    "author_email": null,
    "download_url": "https://files.pythonhosted.org/packages/a9/02/4cd853b2a2ae38880377654857a012ac2d697ef6085b8b6e44192ef95344/adoc_airflow_plugin-2.0.0.tar.gz",
    "platform": null,
    "description": "## Overview\nThe Acceldata Listener plugin integrates Airflow DAGs for automatic observation in ADOC.\n\n## Features\nThe plugin performs the following actions without requiring additional code in your Airflow DAG, unless you disable instrumentation through environment variables.\n\n- **When the DAG starts:**\n  - It creates the pipeline if it does not already exist in ADOC.\n  - It creates a new pipeline run in ADOC.\n\n- **When a TaskInstance starts:**\n  - It creates jobs in ADOC for each of the Airflow operators used in the task.\n  - It constructs job input nodes based on the upstream tasks.\n  - It creates a span and associates it with the jobs.\n  - It emits span events with metadata.\n\n- **When a TaskInstance is completed:**\n  - It emits span events with metadata.\n  - It ends the spans with either success or failure.\n\n- **When the DAG is completed:**\n  - It updates the pipeline run with success or failure in ADOC.\n\n## Prerequisites\nEnsure the following applications are installed on your system:\n- Python V3.8.0 and above ([Download Python](https://www.python.org/downloads/))\n- Airflow V2.5.0 and above ([Apache Airflow](https://airflow.apache.org/docs/apache-airflow/stable/installation.html))\n\nAPI keys are essential for authentication when making calls to ADOC. You can generate API keys in the ADOC UI's Admin Central by visiting the [API Keys](https://docs.acceldata.io/documentation/api-keys) section.\n\n## Configuration\n\n### Plugin Environment Variables\nThe `adoc_airflow_plugin` uses the `acceldata-sdk` to push data to the ADOC backend.\n\n**Mandatory Environment Variables:**\nThe ADOC client requires the following environment variables:\n- `TORCH_CATALOG_URL`: The URL of your ADOC Server instance.\n- `TORCH_ACCESS_KEY`: The API access key generated from the ADOC UI.\n- `TORCH_SECRET_KEY`: The API secret key generated from the ADOC UI.\n\n**Optional Environment Variables:**\nBy default, all DAGs are observed. However, the following environment variables can be set to modify this behavior.\n\n**Note:** The variables for ignoring or observing DAGs are mutually exclusive.\n\n- If the following environment variables match specific DAG IDs, those DAGs will be ignored from observation, while all other DAGs will still be observed:\n  - `DAGIDS_TO_IGNORE`: Comma-separated list of DAG IDs to ignore.\n  - `DAGIDS_REGEX_TO_IGNORE`: Regular expression pattern for DAG IDs to ignore.\n\n- If the following environment variables match specific DAG IDs, only those DAGs will be observed, and all others will be ignored:\n  - `DAGIDS_TO_OBSERVE`: Comma-separated list of DAG IDs to observe.\n  - `DAGIDS_REGEX_TO_OBSERVE`: Regular expression pattern for DAG IDs to observe.\n",
    "bugtrack_url": null,
    "license": null,
    "summary": "Acceldata Airflow Listener Plugin",
    "version": "2.0.0",
    "project_urls": null,
    "split_keywords": [
        "acceldata"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "a9024cd853b2a2ae38880377654857a012ac2d697ef6085b8b6e44192ef95344",
                "md5": "cbdf1383119aeee0fab9a1a0772e2f0f",
                "sha256": "eecdc98039cad60c95303aaab8674465c40c1cf656714aa096e9f78b9ce08e8a"
            },
            "downloads": -1,
            "filename": "adoc_airflow_plugin-2.0.0.tar.gz",
            "has_sig": false,
            "md5_digest": "cbdf1383119aeee0fab9a1a0772e2f0f",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.8",
            "size": 10100,
            "upload_time": "2024-10-03T11:17:58",
            "upload_time_iso_8601": "2024-10-03T11:17:58.810065Z",
            "url": "https://files.pythonhosted.org/packages/a9/02/4cd853b2a2ae38880377654857a012ac2d697ef6085b8b6e44192ef95344/adoc_airflow_plugin-2.0.0.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-10-03 11:17:58",
    "github": false,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "lcname": "adoc-airflow-plugin"
}
        
Elapsed time: 0.30171s