qualytics-cli


Namequalytics-cli JSON
Version 0.1.19 PyPI version JSON
download
home_pagehttps://www.qualytics.co/
SummaryQualytics CLI
upload_time2024-09-24 05:37:24
maintainerNone
docs_urlNone
authorQualytics
requires_python>=3.7
licenseMIT
keywords qualytics data quality
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            # Qualytics CLI

This is a CLI tool for working with the Qualytics API. With this tool, you can manage your configurations, export checks, import checks, and more. It's built on top of the Typer CLI framework and uses the Rich library for enhanced terminal outputs.

## Requirements

- Python 3.7+
- Packages:
  - `typer`
  - `os`
  - `json`
  - `requests`
  - `urllib3`
  - `re`
  - `rich`

## Installation

```bash
pip install qualytics-cli
```

## Usage

### Help

```bash
qualytics --help
```

### Initializing the Configuration

You can set up your Qualytics URL and token using the `init` command:

```bash
qualytics init --url "https://your-qualytics.qualytics.io/" --token "YOUR_TOKEN_HERE"
```

| Option  | Type | Description                                           | Default | Required |
|---------|------|-------------------------------------------------------|---------|----------|
| `--url` | TEXT | The URL to be set. Example: https://your-qualytics.qualytics.io/ | None    | Yes      |
| `--token` | TEXT | The token to be set.                                 | None    | Yes      |

### Qualytics init help

```bash
qualytics init --help
```

### Display Configuration

To view the currently saved configuration:

```bash
qualytics show-config
```

### Export Checks

You can export checks to a file using the `checks export` command:

```bash
qualytics checks export --datastore DATASTORE_ID [--containers CONTAINER_IDS] [--tags TAG_NAMES] [--output LOCATION_TO_BE_EXPORTED]
```

By default, it saves the exported checks to `./qualytics/data_checks.json`. However, you can specify a different output path with the `--output` option.

| Option         | Type            | Description                                             | Default                            | Required |
|----------------|-----------------|---------------------------------------------------------|------------------------------------|----------|
| `--datastore`  | INTEGER         | Datastore ID                                            | None                               | Yes      |
| `--containers` | List of INTEGER | Containers IDs                                          | None                               | No       |
| `--tags`       | List of TEXT    | Tag names                                               | None                               | No       |
| `--status`      | List of TEXT   | Status `Active`, `Draft` or `Archived`                  | None                               | No       |
| `--output`     | TEXT            | Output file path   | ./qualytics/data_checks.json       | No                                 | No       |

### Export Check Templates

You can export check templates to the `_export_check_templates` table to an enrichment datastore.

```bash
qualytics checks export-templates --enrichment_datastore_id ENRICHMENT_DATASTORE_ID [--check_templates CHECK_TEMPLATE_IDS]
```

| Option                   | Type     | Description                                                                | Required |
|--------------------------|----------|----------------------------------------------------------------------------|----------|
| `--enrichment_datastore_id` | INTEGER  | The ID of the enrichment datastore where check templates will be exported. | Yes      |
| `--check_templates`       | TEXT     | Comma-separated list of check template IDs or array-like format. Example: "1, 2, 3" or "[1,2,3]".| No       |

### Import Checks

To import checks from a file:

```bash
qualytics checks import --datastore DATASTORE_ID_LIST [--input LOCATION_FROM_THE_EXPORT]
```

By default, it reads the checks from `./qualytics/data_checks.json`. You can specify a different input file with the `--input` option.

**Note**: Any errors encountered during the importing of checks will be logged in `./qualytics/errors.log`.

| Option       | Type | Description                                                                  | Default                       | Required |
|--------------|------|------------------------------------------------------------------------------|-------------------------------|----------|
| `--datastore`| TEXT | Comma-separated list of Datastore IDs or array-like format. Example: 1,2,3,4,5 or "[1,2,3,4,5]" | None | Yes      |
| `--input`    | TEXT | Input file path                                                              | HOME/.qualytics/data_checks.json | No       |



### Import Check Templates

You can import check templates from a file using the `checks import-templates` command:

```bash
qualytics checks import-templates [--input LOCATION_OF_CHECK_TEMPLATES]
```

By default, it reads the check templates from `./qualytics/data_checks_template.json`. You can specify a different input file with the `--input` option.

| Option    | Type | Description                  | Default                               | Required |
|-----------|------|------------------------------|---------------------------------------|----------|
| `--input` | TEXT | Input file path               | ./qualytics/data_checks_template.json | No       |

### Schedule Metadata Export

Allows you to schedule exports of metadata from your datastores using a specified crontab expression.

```bash
qualytics schedule export-metadata --crontab "CRONTAB_EXPRESSION" --datastore "DATASTORE_ID" [--containers "CONTAINER_IDS"] --options "EXPORT_OPTIONS"
```

| Option       | Type | Description                                                          | Required |
|--------------|------|----------------------------------------------------------------------|----------|
| `--crontab`  | TEXT | Crontab expression inside quotes, specifying when the task should run. Example: "0 * * * *" | Yes      |
| `--datastore`| TEXT | The datastore ID                                                     | Yes      |
| `--containers`| TEXT | Comma-separated list of container IDs or array-like format. Example: "1, 2, 3" or "[1,2,3]" | No       |
| `--options`  | TEXT | Comma-separated list of options to export or "all". Example: "anomalies, checks, field-profiles" | Yes      |

### Run a Catalog Operation on a Datastore

Allows you to trigger a catalog operation on any current datastore (requires admin permissions on the datastore).

```bash
qualytics run catalog --datastore "DATASTORE_ID_LIST" --include "INCLUDE_LIST" --prune --recreate --background
```

| Option         | Type | Description                                                                                         | Required |
|----------------|------|-----------------------------------------------------------------------------------------------------|----------|
| `--datastore`  | TEXT | Comma-separated list of Datastore IDs or array-like format. Example: 1,2,3,4,5 or "[1,2,3,4,5]"     | Yes      |
| `--include`    | TEXT | Comma-separated list of include types or array-like format. Example: "table,view" or "[table,view]" | No       |
| `--prune`      | BOOL | Prune the operation. Do not include if you want prune == false                                      | No       |
| `--recreate`   | BOOL | Recreate the operation. Do not include if you want recreate == false                                | No       |
| `--background` | BOOL | Starts the catalog but does not wait for the operation to finish                                    | No       |

### Run a Profile Operation on a Datastore

Allows you to trigger a profile operation on any current datastore (requires admin permissions on the datastore).

```bash
qualytics run profile --datastore "DATASTORE_ID_LIST" --container_names "CONTAINER_NAMES_LIST" --container_tags "CONTAINER_TAGS_LIST"
--inference_threshold "INFERENCE_THRESHOLD" --infer_as_draft --max_records_analyzed_per_partition "MAX_RECORDS_ANALYZED_PER_PARTITION"
--max_count_testing_sample "MAX_COUNT_TESTING_SAMPLE" --percent_testing_threshold "PERCENT_TESTING_THRESHOLD" --high_correlation_threshold
"HIGH_CORRELATION_THRESHOLD" --greater_than_time "GREATER_THAN_TIME" --greater_than_batch "GREATER_THAN_BATCH" --histogram_max_distinct_values
"HISTOGRAM_MAX_DISTINCT_VALUES" --background
```

| Option                                 | Type     | Description                                                                                                                                      | Required |
|----------------------------------------|----------|--------------------------------------------------------------------------------------------------------------------------------------------------|----------|
| `--datastore`                          | TEXT     | Comma-separated list of Datastore IDs or array-like format. Example: 1,2,3,4,5 or "[1,2,3,4,5]"                                                  | Yes      |
| `--container_names`                    | TEXT     | Comma-separated list of container names or array-like format. Example: "container1,container2" or "[container1,container2]"                      | No       |
| `--container_tags`                     | TEXT     | Comma-separated list of container tags or array-like format. Example: "tag1,tag2" or "[tag1,tag2]"                                               | No       |
| `--inference_threshold`                | INT      | Inference quality checks threshold in profile from 0 to 5. Do not include if inference_threshold == 0                                             | No       |
| `--infer_as_draft`                     | BOOL     | Infer all quality checks in profile as DRAFT. Do not include if you want infer_as_draft == False                                                 | No       |
| `--max_records_analyzed_per_partition` | INT      | Number of max records analyzed per partition                                                                                                     | No       |
| `--max_count_testing_sample`           | INT      | The number of records accumulated during profiling for validation of inferred checks. Capped at 100,000                                           | No       |
| `--percent_testing_threshold`          | FLOAT    | Percent of testing threshold                                                                                                                     | No       |
| `--high_correlation_threshold`         | FLOAT    | Number of correlation threshold                                                                                                                  | No       |
| `--greater_than_time`                  | DATETIME | Only include rows where the incremental field's value is greater than this time. Use one of these formats %Y-%m-%dT%H:%M:%S or %Y-%m-%d %H:%M:%S | No       |
| `--greater_than_batch`                 | FLOAT    | Only include rows where the incremental field's value is greater than this number                                                                | No       |
| `--histogram_max_distinct_values`      | INT      | Number of max distinct values in the histogram                                                                                                   | No       |
| `--background`                         | BOOL     | Starts the profile operation but does not wait for the operation to finish                                                                       | No       |


### Run a Scan Operation on a Datastore

Allows you to trigger a scan operation on a datastore (requires admin permissions on the datastore).

```bash
qualytics run scan --datastore "DATASTORE_ID_LIST" --container_names "CONTAINER_NAMES_LIST" --container_tags "CONTAINER_TAGS_LIST"
--incremental --remediation --max_records_analyzed_per_partition "MAX_RECORDS_ANALYZED_PER_PARTITION" --enrichment_source_record_limit
--greater_than_time "GREATER_THAN_TIME" --greater_than_batch "GREATER_THAN_BATCH" --background
```

| Option                                 | Type     | Description                                                                                                                                      | Required |
|----------------------------------------|----------|--------------------------------------------------------------------------------------------------------------------------------------------------|----------|
| `--datastore`                          | TEXT     | Comma-separated list of Datastore IDs or array-like format. Example: 1,2,3,4,5 or "[1,2,3,4,5]"                                                  | Yes      |
| `--container_names`                    | TEXT     | Comma-separated list of container names or array-like format. Example: "container1,container2" or "[container1,container2]"                      | No       |
| `--container_tags`                     | TEXT     | Comma-separated list of container tags or array-like format. Example: "tag1,tag2" or "[tag1,tag2]"                                                | No       |
| `--incremental`                        | BOOL     | Process only new or updated records since the last incremental scan                                                                              | No       |
| `--remediation`                        | TEXT     | Replication strategy for source tables in the enrichment datastore. Either 'append', 'overwrite', or 'none'                                      | No       |
| `--max_records_analyzed_per_partition` | INT      | Number of max records analyzed per partition. Value must be greater than or equal to 0                                                           | No       |
| `--enrichment_source_record_limit`     | INT      | Limit of enrichment source records per run. Value must be greater than or equal to -1                                                            | No       |
| `--greater_than_time`                  | DATETIME | Only include rows where the incremental field's value is greater than this time. Use one of these formats %Y-%m-%dT%H:%M:%S or %Y-%m-%d %H:%M:%S | No       |
| `--greater_than_batch`                 | FLOAT    | Only include rows where the incremental field's value is greater than this number                                                                | No       |
| `--background`                         | BOOL     | Starts the scan operation but does not wait for the operation to finish                                                                          | No       |

### Check Operation Status

Allows a user to check an operation's status. Useful if a user triggered an operation but had it running in the background.

```bash
qualytics operation check_status --ids "OPERATION_IDS"
```

| Option  | Type     | Description                                                                                                               | Required |
|---------|----------|---------------------------------------------------------------------------------------------------------------------------|----------|
| `--ids` | TEXT     | Comma-separated list of Operation IDs or array-like format. Example: 1,2,3,4,5 or "[1,2,3,4,5]"                           | Yes      |

            

Raw data

            {
    "_id": null,
    "home_page": "https://www.qualytics.co/",
    "name": "qualytics-cli",
    "maintainer": null,
    "docs_url": null,
    "requires_python": ">=3.7",
    "maintainer_email": null,
    "keywords": "Qualytics, Data Quality",
    "author": "Qualytics",
    "author_email": "devops@qualytics.co",
    "download_url": "https://files.pythonhosted.org/packages/08/5b/6c99272eaffd9991974e403d40f3169b72960260ffd92feb7550b01d86bd/qualytics_cli-0.1.19.tar.gz",
    "platform": null,
    "description": "# Qualytics CLI\n\nThis is a CLI tool for working with the Qualytics API. With this tool, you can manage your configurations, export checks, import checks, and more. It's built on top of the Typer CLI framework and uses the Rich library for enhanced terminal outputs.\n\n## Requirements\n\n- Python 3.7+\n- Packages:\n  - `typer`\n  - `os`\n  - `json`\n  - `requests`\n  - `urllib3`\n  - `re`\n  - `rich`\n\n## Installation\n\n```bash\npip install qualytics-cli\n```\n\n## Usage\n\n### Help\n\n```bash\nqualytics --help\n```\n\n### Initializing the Configuration\n\nYou can set up your Qualytics URL and token using the `init` command:\n\n```bash\nqualytics init --url \"https://your-qualytics.qualytics.io/\" --token \"YOUR_TOKEN_HERE\"\n```\n\n| Option  | Type | Description                                           | Default | Required |\n|---------|------|-------------------------------------------------------|---------|----------|\n| `--url` | TEXT | The URL to be set. Example: https://your-qualytics.qualytics.io/ | None    | Yes      |\n| `--token` | TEXT | The token to be set.                                 | None    | Yes      |\n\n### Qualytics init help\n\n```bash\nqualytics init --help\n```\n\n### Display Configuration\n\nTo view the currently saved configuration:\n\n```bash\nqualytics show-config\n```\n\n### Export Checks\n\nYou can export checks to a file using the `checks export` command:\n\n```bash\nqualytics checks export --datastore DATASTORE_ID [--containers CONTAINER_IDS] [--tags TAG_NAMES] [--output LOCATION_TO_BE_EXPORTED]\n```\n\nBy default, it saves the exported checks to `./qualytics/data_checks.json`. However, you can specify a different output path with the `--output` option.\n\n| Option         | Type            | Description                                             | Default                            | Required |\n|----------------|-----------------|---------------------------------------------------------|------------------------------------|----------|\n| `--datastore`  | INTEGER         | Datastore ID                                            | None                               | Yes      |\n| `--containers` | List of INTEGER | Containers IDs                                          | None                               | No       |\n| `--tags`       | List of TEXT    | Tag names                                               | None                               | No       |\n| `--status`      | List of TEXT   | Status `Active`, `Draft` or `Archived`                  | None                               | No       |\n| `--output`     | TEXT            | Output file path   | ./qualytics/data_checks.json       | No                                 | No       |\n\n### Export Check Templates\n\nYou can export check templates to the `_export_check_templates` table to an enrichment datastore.\n\n```bash\nqualytics checks export-templates --enrichment_datastore_id ENRICHMENT_DATASTORE_ID [--check_templates CHECK_TEMPLATE_IDS]\n```\n\n| Option                   | Type     | Description                                                                | Required |\n|--------------------------|----------|----------------------------------------------------------------------------|----------|\n| `--enrichment_datastore_id` | INTEGER  | The ID of the enrichment datastore where check templates will be exported. | Yes      |\n| `--check_templates`       | TEXT     | Comma-separated list of check template IDs or array-like format. Example: \"1, 2, 3\" or \"[1,2,3]\".| No       |\n\n### Import Checks\n\nTo import checks from a file:\n\n```bash\nqualytics checks import --datastore DATASTORE_ID_LIST [--input LOCATION_FROM_THE_EXPORT]\n```\n\nBy default, it reads the checks from `./qualytics/data_checks.json`. You can specify a different input file with the `--input` option.\n\n**Note**: Any errors encountered during the importing of checks will be logged in `./qualytics/errors.log`.\n\n| Option       | Type | Description                                                                  | Default                       | Required |\n|--------------|------|------------------------------------------------------------------------------|-------------------------------|----------|\n| `--datastore`| TEXT | Comma-separated list of Datastore IDs or array-like format. Example: 1,2,3,4,5 or \"[1,2,3,4,5]\" | None | Yes      |\n| `--input`    | TEXT | Input file path                                                              | HOME/.qualytics/data_checks.json | No       |\n\n\n\n### Import Check Templates\n\nYou can import check templates from a file using the `checks import-templates` command:\n\n```bash\nqualytics checks import-templates [--input LOCATION_OF_CHECK_TEMPLATES]\n```\n\nBy default, it reads the check templates from `./qualytics/data_checks_template.json`. You can specify a different input file with the `--input` option.\n\n| Option    | Type | Description                  | Default                               | Required |\n|-----------|------|------------------------------|---------------------------------------|----------|\n| `--input` | TEXT | Input file path               | ./qualytics/data_checks_template.json | No       |\n\n### Schedule Metadata Export\n\nAllows you to schedule exports of metadata from your datastores using a specified crontab expression.\n\n```bash\nqualytics schedule export-metadata --crontab \"CRONTAB_EXPRESSION\" --datastore \"DATASTORE_ID\" [--containers \"CONTAINER_IDS\"] --options \"EXPORT_OPTIONS\"\n```\n\n| Option       | Type | Description                                                          | Required |\n|--------------|------|----------------------------------------------------------------------|----------|\n| `--crontab`  | TEXT | Crontab expression inside quotes, specifying when the task should run. Example: \"0 * * * *\" | Yes      |\n| `--datastore`| TEXT | The datastore ID                                                     | Yes      |\n| `--containers`| TEXT | Comma-separated list of container IDs or array-like format. Example: \"1, 2, 3\" or \"[1,2,3]\" | No       |\n| `--options`  | TEXT | Comma-separated list of options to export or \"all\". Example: \"anomalies, checks, field-profiles\" | Yes      |\n\n### Run a Catalog Operation on a Datastore\n\nAllows you to trigger a catalog operation on any current datastore (requires admin permissions on the datastore).\n\n```bash\nqualytics run catalog --datastore \"DATASTORE_ID_LIST\" --include \"INCLUDE_LIST\" --prune --recreate --background\n```\n\n| Option         | Type | Description                                                                                         | Required |\n|----------------|------|-----------------------------------------------------------------------------------------------------|----------|\n| `--datastore`  | TEXT | Comma-separated list of Datastore IDs or array-like format. Example: 1,2,3,4,5 or \"[1,2,3,4,5]\"     | Yes      |\n| `--include`    | TEXT | Comma-separated list of include types or array-like format. Example: \"table,view\" or \"[table,view]\" | No       |\n| `--prune`      | BOOL | Prune the operation. Do not include if you want prune == false                                      | No       |\n| `--recreate`   | BOOL | Recreate the operation. Do not include if you want recreate == false                                | No       |\n| `--background` | BOOL | Starts the catalog but does not wait for the operation to finish                                    | No       |\n\n### Run a Profile Operation on a Datastore\n\nAllows you to trigger a profile operation on any current datastore (requires admin permissions on the datastore).\n\n```bash\nqualytics run profile --datastore \"DATASTORE_ID_LIST\" --container_names \"CONTAINER_NAMES_LIST\" --container_tags \"CONTAINER_TAGS_LIST\"\n--inference_threshold \"INFERENCE_THRESHOLD\" --infer_as_draft --max_records_analyzed_per_partition \"MAX_RECORDS_ANALYZED_PER_PARTITION\"\n--max_count_testing_sample \"MAX_COUNT_TESTING_SAMPLE\" --percent_testing_threshold \"PERCENT_TESTING_THRESHOLD\" --high_correlation_threshold\n\"HIGH_CORRELATION_THRESHOLD\" --greater_than_time \"GREATER_THAN_TIME\" --greater_than_batch \"GREATER_THAN_BATCH\" --histogram_max_distinct_values\n\"HISTOGRAM_MAX_DISTINCT_VALUES\" --background\n```\n\n| Option                                 | Type     | Description                                                                                                                                      | Required |\n|----------------------------------------|----------|--------------------------------------------------------------------------------------------------------------------------------------------------|----------|\n| `--datastore`                          | TEXT     | Comma-separated list of Datastore IDs or array-like format. Example: 1,2,3,4,5 or \"[1,2,3,4,5]\"                                                  | Yes      |\n| `--container_names`                    | TEXT     | Comma-separated list of container names or array-like format. Example: \"container1,container2\" or \"[container1,container2]\"                      | No       |\n| `--container_tags`                     | TEXT     | Comma-separated list of container tags or array-like format. Example: \"tag1,tag2\" or \"[tag1,tag2]\"                                               | No       |\n| `--inference_threshold`                | INT      | Inference quality checks threshold in profile from 0 to 5. Do not include if inference_threshold == 0                                             | No       |\n| `--infer_as_draft`                     | BOOL     | Infer all quality checks in profile as DRAFT. Do not include if you want infer_as_draft == False                                                 | No       |\n| `--max_records_analyzed_per_partition` | INT      | Number of max records analyzed per partition                                                                                                     | No       |\n| `--max_count_testing_sample`           | INT      | The number of records accumulated during profiling for validation of inferred checks. Capped at 100,000                                           | No       |\n| `--percent_testing_threshold`          | FLOAT    | Percent of testing threshold                                                                                                                     | No       |\n| `--high_correlation_threshold`         | FLOAT    | Number of correlation threshold                                                                                                                  | No       |\n| `--greater_than_time`                  | DATETIME | Only include rows where the incremental field's value is greater than this time. Use one of these formats %Y-%m-%dT%H:%M:%S or %Y-%m-%d %H:%M:%S | No       |\n| `--greater_than_batch`                 | FLOAT    | Only include rows where the incremental field's value is greater than this number                                                                | No       |\n| `--histogram_max_distinct_values`      | INT      | Number of max distinct values in the histogram                                                                                                   | No       |\n| `--background`                         | BOOL     | Starts the profile operation but does not wait for the operation to finish                                                                       | No       |\n\n\n### Run a Scan Operation on a Datastore\n\nAllows you to trigger a scan operation on a datastore (requires admin permissions on the datastore).\n\n```bash\nqualytics run scan --datastore \"DATASTORE_ID_LIST\" --container_names \"CONTAINER_NAMES_LIST\" --container_tags \"CONTAINER_TAGS_LIST\"\n--incremental --remediation --max_records_analyzed_per_partition \"MAX_RECORDS_ANALYZED_PER_PARTITION\" --enrichment_source_record_limit\n--greater_than_time \"GREATER_THAN_TIME\" --greater_than_batch \"GREATER_THAN_BATCH\" --background\n```\n\n| Option                                 | Type     | Description                                                                                                                                      | Required |\n|----------------------------------------|----------|--------------------------------------------------------------------------------------------------------------------------------------------------|----------|\n| `--datastore`                          | TEXT     | Comma-separated list of Datastore IDs or array-like format. Example: 1,2,3,4,5 or \"[1,2,3,4,5]\"                                                  | Yes      |\n| `--container_names`                    | TEXT     | Comma-separated list of container names or array-like format. Example: \"container1,container2\" or \"[container1,container2]\"                      | No       |\n| `--container_tags`                     | TEXT     | Comma-separated list of container tags or array-like format. Example: \"tag1,tag2\" or \"[tag1,tag2]\"                                                | No       |\n| `--incremental`                        | BOOL     | Process only new or updated records since the last incremental scan                                                                              | No       |\n| `--remediation`                        | TEXT     | Replication strategy for source tables in the enrichment datastore. Either 'append', 'overwrite', or 'none'                                      | No       |\n| `--max_records_analyzed_per_partition` | INT      | Number of max records analyzed per partition. Value must be greater than or equal to 0                                                           | No       |\n| `--enrichment_source_record_limit`     | INT      | Limit of enrichment source records per run. Value must be greater than or equal to -1                                                            | No       |\n| `--greater_than_time`                  | DATETIME | Only include rows where the incremental field's value is greater than this time. Use one of these formats %Y-%m-%dT%H:%M:%S or %Y-%m-%d %H:%M:%S | No       |\n| `--greater_than_batch`                 | FLOAT    | Only include rows where the incremental field's value is greater than this number                                                                | No       |\n| `--background`                         | BOOL     | Starts the scan operation but does not wait for the operation to finish                                                                          | No       |\n\n### Check Operation Status\n\nAllows a user to check an operation's status. Useful if a user triggered an operation but had it running in the background.\n\n```bash\nqualytics operation check_status --ids \"OPERATION_IDS\"\n```\n\n| Option  | Type     | Description                                                                                                               | Required |\n|---------|----------|---------------------------------------------------------------------------------------------------------------------------|----------|\n| `--ids` | TEXT     | Comma-separated list of Operation IDs or array-like format. Example: 1,2,3,4,5 or \"[1,2,3,4,5]\"                           | Yes      |\n",
    "bugtrack_url": null,
    "license": "MIT",
    "summary": "Qualytics CLI",
    "version": "0.1.19",
    "project_urls": {
        "GitHub": "https://github.com/Qualytics/qualytics-cli",
        "Homepage": "https://www.qualytics.co/",
        "Userguide": "https://qualytics.github.io/userguide/"
    },
    "split_keywords": [
        "qualytics",
        " data quality"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "3d9091c73c8f910ec07271aff8cb8b0e199c86d3a52bbabf355aa985a265f161",
                "md5": "432c21f72bae40a5134f021e3af6b575",
                "sha256": "6efce944d64e2e46553fb612968baa3790eb95ad5c506ddf50834926eaf5bfe2"
            },
            "downloads": -1,
            "filename": "qualytics_cli-0.1.19-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "432c21f72bae40a5134f021e3af6b575",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.7",
            "size": 17169,
            "upload_time": "2024-09-24T05:37:22",
            "upload_time_iso_8601": "2024-09-24T05:37:22.504820Z",
            "url": "https://files.pythonhosted.org/packages/3d/90/91c73c8f910ec07271aff8cb8b0e199c86d3a52bbabf355aa985a265f161/qualytics_cli-0.1.19-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "085b6c99272eaffd9991974e403d40f3169b72960260ffd92feb7550b01d86bd",
                "md5": "c3fe0d21062d59c23453386a5ff0d7a2",
                "sha256": "812fa59552e4c63705df87e6149a7c361e7324f5191c4599c05a73d9d688bd54"
            },
            "downloads": -1,
            "filename": "qualytics_cli-0.1.19.tar.gz",
            "has_sig": false,
            "md5_digest": "c3fe0d21062d59c23453386a5ff0d7a2",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.7",
            "size": 19219,
            "upload_time": "2024-09-24T05:37:24",
            "upload_time_iso_8601": "2024-09-24T05:37:24.072818Z",
            "url": "https://files.pythonhosted.org/packages/08/5b/6c99272eaffd9991974e403d40f3169b72960260ffd92feb7550b01d86bd/qualytics_cli-0.1.19.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-09-24 05:37:24",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "Qualytics",
    "github_project": "qualytics-cli",
    "travis_ci": false,
    "coveralls": false,
    "github_actions": true,
    "requirements": [],
    "lcname": "qualytics-cli"
}
        
Elapsed time: 8.88396s