# aws-parquet
<br>
[![PyPI version shields.io](https://img.shields.io/pypi/v/aws-parquet.svg)](https://pypi.org/project/aws-parquet/)
[![PyPI license](https://img.shields.io/pypi/l/aws-parquet.svg)](https://pypi.python.org/pypi/)
[![PyPI pyversions](https://img.shields.io/pypi/pyversions/aws-parquet.svg)](https://pypi.python.org/pypi/aws-parquet/)
[![Downloads](https://pepy.tech/badge/aws-parquet/month)](https://pepy.tech/project/aws-parquet)
[![Downloads](https://pepy.tech/badge/aws-parquet)](https://pepy.tech/project/aws-parquet)
`aws-parquet` is a toolkit than enables working with parquet datasets on AWS. It handles AWS S3 reads/writes, AWS Glue catalog updates and AWS Athena queries by providing a simple and intuitive interface.
## Motivation
The goal is to provide a simple and intuitive interface to create and manage parquet datasets on AWS.
`aws-parquet` makes use of the following tools:
- [awswrangler](https://aws-sdk-pandas.readthedocs.io/en/stable/) as an AWS SDK for pandas
- [pandera](https://pandera.readthedocs.io/en/stable/) for pandas-based data validation
- [typeguard](https://typeguard.readthedocs.io/en/stable/userguide.html) and [pydantic](https://docs.pydantic.dev/latest/) for runtime type checking
## Features
`aws-parquet` provides a `ParquetDataset` class that enables the following operations:
- create a parquet dataset that will get registered in AWS Glue
- append new data to the dataset and update the AWS Glue catalog
- read a partition of the dataset and perform proper schema validation and type casting
- overwrite data in the dataset after performing proper schema validation and type casting
- delete a partition of the dataset and update the AWS Glue catalog
- query the dataset using AWS Athena
## How to setup
Using pip:
```bash
pip install aws_parquet
```
## How to use
Create a parquet dataset that will get registered in AWS Glue
```python
import os
from aws_parquet import ParquetDataset
import pandas as pd
import pandera as pa
from pandera.typing import Series
# define your pandera schema model
class MyDatasetSchemaModel(pa.SchemaModel):
col1: Series[int] = pa.Field(nullable=False, ge=0, lt=10)
col2: Series[pa.DateTime]
col3: Series[float]
# configuration
database = "default"
bucket_name = os.environ["AWS_S3_BUCKET"]
table_name = "foo_bar"
path = f"s3://{bucket_name}/{table_name}/"
partition_cols = ["col1", "col2"]
schema = MyDatasetSchemaModel.to_schema()
# create the dataset
dataset = ParquetDataset(
database=database,
table=table_name,
partition_cols=partition_cols,
path=path,
pandera_schema=schema,
)
dataset.create()
```
Append new data to the dataset
```python
df = pd.DataFrame({
"col1": [1, 2, 3],
"col2": ["2021-01-01", "2021-01-02", "2021-01-03"],
"col3": [1.0, 2.0, 3.0]
})
dataset.update(df)
```
Read a partition of the dataset
```python
df = dataset.read({"col2": "2021-01-01"})
```
Overwrite data in the dataset
```python
df_overwrite = pd.DataFrame({
"col1": [1, 2, 3],
"col2": ["2021-01-01", "2021-01-02", "2021-01-03"],
"col3": [4.0, 5.0, 6.0]
})
dataset.update(df_overwrite, overwrite=True)
```
Query the dataset using AWS Athena
```python
df = dataset.query("SELECT col1 FROM foo_bar")
```
Delete a partition of the dataset
```python
dataset.delete({"col1": 1, "col2": "2021-01-01"})
```
Delete the dataset in its entirety
```python
dataset.delete()
```
Raw data
{
"_id": null,
"home_page": "https://github.com/marwan116/aws-parquet/",
"name": "aws-parquet",
"maintainer": "",
"docs_url": null,
"requires_python": ">=3.8,<=3.11",
"maintainer_email": "",
"keywords": "pandas,aws,parquet",
"author": "Marwan Sarieddine",
"author_email": "sarieddine.marwan@gmail.com",
"download_url": "",
"platform": null,
"description": "# aws-parquet\n\n<br>\n\n[![PyPI version shields.io](https://img.shields.io/pypi/v/aws-parquet.svg)](https://pypi.org/project/aws-parquet/)\n[![PyPI license](https://img.shields.io/pypi/l/aws-parquet.svg)](https://pypi.python.org/pypi/)\n[![PyPI pyversions](https://img.shields.io/pypi/pyversions/aws-parquet.svg)](https://pypi.python.org/pypi/aws-parquet/)\n[![Downloads](https://pepy.tech/badge/aws-parquet/month)](https://pepy.tech/project/aws-parquet)\n[![Downloads](https://pepy.tech/badge/aws-parquet)](https://pepy.tech/project/aws-parquet)\n\n`aws-parquet` is a toolkit than enables working with parquet datasets on AWS. It handles AWS S3 reads/writes, AWS Glue catalog updates and AWS Athena queries by providing a simple and intuitive interface.\n\n## Motivation\n\nThe goal is to provide a simple and intuitive interface to create and manage parquet datasets on AWS.\n\n`aws-parquet` makes use of the following tools: \n- [awswrangler](https://aws-sdk-pandas.readthedocs.io/en/stable/) as an AWS SDK for pandas\n- [pandera](https://pandera.readthedocs.io/en/stable/) for pandas-based data validation\n- [typeguard](https://typeguard.readthedocs.io/en/stable/userguide.html) and [pydantic](https://docs.pydantic.dev/latest/) for runtime type checking\n\n## Features\n`aws-parquet` provides a `ParquetDataset` class that enables the following operations:\n\n- create a parquet dataset that will get registered in AWS Glue\n- append new data to the dataset and update the AWS Glue catalog\n- read a partition of the dataset and perform proper schema validation and type casting\n- overwrite data in the dataset after performing proper schema validation and type casting\n- delete a partition of the dataset and update the AWS Glue catalog\n- query the dataset using AWS Athena\n\n\n## How to setup\n\nUsing pip:\n\n```bash\npip install aws_parquet\n```\n\n## How to use\n\nCreate a parquet dataset that will get registered in AWS Glue\n\n```python\nimport os\n\nfrom aws_parquet import ParquetDataset\nimport pandas as pd\nimport pandera as pa\nfrom pandera.typing import Series\n\n# define your pandera schema model\nclass MyDatasetSchemaModel(pa.SchemaModel):\n col1: Series[int] = pa.Field(nullable=False, ge=0, lt=10)\n col2: Series[pa.DateTime]\n col3: Series[float]\n\n# configuration\ndatabase = \"default\"\nbucket_name = os.environ[\"AWS_S3_BUCKET\"]\ntable_name = \"foo_bar\"\npath = f\"s3://{bucket_name}/{table_name}/\"\npartition_cols = [\"col1\", \"col2\"]\nschema = MyDatasetSchemaModel.to_schema()\n\n# create the dataset\ndataset = ParquetDataset(\n database=database,\n table=table_name,\n partition_cols=partition_cols,\n path=path,\n pandera_schema=schema,\n)\n\ndataset.create()\n```\nAppend new data to the dataset\n\n```python\ndf = pd.DataFrame({\n \"col1\": [1, 2, 3],\n \"col2\": [\"2021-01-01\", \"2021-01-02\", \"2021-01-03\"],\n \"col3\": [1.0, 2.0, 3.0]\n})\n\ndataset.update(df)\n```\n\nRead a partition of the dataset\n\n```python\ndf = dataset.read({\"col2\": \"2021-01-01\"})\n```\n\nOverwrite data in the dataset\n\n```python\ndf_overwrite = pd.DataFrame({\n \"col1\": [1, 2, 3],\n \"col2\": [\"2021-01-01\", \"2021-01-02\", \"2021-01-03\"],\n \"col3\": [4.0, 5.0, 6.0]\n})\ndataset.update(df_overwrite, overwrite=True)\n```\n\nQuery the dataset using AWS Athena\n\n```python\ndf = dataset.query(\"SELECT col1 FROM foo_bar\")\n```\n\nDelete a partition of the dataset\n\n```python\ndataset.delete({\"col1\": 1, \"col2\": \"2021-01-01\"})\n```\n\n\nDelete the dataset in its entirety\n\n```python\ndataset.delete()\n```\n",
"bugtrack_url": null,
"license": "MIT",
"summary": "An object-oriented interface for defining parquet datasets for AWS built on top of awswrangler and pandera",
"version": "0.5.0",
"project_urls": {
"Documentation": "https://aws-parquet.readthedocs.io/en/latest/",
"Homepage": "https://github.com/marwan116/aws-parquet/",
"Repository": "https://github.com/marwan116/aws-parquet/"
},
"split_keywords": [
"pandas",
"aws",
"parquet"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "f72ec5b657fda0e9be441f9c941fe42ba992587145c72287c96db351dd9d278b",
"md5": "463aedae7b6ee5a7c564f20c373ffbe0",
"sha256": "e9ec49bf2f11c73cab98e1d186c1fc118496681c33a62914cd52365932734be2"
},
"downloads": -1,
"filename": "aws_parquet-0.5.0-py3-none-any.whl",
"has_sig": false,
"md5_digest": "463aedae7b6ee5a7c564f20c373ffbe0",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": ">=3.8,<=3.11",
"size": 10655,
"upload_time": "2023-06-19T10:30:25",
"upload_time_iso_8601": "2023-06-19T10:30:25.270684Z",
"url": "https://files.pythonhosted.org/packages/f7/2e/c5b657fda0e9be441f9c941fe42ba992587145c72287c96db351dd9d278b/aws_parquet-0.5.0-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2023-06-19 10:30:25",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "marwan116",
"github_project": "aws-parquet",
"travis_ci": false,
"coveralls": false,
"github_actions": false,
"lcname": "aws-parquet"
}