Name | h2o-mlops-scoring-client JSON |
Version |
0.2.1b1
JSON |
| download |
home_page | |
Summary | A Python client library to simplify robust mini-batch scoring against an H2O MLOps scoring endpoint. |
upload_time | 2024-03-11 21:45:45 |
maintainer | |
docs_url | None |
author | H2O.ai |
requires_python | >=3.8,<4 |
license | Proprietary License |
keywords |
|
VCS |
|
bugtrack_url |
|
requirements |
No requirements were recorded.
|
Travis-CI |
No Travis.
|
coveralls test coverage |
No coveralls.
|
# H2O MLOps Scoring Client
A Python client library to simplify robust mini-batch scoring against an H2O MLOps scoring endpoint. It can run on your local PC, a stand alone server, Databricks, or a Spark 3 cluster.
Scoring Pandas data frames is as easy as:
```sh
pip install h2o-mlops-scoring-client
```
```python
import h2o_mlops_scoring_client
scores_df = h2o_mlops_scoring_client.score_data_frame(
mlops_endpoint_url="https://.../model/score",
id_column="ID",
data_frame=df,
)
```
Scoring from a source to a sink is also possible through `pyspark`:
```sh
pip install h2o-mlops-scoring-client[PYSPARK]
```
```python
import h2o_mlops_scoring_client
h2o_mlops_scoring_client.score_source_sink(
mlops_endpoint_url="https://.../model/score",
id_column="ID",
source_data="s3a://...",
source_format=h2o_mlops_scoring_client.Format.CSV,
sink_location="s3a://...",
sink_format=h2o_mlops_scoring_client.Format.PARQUET,
sink_write_mode=h2o_mlops_scoring_client.WriteMode.OVERWRITE
)
```
## Installation
### Requirements
- Linux or Mac OS (Windows is not supported)
- Java (only required for `pyspark` installs)
- Python 3.8 or greater
### Install from PyPI
```sh
pip install h2o-mlops-scoring-client
```
`pyspark` is no longer included in a default install. To include `pyspark`:
```sh
pip install h2o-mlops-scoring-client[PYSPARK]
```
## FAQ
### When should I use the MLOps Scoring Client?
Use when batch scoring processing (authenticating and connecting to source or sink, file/data processing or conversions, etc.) can happen external to H2O AI Cloud but you want to stay within the H2O MLOps workflow (projects, scoring, registry, monitoring, etc.).
### Where does scoring take place?
As the batch scoring processing occurs, the data is sent to an H2O MLOps deployment for scoring. The scores are then returned for the batch scoring processing to complete.
### What Source/Sinks are supported?
The MLOps scoring client can support many source/sinks, including:
- ADLS Gen 2
- Databases with a JDBC driver
- Local file system
- GBQ
- S3
- Snowflake
### What file types are supported?
The MLOps scoring client can read and write:
- CSV
- Parquet
- ORC
- BigQuery tables
- JDBC queries
- JDBC tables
- Snowflake queries
- Snowflake tables
If there's a file type you would like to see supported, please let us know.
### I want model monitoring for batch scoring, can I do that?
Yes. The MLOps Scoring Client uses MLOps scoring endpoints which are automatically monitored.
### Is a Spark installation required?
No. If you're scoring Pandas data frames, then no extra Spark install or configuration is needed. If you want to connect to an external source or sink, you'll need to install `pyspark` and do a small amount of configuration.
Raw data
{
"_id": null,
"home_page": "",
"name": "h2o-mlops-scoring-client",
"maintainer": "",
"docs_url": null,
"requires_python": ">=3.8,<4",
"maintainer_email": "",
"keywords": "",
"author": "H2O.ai",
"author_email": "support@h2o.ai",
"download_url": "",
"platform": null,
"description": "# H2O MLOps Scoring Client\n\nA Python client library to simplify robust mini-batch scoring against an H2O MLOps scoring endpoint. It can run on your local PC, a stand alone server, Databricks, or a Spark 3 cluster.\n\nScoring Pandas data frames is as easy as:\n\n```sh\npip install h2o-mlops-scoring-client\n```\n\n```python\nimport h2o_mlops_scoring_client\n\n\nscores_df = h2o_mlops_scoring_client.score_data_frame(\n mlops_endpoint_url=\"https://.../model/score\",\n id_column=\"ID\",\n data_frame=df,\n)\n```\n\nScoring from a source to a sink is also possible through `pyspark`:\n\n```sh\npip install h2o-mlops-scoring-client[PYSPARK]\n```\n\n```python\nimport h2o_mlops_scoring_client\n\n\nh2o_mlops_scoring_client.score_source_sink(\n mlops_endpoint_url=\"https://.../model/score\",\n id_column=\"ID\",\n source_data=\"s3a://...\",\n source_format=h2o_mlops_scoring_client.Format.CSV,\n sink_location=\"s3a://...\",\n sink_format=h2o_mlops_scoring_client.Format.PARQUET,\n sink_write_mode=h2o_mlops_scoring_client.WriteMode.OVERWRITE\n)\n```\n\n## Installation\n\n### Requirements\n\n- Linux or Mac OS (Windows is not supported)\n- Java (only required for `pyspark` installs)\n- Python 3.8 or greater\n\n### Install from PyPI\n\n```sh\npip install h2o-mlops-scoring-client\n```\n\n`pyspark` is no longer included in a default install. To include `pyspark`:\n\n```sh\npip install h2o-mlops-scoring-client[PYSPARK]\n```\n\n## FAQ\n\n### When should I use the MLOps Scoring Client?\n\nUse when batch scoring processing (authenticating and connecting to source or sink, file/data processing or conversions, etc.) can happen external to H2O AI Cloud but you want to stay within the H2O MLOps workflow (projects, scoring, registry, monitoring, etc.).\n\n### Where does scoring take place?\n\nAs the batch scoring processing occurs, the data is sent to an H2O MLOps deployment for scoring. The scores are then returned for the batch scoring processing to complete.\n\n### What Source/Sinks are supported?\n\nThe MLOps scoring client can support many source/sinks, including:\n\n- ADLS Gen 2\n- Databases with a JDBC driver\n- Local file system\n- GBQ\n- S3\n- Snowflake\n\n### What file types are supported?\n\nThe MLOps scoring client can read and write:\n\n- CSV\n- Parquet\n- ORC\n- BigQuery tables\n- JDBC queries\n- JDBC tables\n- Snowflake queries\n- Snowflake tables\n\nIf there's a file type you would like to see supported, please let us know.\n\n### I want model monitoring for batch scoring, can I do that?\n\nYes. The MLOps Scoring Client uses MLOps scoring endpoints which are automatically monitored.\n\n### Is a Spark installation required?\n\nNo. If you're scoring Pandas data frames, then no extra Spark install or configuration is needed. If you want to connect to an external source or sink, you'll need to install `pyspark` and do a small amount of configuration.\n",
"bugtrack_url": null,
"license": "Proprietary License",
"summary": "A Python client library to simplify robust mini-batch scoring against an H2O MLOps scoring endpoint.",
"version": "0.2.1b1",
"project_urls": {
"Documentation": "https://docs.h2o.ai/mlops/mlops-scoring-client/overview"
},
"split_keywords": [],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "548ec8ffa57a7a1e045236e1e94c6ba4ce27d3a43d412634e5e6a13228193952",
"md5": "c49b25bf131a068aed2bc558a8997558",
"sha256": "d76c9504ccce231f585a41deba01617652029d773562b21dadc387f4239a3540"
},
"downloads": -1,
"filename": "h2o_mlops_scoring_client-0.2.1b1-py3-none-any.whl",
"has_sig": false,
"md5_digest": "c49b25bf131a068aed2bc558a8997558",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": ">=3.8,<4",
"size": 12384,
"upload_time": "2024-03-11T21:45:45",
"upload_time_iso_8601": "2024-03-11T21:45:45.399945Z",
"url": "https://files.pythonhosted.org/packages/54/8e/c8ffa57a7a1e045236e1e94c6ba4ce27d3a43d412634e5e6a13228193952/h2o_mlops_scoring_client-0.2.1b1-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2024-03-11 21:45:45",
"github": false,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"lcname": "h2o-mlops-scoring-client"
}