# cl-hubeau

[](https://pypi.python.org/pypi/cl-hubeau/)

[](https://github.com/psf/black)







Simple hub'eau client for python
This package is currently under active development.
Every API on [Hub'eau](hubeau.eaufrance.fr/) will be covered by this package in
due time.
At this stage, the following APIs are covered by cl-hubeau:
* [phytopharmaceuticals transactions/vente et achat de produits phytopharmaceutiques](https://hubeau.eaufrance.fr/page/api-vente-achat-phytos)
* [watercourses flow/écoulement des cours d'eau](https://hubeau.eaufrance.fr/page/api-ecoulement)
* [drinking water quality/qualité de l'eau potable](https://hubeau.eaufrance.fr/page/api-qualite-eau-potable)
* [hydrometry/hydrométrie](https://hubeau.eaufrance.fr/page/api-hydrometrie)
* [superficial waterbodies quality/qualité des cours d'eau](https://hubeau.eaufrance.fr/page/api-qualite-cours-deau)
* [ground waterbodies quality/qualité des nappes](https://hubeau.eaufrance.fr/page/api-qualite-nappes)
* [piezometry/piézométrie](https://hubeau.eaufrance.fr/page/api-piezometrie)
For any help on available kwargs for each endpoint, please refer
directly to the documentation on hubeau (this will not be covered
by the current documentation).
Assume that each function from cl-hubeau will be consistent with
it's hub'eau counterpart, with the exception of the `size` and
`page` or `cursor` arguments (those will be set automatically by
cl-hubeau to crawl allong the results).
## Parallelization
`cl-hubeau` already uses simple multithreading pools to perform requests.
In order not to endanger the webservers and share ressources among users, a
rate limiter is set to 10 queries per second. This limiter should work fine on
any given machine, whatever the context (even with a new parallelization
overlay).
However `cl-hubeau` should **NOT** be used in containers (or pods) with
parallelization. There is currently no way of tracking the queries' rate
among multiple machines: greedy queries may end up blacklisted by the
team managing Hub'eau.
## Configuration
Starting with `pynsee 0.2.0`, no API keys are needed anymore.
## Support
In case of bugs, please open an issue [on the repo](https://github.com/tgrandje/cl-hubeau/issues).
## Contribution
Any help is welcome. Please refer to the [CONTRIBUTING file](https://github.com/tgrandje/cl-hubeau/CONTRIBUTING.md).
## Licence
GPL-3.0-or-later
## Project Status
This package is currently under active development.
## Basic examples
### Clean cache
```python
from cl_hubeau.utils import clean_all_cache
clean_all_cache()
```
### Phyopharmaceuticals transactions
4 high level functions are available (and one class for low level operations).
Note that high level functions introduce new arguments (`filter_regions` and `filter_departements`
to better target territorial data.
Get all active substances bought (uses a 30 days caching):
```python
from cl_hubeau import phytopharmaceuticals_transactions as pt
df = pt.get_all_active_substances_bought()
# or to get regional data:
df = pt.get_all_active_substances_bought(
type_territoire="Région", code_territoire="32"
)
# or to get departemantal data:
df = pt.get_all_active_substances_bought(
type_territoire="Département", filter_regions="32"
)
# or to get postcode-zoned data:
df = pt.get_all_active_substances_bought(
type_territoire="Zone postale", filter_departements=["59", "62"]
)
```
Get all phytopharmaceutical products bought (uses a 30 days caching):
```python
from cl_hubeau import phytopharmaceuticals_transactions as pt
df = pt.get_all_phytopharmaceutical_products_bought()
# or to get regional data:
df = pt.get_all_phytopharmaceutical_products_bought(
type_territoire="Région", code_territoire="32"
)
# or to get departemantal data:
df = pt.get_all_phytopharmaceutical_products_bought(
type_territoire="Département", filter_regions="32"
)
# or to get postcode-zoned data:
df = pt.get_all_phytopharmaceutical_products_bought(
type_territoire="Zone postale", filter_departements=["59", "62"]
)
```
Get all active substances sold (uses a 30 days caching):
```python
from cl_hubeau import phytopharmaceuticals_transactions as pt
df = pt.get_all_active_substances_sold()
# or to get regional data:
df = pt.get_all_active_substances_sold(
type_territoire="Région", code_territoire="32"
)
# or to get departemantal data:
df = pt.get_all_active_substances_sold(
type_territoire="Département", filter_regions="32"
)
```
Get all phytopharmaceutical products sold (uses a 30 days caching):
```python
from cl_hubeau import phytopharmaceuticals_transactions as pt
df = pt.get_all_phytopharmaceutical_products_sold()
# or to get regional data:
df = pt.get_all_phytopharmaceutical_products_sold(
type_territoire="Région", code_territoire="32"
)
# or to get departemantal data:
df = pt.get_all_phytopharmaceutical_products_sold(
type_territoire="Département", filter_regions="32"
)
```
Low level class to perform the same tasks:
Note that :
* the API is forbidding results > 20k rows and you may need inner loops
* the cache handling will be your responsibility
```python
with pt.PhytopharmaceuticalsSession() as session:
df = session.active_substances_sold(
annee_min=2010,
annee_max=2015,
code_territoire=["32"],
type_territoire="Région",
)
df = session.phytopharmaceutical_products_sold(
annee_min=2010,
annee_max=2015,
code_territoire=["32"],
type_territoire="Région",
eaj="Oui",
unite="l",
)
df = session.active_substances_bought(
annee_min=2010,
annee_max=2015,
code_territoire=["32"],
type_territoire="Région",
)
df = session.phytopharmaceutical_products_bought(
code_territoire=["32"],
type_territoire="Région",
eaj="Oui",
unite="l",
)
```
### Watercourses flow
3 high level functions are available (and one class for low level operations).
Get all stations (uses a 30 days caching):
```python
from cl_hubeau import watercourses_flow
df = watercourses_flow.get_all_stations()
```
Get all observations (uses a 30 days caching):
```python
from cl_hubeau import watercourses_flow
df = watercourses_flow.get_all_observations()
```
Note that this query is heavy, users should restrict it to a given territory when possible.
For instance, you could use :
```python
df = watercourses_flow.get_all_observations(code_region="11")
```
Get all campaigns:
```python
from cl_hubeau import watercourses_flow
df = watercourses_flow.get_all_campaigns()
```
Low level class to perform the same tasks:
Note that :
* the API is forbidding results > 20k rows and you may need inner loops
* the cache handling will be your responsibility
```python
with watercourses_flow.WatercoursesFlowSession() as session:
df = session.get_stations(code_departement="59")
df = session.get_campaigns(code_campagne=[12])
df = session.get_observations(code_station="F6640008")
```
### Drinking water quality
2 high level functions are available (and one class for low level operations).
Get all water networks (UDI) (uses a 30 days caching):
```python
from cl_hubeau import drinking_water_quality
df = drinking_water_quality.get_all_water_networks()
```
Get the sanitary controls's results for nitrates on all networks of Paris, Lyon & Marseille
(uses a 30 days caching) for nitrates
```python
networks = drinking_water_quality.get_all_water_networks()
networks = networks[
networks.nom_commune.isin(["PARIS", "MARSEILLE", "LYON"])
]["code_reseau"].unique().tolist()
df = drinking_water_quality.get_control_results(
codes_reseaux=networks,
code_parametre="1340"
)
```
Note that this query is heavy, even if this was already restricted to nitrates.
In theory, you could also query the API without specifying the substance you're tracking,
but you may hit the 20k threshold and trigger an exception.
You can also call the same function, using official city codes directly:
```python
df = drinking_water_quality.get_control_results(
codes_communes=['59350'],
code_parametre="1340"
)
```
Low level class to perform the same tasks:
Note that :
* the API is forbidding results > 20k rows and you may need inner loops
* the cache handling will be your responsibility
```python
with drinking_water_quality.DrinkingWaterQualitySession() as session:
df = session.get_cities_networks(nom_commune="LILLE")
df = session.get_control_results(code_departement='02', code_parametre="1340")
```
### Hydrometry
4 high level functions are available (and one class for low level operations).
Get all stations (uses a 30 days caching):
```python
from cl_hubeau import hydrometry
gdf = hydrometry.get_all_stations()
```
Get all sites (uses a 30 days caching):
```python
gdf = hydrometry.get_all_sites()
```
Get observations for the first 5 sites (uses a 30 days caching):
_Note that this will also work with stations (instead of sites)._
```python
df = hydrometry.get_observations(gdf["code_site"].head(5).tolist())
```
Get realtime data for the first 5 sites (no cache stored):
A small cache is stored to allow for realtime consumption (cache expires after
only 15 minutes). Please, adopt a responsible usage with this functionnality !
```python
df = hydrometry.get_realtime_observations(gdf["code_site"].head(5).tolist())
```
Low level class to perform the same tasks:
Note that :
* the API is forbidding results > 20k rows and you may need inner loops
* the cache handling will be your responsibility, noticely for realtime data
```python
with hydrometry.HydrometrySession() as session:
df = session.get_stations(code_station="K437311001")
df = session.get_sites(code_departement=['02', '59', '60', '62', '80'], format="geojson")
df = session.get_realtime_observations(code_entite="K437311001")
df = session.get_observations(code_entite="K437311001")
```
### Superficial waterbodies quality
4 high level functions are available (and one class for low level operations).
Get all stations (uses a 30 days caching):
```python
from cl_hubeau import superficial_waterbodies_quality
df = superficial_waterbodies_quality.get_all_stations()
```
Get all operations (uses a 30 days caching):
```python
from cl_hubeau import superficial_waterbodies_quality
df = superficial_waterbodies_quality.get_all_operations()
```
Note that this query is heavy, users should restrict it to a given territory.
For instance, you could use :
```python
df = superficial_waterbodies_quality.get_all_operations(code_region="11")
```
Get all environmental conditions:
```python
from cl_hubeau import superficial_waterbodies_quality
df = superficial_waterbodies_quality.get_all_environmental_conditions()
```
Note that this query is heavy, users should restrict it to a given territory.
For instance, you could use :
```python
df = superficial_waterbodies_quality.get_all_environmental_conditions(code_region="11")
```
Get all physicochemical analyses:
```python
from cl_hubeau import superficial_waterbodies_quality
df = superficial_waterbodies_quality.get_all_analyses()
```
Note that this query is heavy, users should restrict it to a given territory
and given parameters. For instance, you could use :
```python
df = superficial_waterbodies_quality.get_all_analyses(
code_departement="59",
code_parametre="1313"
)
```
Low level class to perform the same tasks:
Note that :
* the API is forbidding results > 20k rows and you may need inner loops
* the cache handling will be your responsibility
```python
with superficial_waterbodies_quality.SuperficialWaterbodiesQualitySession() as session:
df = session.get_stations(code_commune="59183")
df = session.get_operations(code_commune="59183")
df = session.get_environmental_conditions(code_commune="59183")
df = session.get_analyses(code_commune='59183', code_parametre="1340")
```
### Ground waterbodies quality
2 high level functions are available (and one class for low level operations).
Get all stations (uses a 30 days caching):
```python
from cl_hubeau import ground_water_quality
df = ground_water_quality.get_all_stations()
```
Get the tests results for nitrates :
```python
df = ground_water_quality.df = get_all_analyses(code_param="1340")
```
Note that this query is heavy, even if this was already restricted to nitrates, and that it
may fail. In theory, you could even query the API without specifying the substance
you're tracking, but you will hit the 20k threshold and trigger an exception.
In practice, you should call the same function with a territorial restriction or with
specific bss_ids.
For instance, you could use official city codes directly:
```python
df = ground_water_quality.get_all_analyses(
num_departement=["59"]
code_param="1340"
)
```
Note: a bit of caution is needed here, as the arguments are **NOT** the same
in the two endpoints. Please have a look at the documentation on
[hubeau](https://hubeau.eaufrance.fr/page/api-qualite-nappes#/qualite-nappes/analyses).
For instance, the city's number is called `"code_insee_actuel"` on analyses' endpoint
and `"code_commune"` on station's.
Low level class to perform the same tasks:
Note that :
* the API is forbidding results > 20k rows and you may need inner loops
* the cache handling will be your responsibility
```python
with ground_water_quality.GroundWaterQualitySession() as session:
df = session.get_stations(bss_id="01832B0600")
df = session.get_analyses(
bss_ids=["BSS000BMMA"],
code_param="1461",
)
```
### Piezometry
3 high level functions are available (and one class for low level operations).
Get all piezometers (uses a 30 days caching):
```python
from cl_hubeau import piezometry
gdf = piezometry.get_all_stations()
```
Get chronicles for the first 100 piezometers (uses a 30 days caching):
```python
df = piezometry.get_chronicles(gdf["code_bss"].head(100).tolist())
```
Get realtime data for the first 100 piezometers:
A small cache is stored to allow for realtime consumption (cache expires after
only 15 minutes). Please, adopt a responsible usage with this functionnality !
```python
df = get_realtime_chronicles(gdf["code_bss"].head(100).tolist())
```
Low level class to perform the same tasks:
Note that :
* the API is forbidding results > 20k rows and you may need inner loops
* the cache handling will be your responsibility, noticely for realtime data
```python
with piezometry.PiezometrySession() as session:
df = session.get_chronicles(code_bss="07548X0009/F")
df = session.get_stations(code_departement=['02', '59', '60', '62', '80'], format="geojson")
df = session.get_chronicles_real_time(code_bss="07548X0009/F")
```
### Convenience functions
In order to ease queries on hydrographic territories, some convenience functions
have been added to this module.
In these process, we are harvesting official geodatasets which are not available on hub'eau;
afterwards, simple geospatial joins are performed with the latest geodataset of french cities.
These are **convenience** tools and there **will** be approximations (geographical precision
of both datasets might not match).
#### SAGE (Schéma d'Aménagement et de Gestion des Eaux)
You can retrieve a SAGE's communal components using the following snippet:
```python
from cl_hubeau.utils import cities_for_sage
d = cities_for_sage()
```
The official geodataset is eaufrance's SAGE.
Raw data
{
"_id": null,
"home_page": "https://tgrandje.github.io/cl-hubeau",
"name": "cl-hubeau",
"maintainer": "Thomas Grandjean",
"docs_url": null,
"requires_python": "<4.0,>=3.9",
"maintainer_email": "thomas.grandjean@developpement-durable.gouv.fr",
"keywords": "france, water, hydrology",
"author": "Thomas Grandjean",
"author_email": "thomas.grandjean@developpement-durable.gouv.fr",
"download_url": "https://files.pythonhosted.org/packages/51/fe/68fccabb14984ef7328d4f43332952a22e84394b5d2623aa607d690546b6/cl_hubeau-0.8.1.tar.gz",
"platform": null,
"description": "# cl-hubeau\n\n\n[](https://pypi.python.org/pypi/cl-hubeau/)\n\n\n[](https://github.com/psf/black)\n\n\n\n\n\n\n\n\n\n\nSimple hub'eau client for python\n\nThis package is currently under active development.\nEvery API on [Hub'eau](hubeau.eaufrance.fr/) will be covered by this package in\ndue time.\n\nAt this stage, the following APIs are covered by cl-hubeau:\n* [phytopharmaceuticals transactions/vente et achat de produits phytopharmaceutiques](https://hubeau.eaufrance.fr/page/api-vente-achat-phytos)\n* [watercourses flow/\u00e9coulement des cours d'eau](https://hubeau.eaufrance.fr/page/api-ecoulement)\n* [drinking water quality/qualit\u00e9 de l'eau potable](https://hubeau.eaufrance.fr/page/api-qualite-eau-potable)\n* [hydrometry/hydrom\u00e9trie](https://hubeau.eaufrance.fr/page/api-hydrometrie)\n* [superficial waterbodies quality/qualit\u00e9 des cours d'eau](https://hubeau.eaufrance.fr/page/api-qualite-cours-deau)\n* [ground waterbodies quality/qualit\u00e9 des nappes](https://hubeau.eaufrance.fr/page/api-qualite-nappes)\n* [piezometry/pi\u00e9zom\u00e9trie](https://hubeau.eaufrance.fr/page/api-piezometrie)\n\n\nFor any help on available kwargs for each endpoint, please refer\ndirectly to the documentation on hubeau (this will not be covered\nby the current documentation).\n\nAssume that each function from cl-hubeau will be consistent with\nit's hub'eau counterpart, with the exception of the `size` and\n`page` or `cursor` arguments (those will be set automatically by\ncl-hubeau to crawl allong the results).\n\n## Parallelization\n\n`cl-hubeau` already uses simple multithreading pools to perform requests.\nIn order not to endanger the webservers and share ressources among users, a\nrate limiter is set to 10 queries per second. This limiter should work fine on\nany given machine, whatever the context (even with a new parallelization\noverlay).\n\nHowever `cl-hubeau` should **NOT** be used in containers (or pods) with\nparallelization. There is currently no way of tracking the queries' rate\namong multiple machines: greedy queries may end up blacklisted by the\nteam managing Hub'eau.\n\n\n## Configuration\n\nStarting with `pynsee 0.2.0`, no API keys are needed anymore.\n\n## Support\n\nIn case of bugs, please open an issue [on the repo](https://github.com/tgrandje/cl-hubeau/issues).\n\n## Contribution\nAny help is welcome. Please refer to the [CONTRIBUTING file](https://github.com/tgrandje/cl-hubeau/CONTRIBUTING.md).\n\n## Licence\nGPL-3.0-or-later\n\n## Project Status\n\nThis package is currently under active development.\n\n## Basic examples\n\n### Clean cache\n\n```python\nfrom cl_hubeau.utils import clean_all_cache\nclean_all_cache()\n```\n\n### Phyopharmaceuticals transactions\n\n4 high level functions are available (and one class for low level operations).\n\nNote that high level functions introduce new arguments (`filter_regions` and `filter_departements`\nto better target territorial data.\n\nGet all active substances bought (uses a 30 days caching):\n\n```python\nfrom cl_hubeau import phytopharmaceuticals_transactions as pt\ndf = pt.get_all_active_substances_bought()\n\n# or to get regional data:\ndf = pt.get_all_active_substances_bought(\n type_territoire=\"R\u00e9gion\", code_territoire=\"32\"\n )\n\n# or to get departemantal data:\ndf = pt.get_all_active_substances_bought(\n type_territoire=\"D\u00e9partement\", filter_regions=\"32\"\n )\n\n# or to get postcode-zoned data:\ndf = pt.get_all_active_substances_bought(\n type_territoire=\"Zone postale\", filter_departements=[\"59\", \"62\"]\n )\n```\n\nGet all phytopharmaceutical products bought (uses a 30 days caching):\n\n```python\nfrom cl_hubeau import phytopharmaceuticals_transactions as pt\ndf = pt.get_all_phytopharmaceutical_products_bought()\n\n# or to get regional data:\ndf = pt.get_all_phytopharmaceutical_products_bought(\n type_territoire=\"R\u00e9gion\", code_territoire=\"32\"\n )\n\n# or to get departemantal data:\ndf = pt.get_all_phytopharmaceutical_products_bought(\n type_territoire=\"D\u00e9partement\", filter_regions=\"32\"\n )\n\n# or to get postcode-zoned data:\ndf = pt.get_all_phytopharmaceutical_products_bought(\n type_territoire=\"Zone postale\", filter_departements=[\"59\", \"62\"]\n )\n```\n\nGet all active substances sold (uses a 30 days caching):\n\n```python\nfrom cl_hubeau import phytopharmaceuticals_transactions as pt\ndf = pt.get_all_active_substances_sold()\n\n# or to get regional data:\ndf = pt.get_all_active_substances_sold(\n type_territoire=\"R\u00e9gion\", code_territoire=\"32\"\n )\n\n# or to get departemantal data:\ndf = pt.get_all_active_substances_sold(\n type_territoire=\"D\u00e9partement\", filter_regions=\"32\"\n )\n```\n\nGet all phytopharmaceutical products sold (uses a 30 days caching):\n\n```python\nfrom cl_hubeau import phytopharmaceuticals_transactions as pt\ndf = pt.get_all_phytopharmaceutical_products_sold()\n\n# or to get regional data:\ndf = pt.get_all_phytopharmaceutical_products_sold(\n type_territoire=\"R\u00e9gion\", code_territoire=\"32\"\n )\n\n# or to get departemantal data:\ndf = pt.get_all_phytopharmaceutical_products_sold(\n type_territoire=\"D\u00e9partement\", filter_regions=\"32\"\n )\n```\n\nLow level class to perform the same tasks:\n\nNote that :\n\n* the API is forbidding results > 20k rows and you may need inner loops\n* the cache handling will be your responsibility\n\n```python\nwith pt.PhytopharmaceuticalsSession() as session:\n df = session.active_substances_sold(\n annee_min=2010,\n annee_max=2015,\n code_territoire=[\"32\"],\n type_territoire=\"R\u00e9gion\",\n )\n df = session.phytopharmaceutical_products_sold(\n annee_min=2010,\n annee_max=2015,\n code_territoire=[\"32\"],\n type_territoire=\"R\u00e9gion\",\n eaj=\"Oui\",\n unite=\"l\",\n )\n df = session.active_substances_bought(\n annee_min=2010,\n annee_max=2015,\n code_territoire=[\"32\"],\n type_territoire=\"R\u00e9gion\",\n )\n df = session.phytopharmaceutical_products_bought(\n code_territoire=[\"32\"],\n type_territoire=\"R\u00e9gion\",\n eaj=\"Oui\",\n unite=\"l\",\n )\n\n```\n\n### Watercourses flow\n\n3 high level functions are available (and one class for low level operations).\n\nGet all stations (uses a 30 days caching):\n\n```python\nfrom cl_hubeau import watercourses_flow\ndf = watercourses_flow.get_all_stations()\n```\n\nGet all observations (uses a 30 days caching):\n\n```python\nfrom cl_hubeau import watercourses_flow\ndf = watercourses_flow.get_all_observations()\n```\n\nNote that this query is heavy, users should restrict it to a given territory when possible.\nFor instance, you could use :\n```python\ndf = watercourses_flow.get_all_observations(code_region=\"11\")\n```\n\nGet all campaigns:\n\n```python\nfrom cl_hubeau import watercourses_flow\ndf = watercourses_flow.get_all_campaigns()\n```\n\nLow level class to perform the same tasks:\n\n\nNote that :\n\n* the API is forbidding results > 20k rows and you may need inner loops\n* the cache handling will be your responsibility\n\n```python\nwith watercourses_flow.WatercoursesFlowSession() as session:\n df = session.get_stations(code_departement=\"59\")\n df = session.get_campaigns(code_campagne=[12])\n df = session.get_observations(code_station=\"F6640008\")\n\n```\n\n### Drinking water quality\n\n2 high level functions are available (and one class for low level operations).\n\n\nGet all water networks (UDI) (uses a 30 days caching):\n\n```python\nfrom cl_hubeau import drinking_water_quality\ndf = drinking_water_quality.get_all_water_networks()\n```\n\nGet the sanitary controls's results for nitrates on all networks of Paris, Lyon & Marseille\n(uses a 30 days caching) for nitrates\n\n```python\nnetworks = drinking_water_quality.get_all_water_networks()\nnetworks = networks[\n networks.nom_commune.isin([\"PARIS\", \"MARSEILLE\", \"LYON\"])\n ][\"code_reseau\"].unique().tolist()\n\ndf = drinking_water_quality.get_control_results(\n codes_reseaux=networks,\n code_parametre=\"1340\"\n)\n```\n\nNote that this query is heavy, even if this was already restricted to nitrates.\nIn theory, you could also query the API without specifying the substance you're tracking,\nbut you may hit the 20k threshold and trigger an exception.\n\nYou can also call the same function, using official city codes directly:\n```python\ndf = drinking_water_quality.get_control_results(\n codes_communes=['59350'],\n code_parametre=\"1340\"\n)\n```\n\nLow level class to perform the same tasks:\n\n\nNote that :\n\n* the API is forbidding results > 20k rows and you may need inner loops\n* the cache handling will be your responsibility\n\n```python\nwith drinking_water_quality.DrinkingWaterQualitySession() as session:\n df = session.get_cities_networks(nom_commune=\"LILLE\")\n df = session.get_control_results(code_departement='02', code_parametre=\"1340\")\n\n```\n\n### Hydrometry\n\n4 high level functions are available (and one class for low level operations).\n\n\nGet all stations (uses a 30 days caching):\n\n```python\nfrom cl_hubeau import hydrometry\ngdf = hydrometry.get_all_stations()\n```\n\nGet all sites (uses a 30 days caching):\n\n```python\ngdf = hydrometry.get_all_sites()\n```\n\nGet observations for the first 5 sites (uses a 30 days caching):\n_Note that this will also work with stations (instead of sites)._\n\n```python\ndf = hydrometry.get_observations(gdf[\"code_site\"].head(5).tolist())\n```\n\nGet realtime data for the first 5 sites (no cache stored):\n\nA small cache is stored to allow for realtime consumption (cache expires after\nonly 15 minutes). Please, adopt a responsible usage with this functionnality !\n\n\n```python\ndf = hydrometry.get_realtime_observations(gdf[\"code_site\"].head(5).tolist())\n```\n\nLow level class to perform the same tasks:\n\n\nNote that :\n\n* the API is forbidding results > 20k rows and you may need inner loops\n* the cache handling will be your responsibility, noticely for realtime data\n\n```python\nwith hydrometry.HydrometrySession() as session:\n df = session.get_stations(code_station=\"K437311001\")\n df = session.get_sites(code_departement=['02', '59', '60', '62', '80'], format=\"geojson\")\n df = session.get_realtime_observations(code_entite=\"K437311001\")\n df = session.get_observations(code_entite=\"K437311001\")\n\n```\n\n### Superficial waterbodies quality\n\n4 high level functions are available (and one class for low level operations).\n\n\nGet all stations (uses a 30 days caching):\n\n```python\nfrom cl_hubeau import superficial_waterbodies_quality\ndf = superficial_waterbodies_quality.get_all_stations()\n```\n\nGet all operations (uses a 30 days caching):\n\n```python\nfrom cl_hubeau import superficial_waterbodies_quality\ndf = superficial_waterbodies_quality.get_all_operations()\n```\n\nNote that this query is heavy, users should restrict it to a given territory.\nFor instance, you could use :\n```python\ndf = superficial_waterbodies_quality.get_all_operations(code_region=\"11\")\n```\n\nGet all environmental conditions:\n\n```python\nfrom cl_hubeau import superficial_waterbodies_quality\ndf = superficial_waterbodies_quality.get_all_environmental_conditions()\n```\n\nNote that this query is heavy, users should restrict it to a given territory.\nFor instance, you could use :\n```python\ndf = superficial_waterbodies_quality.get_all_environmental_conditions(code_region=\"11\")\n```\n\nGet all physicochemical analyses:\n```python\nfrom cl_hubeau import superficial_waterbodies_quality\ndf = superficial_waterbodies_quality.get_all_analyses()\n```\n\nNote that this query is heavy, users should restrict it to a given territory\nand given parameters. For instance, you could use :\n```python\ndf = superficial_waterbodies_quality.get_all_analyses(\n code_departement=\"59\",\n code_parametre=\"1313\"\n )\n```\n\n\nLow level class to perform the same tasks:\n\n\nNote that :\n\n* the API is forbidding results > 20k rows and you may need inner loops\n* the cache handling will be your responsibility\n\n```python\nwith superficial_waterbodies_quality.SuperficialWaterbodiesQualitySession() as session:\n df = session.get_stations(code_commune=\"59183\")\n df = session.get_operations(code_commune=\"59183\")\n df = session.get_environmental_conditions(code_commune=\"59183\")\n df = session.get_analyses(code_commune='59183', code_parametre=\"1340\")\n\n```\n\n### Ground waterbodies quality\n\n2 high level functions are available (and one class for low level operations).\n\n\nGet all stations (uses a 30 days caching):\n\n```python\nfrom cl_hubeau import ground_water_quality\ndf = ground_water_quality.get_all_stations()\n```\n\nGet the tests results for nitrates :\n\n```python\ndf = ground_water_quality.df = get_all_analyses(code_param=\"1340\")\n```\n\nNote that this query is heavy, even if this was already restricted to nitrates, and that it\nmay fail. In theory, you could even query the API without specifying the substance\nyou're tracking, but you will hit the 20k threshold and trigger an exception.\n\nIn practice, you should call the same function with a territorial restriction or with\nspecific bss_ids.\nFor instance, you could use official city codes directly:\n\n```python\ndf = ground_water_quality.get_all_analyses(\n num_departement=[\"59\"]\n code_param=\"1340\"\n)\n```\n\nNote: a bit of caution is needed here, as the arguments are **NOT** the same\nin the two endpoints. Please have a look at the documentation on\n[hubeau](https://hubeau.eaufrance.fr/page/api-qualite-nappes#/qualite-nappes/analyses).\nFor instance, the city's number is called `\"code_insee_actuel\"` on analyses' endpoint\nand `\"code_commune\"` on station's.\n\nLow level class to perform the same tasks:\n\n\nNote that :\n\n* the API is forbidding results > 20k rows and you may need inner loops\n* the cache handling will be your responsibility\n\n```python\nwith ground_water_quality.GroundWaterQualitySession() as session:\n df = session.get_stations(bss_id=\"01832B0600\")\n df = session.get_analyses(\n bss_ids=[\"BSS000BMMA\"],\n code_param=\"1461\",\n )\n```\n\n### Piezometry\n\n3 high level functions are available (and one class for low level operations).\n\nGet all piezometers (uses a 30 days caching):\n\n```python\nfrom cl_hubeau import piezometry\ngdf = piezometry.get_all_stations()\n```\n\nGet chronicles for the first 100 piezometers (uses a 30 days caching):\n\n```python\ndf = piezometry.get_chronicles(gdf[\"code_bss\"].head(100).tolist())\n```\n\nGet realtime data for the first 100 piezometers:\n\nA small cache is stored to allow for realtime consumption (cache expires after\nonly 15 minutes). Please, adopt a responsible usage with this functionnality !\n\n```python\ndf = get_realtime_chronicles(gdf[\"code_bss\"].head(100).tolist())\n```\n\nLow level class to perform the same tasks:\n\nNote that :\n\n* the API is forbidding results > 20k rows and you may need inner loops\n* the cache handling will be your responsibility, noticely for realtime data\n\n```python\nwith piezometry.PiezometrySession() as session:\n df = session.get_chronicles(code_bss=\"07548X0009/F\")\n df = session.get_stations(code_departement=['02', '59', '60', '62', '80'], format=\"geojson\")\n df = session.get_chronicles_real_time(code_bss=\"07548X0009/F\")\n```\n\n\n\n\n\n### Convenience functions\n\nIn order to ease queries on hydrographic territories, some convenience functions\nhave been added to this module.\n\nIn these process, we are harvesting official geodatasets which are not available on hub'eau;\nafterwards, simple geospatial joins are performed with the latest geodataset of french cities.\n\nThese are **convenience** tools and there **will** be approximations (geographical precision\nof both datasets might not match).\n\n#### SAGE (Sch\u00e9ma d'Am\u00e9nagement et de Gestion des Eaux)\n\nYou can retrieve a SAGE's communal components using the following snippet:\n\n```python\n\nfrom cl_hubeau.utils import cities_for_sage\n\nd = cities_for_sage()\n```\n\nThe official geodataset is eaufrance's SAGE.\n",
"bugtrack_url": null,
"license": "GPL-3.0-or-later",
"summary": "Hubeau client to collect data from the different APIs",
"version": "0.8.1",
"project_urls": {
"Documentation": "https://tgrandje.github.io/cl-hubeau",
"Homepage": "https://tgrandje.github.io/cl-hubeau",
"Repository": "https://github.com/tgrandje/cl-hubeau/"
},
"split_keywords": [
"france",
" water",
" hydrology"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "500e2c4a79e94bec73813e886bb6b2dea48c44dd277a68e7ccd350acbf15545f",
"md5": "efc4133c5bd1e6cf6bdcd52f33f352d6",
"sha256": "a6ff8c53d08cbeca9d0837c7809a44a99d4dbc23b7759029760e7a2733c30598"
},
"downloads": -1,
"filename": "cl_hubeau-0.8.1-py3-none-any.whl",
"has_sig": false,
"md5_digest": "efc4133c5bd1e6cf6bdcd52f33f352d6",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": "<4.0,>=3.9",
"size": 46790,
"upload_time": "2025-07-09T09:49:53",
"upload_time_iso_8601": "2025-07-09T09:49:53.868793Z",
"url": "https://files.pythonhosted.org/packages/50/0e/2c4a79e94bec73813e886bb6b2dea48c44dd277a68e7ccd350acbf15545f/cl_hubeau-0.8.1-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "51fe68fccabb14984ef7328d4f43332952a22e84394b5d2623aa607d690546b6",
"md5": "14ce4f0e636eda1a08aad0156cc1bade",
"sha256": "777a8071f441e9098744153430619ab7880fbccf72042b2cb468352e2aaa6cbf"
},
"downloads": -1,
"filename": "cl_hubeau-0.8.1.tar.gz",
"has_sig": false,
"md5_digest": "14ce4f0e636eda1a08aad0156cc1bade",
"packagetype": "sdist",
"python_version": "source",
"requires_python": "<4.0,>=3.9",
"size": 34112,
"upload_time": "2025-07-09T09:49:55",
"upload_time_iso_8601": "2025-07-09T09:49:55.173630Z",
"url": "https://files.pythonhosted.org/packages/51/fe/68fccabb14984ef7328d4f43332952a22e84394b5d2623aa607d690546b6/cl_hubeau-0.8.1.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2025-07-09 09:49:55",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "tgrandje",
"github_project": "cl-hubeau",
"travis_ci": false,
"coveralls": false,
"github_actions": true,
"lcname": "cl-hubeau"
}