Name | exso JSON |
Version |
0.0.0
JSON |
| download |
home_page | |
Summary | A powerful data colllector |
upload_time | 2023-05-28 16:09:21 |
maintainer | |
docs_url | None |
author | |
requires_python | >=3.10 |
license | Attribution-NonCommercial-NoDerivatives 4.0 International ======================================================================= Creative Commons Corporation ("Creative Commons") is not a law firm and does not provide legal services or legal advice. Distribution of Creative Commons public licenses does not create a lawyer-client or other relationship. Creative Commons makes its licenses and related information available on an "as-is" basis. Creative Commons gives no warranties regarding its licenses, any material licensed under their terms and conditions, or any related information. Creative Commons disclaims all liability for damages resulting from their use to the fullest extent possible. Using Creative Commons Public Licenses Creative Commons public licenses provide a standard set of terms and conditions that creators and other rights holders may use to share original works of authorship and other material subject to copyright and certain other rights specified in the public license below. The following considerations are for informational purposes only, are not exhaustive, and do not form part of our licenses. Considerations for licensors: Our public licenses are intended for use by those authorized to give the public permission to use material in ways otherwise restricted by copyright and certain other rights. Our licenses are irrevocable. Licensors should read and understand the terms and conditions of the license they choose before applying it. Licensors should also secure all rights necessary before applying our licenses so that the public can reuse the material as expected. Licensors should clearly mark any material not subject to the license. This includes other CC- licensed material, or material used under an exception or limitation to copyright. More considerations for licensors: wiki.creativecommons.org/Considerations_for_licensors Considerations for the public: By using one of our public licenses, a licensor grants the public permission to use the licensed material under specified terms and conditions. If the licensor's permission is not necessary for any reason--for example, because of any applicable exception or limitation to copyright--then that use is not regulated by the license. Our licenses grant only permissions under copyright and certain other rights that a licensor has authority to grant. Use of the licensed material may still be restricted for other reasons, including because others have copyright or other rights in the material. A licensor may make special requests, such as asking that all changes be marked or described. Although not required by our licenses, you are encouraged to respect those requests where reasonable. More considerations for the public: wiki.creativecommons.org/Considerations_for_licensees ======================================================================= Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International Public License By exercising the Licensed Rights (defined below), You accept and agree to be bound by the terms and conditions of this Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International Public License ("Public License"). To the extent this Public License may be interpreted as a contract, You are granted the Licensed Rights in consideration of Your acceptance of these terms and conditions, and the Licensor grants You such rights in consideration of benefits the Licensor receives from making the Licensed Material available under these terms and conditions. Section 1 -- Definitions. a. Adapted Material means material subject to Copyright and Similar Rights that is derived from or based upon the Licensed Material and in which the Licensed Material is translated, altered, arranged, transformed, or otherwise modified in a manner requiring permission under the Copyright and Similar Rights held by the Licensor. For purposes of this Public License, where the Licensed Material is a musical work, performance, or sound recording, Adapted Material is always produced where the Licensed Material is synched in timed relation with a moving image. b. Copyright and Similar Rights means copyright and/or similar rights closely related to copyright including, without limitation, performance, broadcast, sound recording, and Sui Generis Database Rights, without regard to how the rights are labeled or categorized. For purposes of this Public License, the rights specified in Section 2(b)(1)-(2) are not Copyright and Similar Rights. c. Effective Technological Measures means those measures that, in the absence of proper authority, may not be circumvented under laws fulfilling obligations under Article 11 of the WIPO Copyright Treaty adopted on December 20, 1996, and/or similar international agreements. d. Exceptions and Limitations means fair use, fair dealing, and/or any other exception or limitation to Copyright and Similar Rights that applies to Your use of the Licensed Material. e. Licensed Material means the artistic or literary work, database, or other material to which the Licensor applied this Public License. f. Licensed Rights means the rights granted to You subject to the terms and conditions of this Public License, which are limited to all Copyright and Similar Rights that apply to Your use of the Licensed Material and that the Licensor has authority to license. g. Licensor means the individual(s) or entity(ies) granting rights under this Public License. h. NonCommercial means not primarily intended for or directed towards commercial advantage or monetary compensation. For purposes of this Public License, the exchange of the Licensed Material for other material subject to Copyright and Similar Rights by digital file-sharing or similar means is NonCommercial provided there is no payment of monetary compensation in connection with the exchange. i. Share means to provide material to the public by any means or process that requires permission under the Licensed Rights, such as reproduction, public display, public performance, distribution, dissemination, communication, or importation, and to make material available to the public including in ways that members of the public may access the material from a place and at a time individually chosen by them. j. Sui Generis Database Rights means rights other than copyright resulting from Directive 96/9/EC of the European Parliament and of the Council of 11 March 1996 on the legal protection of databases, as amended and/or succeeded, as well as other essentially equivalent rights anywhere in the world. k. You means the individual or entity exercising the Licensed Rights under this Public License. Your has a corresponding meaning. Section 2 -- Scope. a. License grant. 1. Subject to the terms and conditions of this Public License, the Licensor hereby grants You a worldwide, royalty-free, non-sublicensable, non-exclusive, irrevocable license to exercise the Licensed Rights in the Licensed Material to: a. reproduce and Share the Licensed Material, in whole or in part, for NonCommercial purposes only; and b. produce and reproduce, but not Share, Adapted Material for NonCommercial purposes only. 2. Exceptions and Limitations. For the avoidance of doubt, where Exceptions and Limitations apply to Your use, this Public License does not apply, and You do not need to comply with its terms and conditions. 3. Term. The term of this Public License is specified in Section 6(a). 4. Media and formats; technical modifications allowed. The Licensor authorizes You to exercise the Licensed Rights in all media and formats whether now known or hereafter created, and to make technical modifications necessary to do so. The Licensor waives and/or agrees not to assert any right or authority to forbid You from making technical modifications necessary to exercise the Licensed Rights, including technical modifications necessary to circumvent Effective Technological Measures. For purposes of this Public License, simply making modifications authorized by this Section 2(a) (4) never produces Adapted Material. 5. Downstream recipients. a. Offer from the Licensor -- Licensed Material. Every recipient of the Licensed Material automatically receives an offer from the Licensor to exercise the Licensed Rights under the terms and conditions of this Public License. b. No downstream restrictions. You may not offer or impose any additional or different terms or conditions on, or apply any Effective Technological Measures to, the Licensed Material if doing so restricts exercise of the Licensed Rights by any recipient of the Licensed Material. 6. No endorsement. Nothing in this Public License constitutes or may be construed as permission to assert or imply that You are, or that Your use of the Licensed Material is, connected with, or sponsored, endorsed, or granted official status by, the Licensor or others designated to receive attribution as provided in Section 3(a)(1)(A)(i). b. Other rights. 1. Moral rights, such as the right of integrity, are not licensed under this Public License, nor are publicity, privacy, and/or other similar personality rights; however, to the extent possible, the Licensor waives and/or agrees not to assert any such rights held by the Licensor to the limited extent necessary to allow You to exercise the Licensed Rights, but not otherwise. 2. Patent and trademark rights are not licensed under this Public License. 3. To the extent possible, the Licensor waives any right to collect royalties from You for the exercise of the Licensed Rights, whether directly or through a collecting society under any voluntary or waivable statutory or compulsory licensing scheme. In all other cases the Licensor expressly reserves any right to collect such royalties, including when the Licensed Material is used other than for NonCommercial purposes. Section 3 -- License Conditions. Your exercise of the Licensed Rights is expressly made subject to the following conditions. a. Attribution. 1. If You Share the Licensed Material, You must: a. retain the following if it is supplied by the Licensor with the Licensed Material: i. identification of the creator(s) of the Licensed Material and any others designated to receive attribution, in any reasonable manner requested by the Licensor (including by pseudonym if designated); ii. a copyright notice; iii. a notice that refers to this Public License; iv. a notice that refers to the disclaimer of warranties; v. a URI or hyperlink to the Licensed Material to the extent reasonably practicable; b. indicate if You modified the Licensed Material and retain an indication of any previous modifications; and c. indicate the Licensed Material is licensed under this Public License, and include the text of, or the URI or hyperlink to, this Public License. For the avoidance of doubt, You do not have permission under this Public License to Share Adapted Material. 2. You may satisfy the conditions in Section 3(a)(1) in any reasonable manner based on the medium, means, and context in which You Share the Licensed Material. For example, it may be reasonable to satisfy the conditions by providing a URI or hyperlink to a resource that includes the required information. 3. If requested by the Licensor, You must remove any of the information required by Section 3(a)(1)(A) to the extent reasonably practicable. Section 4 -- Sui Generis Database Rights. Where the Licensed Rights include Sui Generis Database Rights that apply to Your use of the Licensed Material: a. for the avoidance of doubt, Section 2(a)(1) grants You the right to extract, reuse, reproduce, and Share all or a substantial portion of the contents of the database for NonCommercial purposes only and provided You do not Share Adapted Material; b. if You include all or a substantial portion of the database contents in a database in which You have Sui Generis Database Rights, then the database in which You have Sui Generis Database Rights (but not its individual contents) is Adapted Material; and c. You must comply with the conditions in Section 3(a) if You Share all or a substantial portion of the contents of the database. For the avoidance of doubt, this Section 4 supplements and does not replace Your obligations under this Public License where the Licensed Rights include other Copyright and Similar Rights. Section 5 -- Disclaimer of Warranties and Limitation of Liability. a. UNLESS OTHERWISE SEPARATELY UNDERTAKEN BY THE LICENSOR, TO THE EXTENT POSSIBLE, THE LICENSOR OFFERS THE LICENSED MATERIAL AS-IS AND AS-AVAILABLE, AND MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND CONCERNING THE LICENSED MATERIAL, WHETHER EXPRESS, IMPLIED, STATUTORY, OR OTHER. THIS INCLUDES, WITHOUT LIMITATION, WARRANTIES OF TITLE, MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, NON-INFRINGEMENT, ABSENCE OF LATENT OR OTHER DEFECTS, ACCURACY, OR THE PRESENCE OR ABSENCE OF ERRORS, WHETHER OR NOT KNOWN OR DISCOVERABLE. WHERE DISCLAIMERS OF WARRANTIES ARE NOT ALLOWED IN FULL OR IN PART, THIS DISCLAIMER MAY NOT APPLY TO YOU. b. TO THE EXTENT POSSIBLE, IN NO EVENT WILL THE LICENSOR BE LIABLE TO YOU ON ANY LEGAL THEORY (INCLUDING, WITHOUT LIMITATION, NEGLIGENCE) OR OTHERWISE FOR ANY DIRECT, SPECIAL, INDIRECT, INCIDENTAL, CONSEQUENTIAL, PUNITIVE, EXEMPLARY, OR OTHER LOSSES, COSTS, EXPENSES, OR DAMAGES ARISING OUT OF THIS PUBLIC LICENSE OR USE OF THE LICENSED MATERIAL, EVEN IF THE LICENSOR HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH LOSSES, COSTS, EXPENSES, OR DAMAGES. WHERE A LIMITATION OF LIABILITY IS NOT ALLOWED IN FULL OR IN PART, THIS LIMITATION MAY NOT APPLY TO YOU. c. The disclaimer of warranties and limitation of liability provided above shall be interpreted in a manner that, to the extent possible, most closely approximates an absolute disclaimer and waiver of all liability. Section 6 -- Term and Termination. a. This Public License applies for the term of the Copyright and Similar Rights licensed here. However, if You fail to comply with this Public License, then Your rights under this Public License terminate automatically. b. Where Your right to use the Licensed Material has terminated under Section 6(a), it reinstates: 1. automatically as of the date the violation is cured, provided it is cured within 30 days of Your discovery of the violation; or 2. upon express reinstatement by the Licensor. For the avoidance of doubt, this Section 6(b) does not affect any right the Licensor may have to seek remedies for Your violations of this Public License. c. For the avoidance of doubt, the Licensor may also offer the Licensed Material under separate terms or conditions or stop distributing the Licensed Material at any time; however, doing so will not terminate this Public License. d. Sections 1, 5, 6, 7, and 8 survive termination of this Public License. Section 7 -- Other Terms and Conditions. a. The Licensor shall not be bound by any additional or different terms or conditions communicated by You unless expressly agreed. b. Any arrangements, understandings, or agreements regarding the Licensed Material not stated herein are separate from and independent of the terms and conditions of this Public License. Section 8 -- Interpretation. a. For the avoidance of doubt, this Public License does not, and shall not be interpreted to, reduce, limit, restrict, or impose conditions on any use of the Licensed Material that could lawfully be made without permission under this Public License. b. To the extent possible, if any provision of this Public License is deemed unenforceable, it shall be automatically reformed to the minimum extent necessary to make it enforceable. If the provision cannot be reformed, it shall be severed from this Public License without affecting the enforceability of the remaining terms and conditions. c. No term or condition of this Public License will be waived and no failure to comply consented to unless expressly agreed to by the Licensor. d. Nothing in this Public License constitutes or may be interpreted as a limitation upon, or waiver of, any privileges and immunities that apply to the Licensor or You, including from the legal processes of any jurisdiction or authority. ======================================================================= Creative Commons is not a party to its public licenses. Notwithstanding, Creative Commons may elect to apply one of its public licenses to material it publishes and in those instances will be considered the “Licensor.” The text of the Creative Commons public licenses is dedicated to the public domain under the CC0 Public Domain Dedication. Except for the limited purpose of indicating that material is shared under a Creative Commons public license or as otherwise permitted by the Creative Commons policies published at creativecommons.org/policies, Creative Commons does not authorize the use of the trademark "Creative Commons" or any other trademark or logo of Creative Commons without its prior written consent including, without limitation, in connection with any unauthorized modifications to any of its public licenses or any other arrangements, understandings, or agreements concerning use of licensed material. For the avoidance of doubt, this paragraph does not form part of the public licenses. Creative Commons may be contacted at creativecommons.org. |
keywords |
admie
energy-markets
greece
henex
ipto
power-exchange
system-operation
|
VCS |
|
bugtrack_url |
|
requirements |
No requirements were recorded.
|
Travis-CI |
No Travis.
|
coveralls test coverage |
No coveralls.
|
# ExSO
An analytical framework for the Greek Power&Gas System Operation ("SO") and Market Exchange ("Ex") Data.
-----
## <span style="color: #cdd613"> What is it? </span>
**exso** provides an integrated framework for retrieving, extracting, transforming, loading and analyzing timeseries data for the Greek Power&Gas sector.
- The core of the project is to provide an automated, versatile and robust framework for:
- Downloading raw files ("the Datalake"), as reported by the Publishing Entities (ADMIE/IPTO, HEnEX, Desfa, ...)
- Parsing/converting raw files to flat, clean, high-quality timeseries
- Inserting/updating the parsed data to a local, self-maintained database ("the Database")
- Providing an API for accecssing, slicing, transforming, analyzing, and visualizing the local Database.
- The local database consists of a tree structure of local directories and .csv files. The resons we opted for csv-based format are aligned with the [Rationale](#rationale) of the project:
- Anyone can access a csv file without needing programming or SQL skills
- No local/remote database server required
- No significant loss of speed
-----
## Main Features
- Get **info** about implemented reports, their content, their availability periods, metadatata, etc.
- **Create** a local database of Market and System data (flat, seamless timeseries over the whole availability interval of each report)
- **Update** (hot/cold-start) the datalake and database for all or some of the implemented reports
- Interactive **Visualization**
- **Time-slicing** operations (timezone change, from/to time slicing)
- **Exporting/Extracting** visualizations and/or time-sliced data to a "sandbox" location (data in the database should not be modified in any way)
### <span style="color: #cdd613 "> Implemented Reports </span>
([see more here](#implemented-reports))
| id | Report Name | id | Report Name | id | Report Name | id | Report Name |
|----:|:-------------------------------------|----:|:-------------------------------------|----:|:-------------------------------------|---:|:-------------------------------------|
| 1 | AdhocISPResults | 21 | IDM_CRIDA2_AggDemandSupplyCurves | 41 | ISP1DayAheadRESForecast | 61| SystemRealizationSCADA |
| 2 | BalancingCapacityProduct | 22 | IDM_CRIDA2_MarketCoupling | 42 | ISP1ISPResults |
| 3 | BalancingEnergyProduct | 23 | IDM_CRIDA2_Results | 43 | ISP1Requirements |
| 4 | DAM_AggDemandSupplyCurves | 24 | IDM_CRIDA2_ResultsSummary | 44 | ISP1UnitAvailabilities |
| 5 | DAM_BlockOrders | 25 | IDM_CRIDA3_AggDemandSupplyCurves | 45 | ISP2DayAheadLoadForecast |
| 6 | DAM_GasVTP | 26 | IDM_CRIDA3_MarketCoupling | 46 | ISP2DayAheadRESForecast |
| 7 | DAM_MarketCoupling | 27 | IDM_CRIDA3_Results | 47 | ISP2ISPResults |
| 8 | DAM_PhysicalDeliveriesOfftakes | 28 | IDM_CRIDA3_ResultsSummary | 48 | ISP2Requirements |
| 9 | DAM_PreMarketSummary | 29 | IDM_LIDA1_AggDemandSupplyCurves | 49 | ISP2UnitAvailabilities |
| 10 | DAM_Results | 30 | IDM_LIDA1_Results | 50 | ISP3IntraDayLoadForecast |
| 11 | DAM_ResultsSummary | 31 | IDM_LIDA1_ResultsSummary | 51 | ISP3IntraDayRESForecast |
| 12 | DayAheadLoadForecast | 32 | IDM_LIDA2_AggDemandSupplyCurves | 52 | ISP3ISPResults |
| 13 | DayAheadRESForecast | 33 | IDM_LIDA2_Results | 53 | ISP3Requirements |
| 14 | DAS | 34 | IDM_LIDA2_ResultsSummary | 54 | ISP3UnitAvailabilities |
| 15 | DayAheadSchedulingUnitAvailabilities | 35 | IDM_LIDA3_AggDemandSupplyCurves | 55 | LTPTRsNominationsSummary |
| 16 | HVCUSTCONS | 36 | IDM_LIDA3_Results | 56 | RealTimeSCADARES |
| 17 | IDM_CRIDA1_AggDemandSupplyCurves | 37 | IDM_LIDA3_ResultsSummary | 57 | RealTimeSCADASystemLoad |
| 18 | IDM_CRIDA1_MarketCoupling | 38 | IDM_XBID_Results | 58 | ReservoirFillingRate |
| 19 | IDM_CRIDA1_Results | 39 | IMBABE | 59 | RESMV |
| 20 | IDM_CRIDA1_ResultsSummary | 40 | ISP1DayAheadLoadForecast | 60 | RESMVLVPROD |
-----
## Rationale
**Publicly-available does not always mean publicly-accessible**
- Market players, TSOs, and professionals in the energy sector may or may not already have access to some of the data made accessible by **exso**, through paid or "mebers only" subscriptions (e.g. market participants).
- Individuals, researchers, and in (surprisingly) many cases professionals are either not entitled, or not willing to pay for high-quality data access.
- Even when an interested party is willing to pay for high-quality, long-term timeseries data, it's not clear where would he/she attend to.
- To our knowledge, no commercial or "members-only" database provides any of the variety, the duration, the reliability and the transparency that **exso** provides.
- We strongly believe in open access and transparency. **ExSO** is a project aiming to render publicly-available data in the scope of the Greek Power&Gas sector, utilizable and accessible by anyone, expert or not.
-----
## Installation
Required Python Version >=3.10
```sh
pip install exso
```
- Note: If you are connected through your company's access point, and your company has restrictions against [PyPI](https://pypi.org/) and/or [GitHub](https://github.com/), you may be unable to install, not just ***exso*** but any open-source python package.
### For non-programmers
- Regardless of whether you have another Python installation in your PC or not, go ahead and install the [latest Python version](https://www.python.org/ftp/python/3.11.3/python-3.11.3-amd64.exe). Opt-in for the "Add Python to PATH" option during installation pop-up.
- ***exso*** is tested for Python >=3.10. Any previous version (<=3.9) will most likely be partially or fully incompatible.
- After installation is complete, **open a windows terminal** (In windows search, type "cmd" and hit enter)
- **Create a virtual environment** ([more on virtual environments (venvs) and how they work](https://docs.python.org/3/library/venv.html#how-venvs-work))
- First, locate which python versions are installed by typing (in the command-line terminal):
```sh
py -0
```
- Take note of the latest version (let's assume it is 3.11), and type:
```sh
py -3.11 -m venv "C:\Users\yourUserNameHere\exso_venv"
# (Hit Enter) Now the vitrtual environment is created.
```
- #### Activate the Virtual Environment
```sh
# (replace yourUserNameHere with your actual windows username, and hit enter)
"C:\Users\yourUserNameHere\exso_venv\Scripts\activate.bat"
```
- Install the ***exso*** package by typing (in the **same command line session**)
```sh
pip install exso
```
-----
# ***exso*** API
***exso*** can be used either through the **command line interface** ("CLI-based" for short), intended for only the core usage, **or** as an importable **python package** through any IDE ("IDE-based" for short), intended and allowing more advanced usage.
- At the moment, the CLI-based API has some namespace inconsistencies compared to the IDE-based API
- e.g. set_system_formats instead of _modify_system_formats(...), a.o.
- They can only become annoying if frequently switching from CLI-based to IDE-based usage, which is not really probable, but they will nonetheless be conformed by the next ***exso*** version.
### For non-programmers
* Both the CLI-based and the IDE-based usages are fairly straight-forward
* For Command-Line interface, see the [CLI documentation](#command-line-interface-cli)
* For IDE-based usage:
* Either install a proper IDE (e.g. [Pycharm Community](https://www.jetbrains.com/pycharm/download/download-thanks.html?platform=windows&code=PCC) is an excellent IDE but there's some learning curve involved), OR
* Use the python-interface:
* Open a command-line terminal and [Activate the exso virtual environment](#activate-the-virtual-environment)
* Then, simply write "python" and hit enter
![img_1.png](resources/img_1.png)
* Now, follow the [IDE-based usage documentation](#ide-based-usage)
* It is **critical** that you read the [System Formats](#system-formats) section before you get going.
* If you plan to use the exso API also for querying/time-slicing/visualizing data, you should read the documentation on [Database](#database), [Nodes](#nodes), and the [examples snippets](#examples-use-cases-special-attention)
-----
### Command Line Interface (CLI)
Open a windows terminal, and [activate the exso virtual environment](#activate-the-virtual-environment)
- List available reports and text descriptions (available in the sense of ***exso***-available, not necessarily already present in datalake/database)
```sh
py -m exso info
```
- Set system formats:
```sh
# Recommended: The below stand-alone command, permanently informs exso on your system format settings, and does not require to
# be explicitly specified again
py -m exso set_system_formats --decimal_sep "your decimal separator" --list_sep "your list separator"
# Otherwise, you can still specify the --decimal_sep and/or --list_sep arguments, but if the mode argument is not "set_system_formats",
# the modification will be valid just for this run of exso, and will revert to default values afterwards.
```
- Database & Datalake update (or first-time setup)
- **IMPORTANT:** Whenever an argument refers to a path (e.g. path/to/whatever), this should better be placed **inside double quotes** to ensure smooth operation (avoid issues with empty spaces in paths).
```sh
py -m exso update -rl path/to/root/datalake -rb path/to/root/database # or...
py -m exso update -rl path/to/root/datalake -rb path/to/root/database --which ReportName1 ReportName2 ReportName3
```
- You can find the full path to your specified directories by clicking on the address bar of windows explorer
![img.png](resources/img.png#center)
- Extraction / Timezone Conversion / Timeslicing / Visualizing (See section [Node Locators](#node-locators))
```sh
py -m exso query -rb path/to/root/database -loc NodeLocator -output_dir path/to/some/dir -tz desiredTimezone -from YYYY-MM-DD -until YYYY-MM-DD -extract -plot -stacked
```
- **Hint**: To avoid too much copy-pasting, the default values for root-lake and root-base are ".\datalake" and ".\database"
- This means, that (after having activated the virtual environment), if you navigate to a desired directory, you can launch the command line without specifying the -rl and -rb arguments
```sh
# Activate venv (indicative name of venv = "deploy_exso")
C:\Users\theUser>Desktop\VENVs\PythonVENVs\deploy_exso\Scripts\activate.bat
# Navigate to desired directory, that will host / is already hosting both the datalake and the database
(deploy_exso) C:\Users\theUser>cd Desktop\exso_data
# you can now launch exso without specifying root-lake and/or root-base paths
(deploy_exso) C:\Users\theUser\Desktop\exso_data>py -m exso update
```
-
#### Notes on CLI arguments
- **<span style="color: #2fb2b6"> -rl</span>**: path to the (desired or existing) root datalake directory
- **<span style="color: #2fb2b6"> -rb</span>**: path to the (desired or existing) root databake directory
- **<span style="color: #2fb2b6"> --which</span>**: if the positional argument is **update**, you can optionally enter only specific report names to update (space separated, case-sensitive, as they are listed in the [Implemented Reports](#implemented-reports) section)
- the <span style="color: #2fb2b6"> **-extract** </span> argument, if passed, must be accompanied by the -output_dir argument
- the <span style="color: #2fb2b6"> **-plot** </span> argument is valid only if the node corresponding to the given NodeLocator is a file or property node (not a directory)
- the <span style="color: #2fb2b6"> **-stacked** </span> argument makes the plot (if given) a stacked-area plot
- the <span style="co[exso_CLI.doc](src%2Fexso_CLI.doc)lor: #2fb2b6"> **-tz** </span> argument is optional. If given, it will convert the database data (which is in UTC) to the specified timezone (including daylight-saving shifts)
- Recommendation: Don't enter country-specific timezone names. Prefer broader timezones (e.g. EET, CET, UTC, etc.).
- the <span style="color: #2fb2b6"> **-from** </span> argument is optional (can be combined with -until): format "YYYY-MM-DD HH:MM"
- the <span style="color: #2fb2b6"> **-until** </span> argument is optional (can be combined with -from): format "YYYY-MM-DD HH:MM"
#### CLI formal documentation
```sh
py -m exso --help
```
```sh
usage: py -m exso [-h] [-rl ROOT_LAKE] [-rb ROOT_BASE] [--which WHICH [WHICH ...]] [--val_report VAL_REPORT] [--val_dates VAL_DATES [VAL_DATES ...]] [--val_fields VAL_FIELDS [VAL_FIELDS ...]] [-loc QUERY_LOCATOR]
[-output_dir QUERY_OUTPUT_DIR] [-tz QUERY_TZ] [-from QUERY_FROM] [-until QUERY_UNTIL] [-extract] [-plot] [-stacked] [--decimal_sep DECIMAL_SEP] [--list_sep LIST_SEP]
{info,update,validate,query,set_system_formats}
positional arguments:
{info,update,validate,query,set_system_formats}
options:
-h, --help show this help message and exit
-rl ROOT_LAKE, --root_lake ROOT_LAKE
-rb ROOT_BASE, --root_base ROOT_BASE
--which WHICH [WHICH ...]
--which argument can be either 'all' (default), or a list of valid report-names (space-separated)
--val_report VAL_REPORT
report name you wish to validate.
--val_dates VAL_DATES [VAL_DATES ...]
space separated date(s) to validate. format: YYYY-M-D
--val_fields VAL_FIELDS [VAL_FIELDS ...]
"Field(s)" are the filenames, as to be found in the database folder of a specific report (space separated).
-loc QUERY_LOCATOR, --query_locator QUERY_LOCATOR
'locator' means a unique identifier of database objects. example: root.admie.isp1ispresults, will extract the whole database of this report and transform it / slice it depending on the rest of the options you
set.
-output_dir QUERY_OUTPUT_DIR, --query_output_dir QUERY_OUTPUT_DIR
If specified, it will be used to save the generated plot (if -plot), and/or the extracted timeslice (if -extract).
-tz QUERY_TZ, --query_tz QUERY_TZ
-from QUERY_FROM, --query_from QUERY_FROM
Start date(time) of query (YYYY-M-D [H:M])
-until QUERY_UNTIL, --query_until QUERY_UNTIL
End date(time) of query (YYYY-M-D [H:M])
-extract, --query_extract
If added, it means you wish to EXTRACT the specified query (among possible other actions)
-plot, --query_plot If added, it means you wish to PLOT the upstream query (among possible other actions)
-stacked, --plot_stacked
If added, it means you wish the PLOT specified, to be a stacked-area plot
--decimal_sep DECIMAL_SEP
--list_sep LIST_SEP
```
-----
# IDE-based Usage
## Basic update
The below script will download and insert to the database all (61) currently supported reports. For more information continue reading.
```sh
from exso import Updater
# define where you want the datalake and the database to be stored in the disk
root_lake = r"path\to\desired\datalake\directory" # e.g. r"C:\Users\your_username\exsodata\datalake"
root_base = r"path\to\desired\database\directory" # e.g. r"C:\Users\your_username\exsodata\database"
# root_lake and root_base can also be pathlib.Path objects
upd = Updater(root_lake, root_base, all=True)
upd.run()
```
#### Update Progress
Progress bars will be displayed for every kind of operation for each report, as demonstrated in the figure below, for e.g. report = DAM_Results:
![progress_bar_example.png](resources/progress_bar_example.png)
#### Performance & System Requirements
A **full cold-start** process of all 61 reports, might take from **2 to over 5 hours**, depending on internet speed, processing power, memory and non-***exso*** PC usage. Indicative time requirements:
- Pentium-tier processors + 4GB RAM --> ~6 hours (not a good idea in general)
- 8th-gen high-performance i5 processor + 16GB --> 2.5 hours
- 12th-gen high-performance i7 processor + 16GB --> 1.5 hours
After the first database update (creation actually), each **daily or weekly update** process takes a matter of **a few minutes**.
***exso*** is not optimized for performance (and probably won't be anytime soon):
- Publishing parties keep on changing formats, contents, adding/removing report sections, modifying string representations, etc.
- This requires very "low-level" (in variable level, not machine-level) control over reading, parsing, datetime-indexing in order to work for all reports
- Frankly performance is not essential:
- The blow is taken once, at the database initialization: Daily, weekly, monthly updates are really a matter of minutes.
- The database querying speeds are fast (speed in data operations is always refering to a specific context). The computational cost lies in parsing the datalake files.
The combined Database & Datalake takes up approximately **4GB of disk space**
-----
## Implemented Reports
***exso*** currently supports a total of 61 reports from ADMIE/IPTO and HEnEx. Some of them are obsolete (but still useful for analytics), and some ongoing.
In order to review the available reports or decide which ones fit your needs and update only those, a Report.Pool object can be useful.
```sh
from exso import Report
# Instantiate a report Pool object
rp = Report.Pool()
description = rp.get_text_description() # returns a dictionary of available reports. e.g. {report_name1: report1_description, report_name2: report2_description}
# to get the description of a specific report, you can print:
report_name = "select a valid report name"
print(description[report_name])
# The .get_available() method, returns dataframe with available reports, and their basic metadata
# To get only a list of names, set only_names <- False
metadata = rp.get_available(only_names = False)
```
![pool_get_available.png](resources/get_available().png)
-----
## Custom Update
```sh
# Now, if you conclude that you want e.g. 3 specific reports (at least for now), you can:
interesting_reports = ['reportname1', 'reportname2', 'reportname3']
upd = Updater(root_lake, root_base, reports_pool = rp, some = interesting_reports) # given that you have already instantiated a Report.Pool object
# or, for future use
upd = Updater(root_lake, root_base, some = interesting_reports) # with no need of instantiating a Report.Pool object
# Very often, a specific datalake-file (excel file) for a given date, may have multiple versions. (e.g. YYYYMMDD_report_01.xlsx, YYYYMMDD_report_02.xlsx)
# The use_lake_version argument allows to select which version to use.
# By default it is set to 'latest'. Other options: 'first' or any natural number
upd.run(use_lake_version = 'latest')
# The below query, will use datalake-files whose version is 4, given that the file actually has 4 or more versions.
# If the specific file has only two versions, then version 2 will be used.
# The integer lake version is interpreted as: "Use this version, or the most recent version available prior to it"
upd.run(use_lake_version = 4)
```
-----
# Datalake
The datalake consists of raw excel (.xls, or .xlsx, or .zip of .xls*) reports, as published by the publishing parties.
- Each report is published (is available) over a specific date range (some reports may be no longer actively updated but still useful for historical analysis)
- Each report is published on a specific frequency (e.g. each day, each week, each month, etc.)
- Each report file content, spans over various horizons (e.g. one day-long, one week-long, one month-long, etc.)
- Each report file consists of one or more excel sheets
- Each report is expressed in a specific timezone (EET, UTC or CET) and may or may not have well-defined daylight-saving switches.
-----
# Datalake → Database
***exso*** performs:
- For each report (report-type)
- Datalake Update (download raw excel files if there are newly published data)
- For each report file (i.e. for each date that this report was/is published)
- For each report file excel-sheet
- Data parsing, data cleaning, datetime conforming, disambiguation, joining
- Database "upsert" (update / insert)
-----
# Database
The database, automatically created and maintained through the ***exso*** package, has a tree-like structure.
→ **After the creation of the local database** ([update mode](#basic-update)), you can access it through the ***exso*** API.
- When the tree is initialized, it only retrieves the structure of the directories and files, and the column-names of each .csv file
- It does not read and load to memory the whole database. This can be done by the __call__ method of a Node object ([see Data Access section](#data-access))
### Database Tree Visualization
```sh
from exso import Tree
tree = Tree(root_base)
tree.make_tree()
# You can make a quick visualization of the directory structure and contents:
tree.visualize()
```
![database_viz.png](resources/database_Viz.png)
-----
### Datalake vs Database Comparison
#### Reports (aka report-names, or report-types)
- ***Report Names are almost always the same as the official report-names*** published by the publishing entities, and that's how they appear both in the datalake and in the database
- When using the "report"-kind to access some data, and having doubts on the correct names/strings to use, the name-to-use can be found:
- in the [Implemented Reports](#implemented-reports)
- By actually moving through the database directories (reports are directories)
- Or by using the [Database Tree Visualization](#database-tree-visualization)
- Or by accessing a "publisher"-kind Node-object's .children attributes
```sh
publisher = Tree['root.some_publisher']
reports = publisher.children
```
#### Excel Sheets / Fields
- Raw excel sheets may have peculiar naming conventions in the raw files.
- Excel ***sheet names are referred to as "fields" in the database-scope***, and may differ from the actual excel-sheet names of the raw datalake
- When using the "field"-kind to access some data (after the database creation/update), and having doubts on the correct names/strings, to use the name-to-use can be found in the actual database folder structure (Fields are directories in the database, NOT files).
- You can do this by actually moving through the database directories
- Or by using the [Database Tree Visualization](#database-tree-visualization)
- Or by accessing a "report"-kind Node-object's .children attributes
```sh
report = Tree['root.some_publisher.some_report']
fields = report.children
```
#### Database files
- Because each raw excel **sheet** sometimes contains a lot of data that are sometimes heterogenonus:
- Each raw excel sheet (corresponding to a database-field) is broken down in one or more "subfields", which are the actual final .csv files
- When using the "file"-kind to access some data (after the database creation/update), and having doubts on the correct names/strings, the name-to-use can be found in the actual database folder structure ("Files" are .csv files in the database).
- You can do this by actually moving through the database directories
- Or by using the [Database Tree Visualization](#database-tree-visualization)
- Or by accessing a "field"-kind Node-object's .children attributes
```sh
field = Tree['root.some_publisher.some_report.some_field']
files = field.children
```
#### Database Columns / Properties
- Each database file has one or more columns or "properties" (apart from the datetime column)
- When using the "property"-kind to access some data (after the database creation/update), and having doubts on the correct names/strings, the name-to-use can be found in the actual database folder structure ("Properties" are columns in the .csv files in the database).
- You can do this by manually opening the database .csv files
- **Not available** through the [Database Tree Visualization](#database-tree-visualization)
- Or by accessing a "file"-kind Node-object's .children attributes
```sh
file = Tree['root.some_publisher.some_report.some_field.some_file']
properties = file.children
```
-----
## Nodes
Each branch (a sub-tree) or leaf (an end-Node) of the Database Tree is a Node object. Each Node has specific [attributes](#node-attributes) (both in the literal, and in programming notion):
These attributes assist in accessing, manipulating, visualizing or exporting any combination of required system or market properties.
-----
### Node Attributes
The text below serves both a descriptive and a definitive purpose. (So, from now on, the term "kind" will have the meaning defined in this section)
- .name
- a (descriptive) string. Sometimes, names are automatically given from the raw files, while other times there are some alterations. Names are generally non-unique accross the tree, but unique within the children of one node.
- .path
- physical path in the disk (directory or file)
- .dna
- a concatenation of all the node's parents, dot-separated, and **case insensitive** (e.g. "root.henex.DaM_ReSuLtS")
- .kind
- In the ***exso*** database, nodes can be of one of the 6 following kinds:
- "root" (parent of all nodes)
- "publisher" (parent of all reports, published by that publisher)
- "report" (the name of the report, e.g. "ISP1ISPResults")
- "field" (the name of the sheet of the original report-excel, e.g. ISP_Schedule. *with some exceptions)
- "file" (a csv file containing some or all of the sheet (field)-data e.g. "Load")
- "property" (a column of the csv file, e.g. "Net Load")
- .parent
- .children
- .siblings
- .ascendants
- .descendants
- These are pretty much self-explanatory. They refer to Node objects, or Groups of Node objects (e.g. node.children returns a Group object, but can be accessed as a Node, e.g. node.children.dna, will return a list of dnas of that node's children)
-----
### Node Locators
Node Locators are unique Node identifications. Nodes can be uniquely accessed in more than one ways. The three main node locator types are:
- DNA locators
- Path locators
- Successive children locators
In all three cases, nodes are accessed through a succession chain:
- **root > publisher > reportName > fieldName > fileName** [>columnName]
- root: Literally the database root, which is annotated simply as "root"
- publisher: currently admie/henex
- For reportName, fieldName, fileName, columnName, consult with sections
-
- reportName:
For better demonstration, we'll use the example of ISP Activations/Redispatch, of a non-schedulued ISP (report_name = "AdhocISPResults"), only for Hydroelectric Units (see database visualization above)
The file is called **"Hydro.csv"** and is located in the directory **"root/admie/AdhocISPResults/ISP_Activations"**. All three methods below will return the desired Node object.
- #### DNA Locator
```sh
tree['root.admie.adhocispresults.isp_activations.hydro'] # lower/upper case unimportant
```
- #### Path Locator
```sh
tree["C:\path_to_root_database\admie\AdhocISPResults\ISP_Activations\Hydro.csv"] # exact path must be provided
```
- #### Successive children locators
```sh
tree['root']['admie']['AdhocISPResults']['ISP_Activations']['Hydro'] # case sensitive: it accesses the names of the children of each successive node access
```
-----
### Data Access
Data can be accessed, visualized, manipulated and extracted through Node operations. **(retrieve, export, plot)**
- Once a node has been called in any way (for export, for plot or just retrieval), the whole node is read and stored in memory.
- **Any timezone, or time-slicing operations, only affect the returned data.**
- The node's data remain intact and always in UTC timezone
Optional arguments (**kwargs) are common for all three operations, and they can modify the returned time-range (from/to), and the returned timezone.
- tz_pipe: None or list
- If a list is provided: [database_timezone, target_timezone, (None)]
- The databases timezone is UTC and shall be set to UTC
- The target_timezone can be any pytz compatible timezone: Try to stick to wider zones (EET, CET, UTC, GMT) and not country-wide timezones (e.g. Europe/Athens, etc.)
- Last argument:
- If is None, this means, after tz-conversion, truncate the timezone information (e.g. 2022-1-1 00:00 +02:00 will become --> 2022-1-1 00:00)
- If it doesn't exist (e.g. the list is [database_tz, target_Tz], the returned data will contain the tz-information)
- start_date / end_date: it can accept pandas Timestamps and datetime.datetime objects, but also strings formatted as YYYY-MM-DD HH:MM
#### Usage
Retrieve a node's data:
```sh
node_data = node() # data is actually read from the database files
# retrieve data only for January '22, converted to EET timezone:
# data is retrieved from memory
node_data_range = node(tz_pipe = ['utc', 'eet', None], start_date = '2022-1-1', end_date = '2022-1-31')
# retrieve all the node's data, in utc (data is retrieved from memory)
node_data = node() # node's internal data not affected by an intermediate timezone/timeslicing operation.
```
Export a node's data:
```sh
node.export(to_path, **kwargs)
```
Plot a node's data (node must be of file- or property-kind):
- Plotting whole files of considerable size (i.e. > 5-10 MB) may require considerable time
```sh
node.plot() # this will plot the file timeseries in UTC over its whole span
# example: stacked-area plot, in EET timezone, only for the period after 1st-Jan 2023. Also save the plot somewhere specified.
node.plot(tz_pipe = ['utc', 'eet', None], start_date = '2023-1-1', show = True, save_path = "C:/Users/Desktop/my_plot.html", area = True)
```
-----
#### Remarks
- The structure of the ***exso*** database, was built as a "one-size-fits-all" solution. This may work intuitively in information-rich reports, and less intuitively in very simple reports:
```sh
# access sequence with DNA locators
tree['root.<publisher>.<report_name>.<sheet/field>.<filename/subfield>]
```
- e.g. report = ISP1ISPResults
- raw datalake file has **8 sheets**, and each sheet has **multiple subfields** (e.g. thermal dispatch, hydro dispatch, load, reserve requirements, a.o.)
- in order to access e.g. thermal dispatch: tree['root.admie.isp1ispresults.isp_schedule.thermal]
- --> intuitive
- e.g. report = ISP1DayAheadLoadForecast
- raw datalake file has a **single sheet**, with a **single subfield** (30-min Load Forecast)
- in order to access it: tree['root.admie.ISP1DayAheadLoadForecast.LoadForecast.LoadForecast']
- --> seems repetitive, but that's how it is.
-----
# Examples, Use-cases, Special Attention
## Database Nodes and NodesLocators
- Accessing nodes, transforming, and custom-exporting.
```sh
# Access a node:
isp1 = tree['root.admie.isp1ispresults'] # dict-like acess, dot separated, case INsensitive. --> the recommended usage
#Alternative methods
isp1 = tree.get_node(locator = 'root.admie.isp1ispresults') # get_node method accepts more than DNA locators
isp1 = tree.get_node(locator = 'path/to/database/admie/isp1ispresults')
isp1 = tree['root']['admie']['ISP1ISPResults']
# you can access a nodes' children
print(isp1.children)
# and get their dnas, names, paths
print(isp1.children.dna)
print(isp1.children.name)
isp1loadforecast_node = tree['root.admie.isp1ispresults.isp_schedule.load']
# Calling a node, returns the nodes' contents (dataframe, or dict (of dicts) of dataframes
isp1loadforecast_df = isp1loadforecast_node() # returns all available data in UTC timezone
## typical use-case: query some specific date range, convert it to a desired timezone and store it somewhere else, to send it to a colleague or do some excel graphs on it
isp1 = tree['root.admie.isp1ispresults'] # dot separated, case INsensitive
# export the whole data in utc timezone
isp1.export(to_path = "where/to/export/full_data_utc")
# export custom range in custom timezone
isp1.export(to_path = "where/to/export/sliced_data_eet", tz_pipe = ['utc', 'eet', None], start_date = '2022-1-1 00:00', end_date = '2022-12-31 23:30')
# Note: the start_date:end_date filter is applied AFTER the timezone conversion (if given)
# tz_pipe: a list of timezone operations
#The first argument must ALWAYS be UTC (that's the timezone that the csv files are initially on)
##e.g. 01/01/2021 00:00 will become 01/01/2021 00:00+00:00
#The second argument (optional) is a timezone to convert the data to. (e.g. EET)
##e.g. 01/01/2021 00:00+00:00 will become: 2021-01-01 02:00+02:00
#The third argument (optional) can only be None, if provided. It means, truncate the timezone information from the converted data
##e.g. 2021-01-01 02:00+02:00 will become: 2021-01-01 02:00
## IMPORTANT:
# The optional arguments tz_pipe, start_date, end_date do not persist in memory!!
# This means that, the returned dataframe or dict of dataframe will have the requested charasteristics, but the node keeps its original information
# That is, if to be re-called without arguments, it will immediately return its raw content: UTC, tz-unaware, full available range
```
## Visualization
***exso*** utilizes the (extremely helpful and interactive) package [plotly](https://plotly.com/python) for data visualization.
The visualization of a Node object is as simple as calling its .plot() method:
Graphs can be zoomed in/out, rescaled, columns can be toggled-on/off in real time.
→ By default, ***exso*** will **omit to plot any columns that are Zero or NaN** over the whole selected timerange, in order to nake the plot lighter.
```sh
isp1_thermal_gen = t['root.admie.isp1ispresults.isp_schedule.thermal']
fig = isp1_thermal_gen.plot(area = True, start_date = '2022-1-1', end_date = '2022-1-10', tz_pipe = ['UTC', 'EET', None], show = True, save_path = None)
# the returned figure is of type "plotly.graph_objs._figure.Figure", meaning, you can set "show"=False, and update the layout with normal plotly usage before displaying it.
# Some very basic modification-options (title, x&y labels) will be supported directly through the exso.Node object
```
![plotly_viz.png](resources/plotly_viz.png)
#### When calling the .plot() method of a Node:
- Any columns containing only NaN or only zero & nan values are **dropped**, in order to get a cleaner graph.
- The area = True/False argument, modifies whether the plot will be a stacked area, or a line plot
- start/end_dates, tz_pipe work exactly as in the __call__ method (node())
- show = True/False argument, controls whether to automatically display the graph when its rendered (in both cases, a figure object is returned)
- The save_path argument accepts a Path-like entry (.html), to locally save the graph (regardless of whether show=True/False)
## Disclaimers
- Any modification on the datalake and database structure, file and directory names, additions/deletions, will probably cause malfunctions.
- If you must open a database csv in-place (directly from the database directory, e.g. for quick inspection), which you shouldn't, you should at least not perform any saves (even if apparently, no changes were made)
- If you accidentally modify e.g. a database file, and the Updater malfunctions, you can manually delete the whole reports folder, and re-run the Updater.
## System Formats
By default, ***exso*** uses:
- "," (comma) for list separation
- "." (dot) for decimal point
If your system settings are different, and don't want to change them, you can:
- Modify ***exso***'s default values just **once**
- Modify ***exso***'s default values **persistently** → recommended option
If your default/desired system formats do not comply with ***exso***'s defaults, these lines should be placed **before executing** any ***exso*** operations in order to work properly:
- IDE-based usage
```sh
import exso
# modify persistently
exso._modify_system_formats(decimal_sep = "your decimal separator", list_sep = "your list separator")
# modify just once
exso._list_sep = "your list separator"
exso._decimal_sep = "you decimal point separator"
```
- CLI-based usage
```sh
# modify persistently
py -m exso set_system_formats --decimal_sep "your decimal separator" --list_sep "your list separator"
# modify just once
py -m exso <whatever command and config> --decimal_sep "your decimal separator" --list_sep "your list separator"
```
----
## Data Validation
The frequent changes in formats, reporting properties, periodic or one-time mislabeling, timezone ambiguity etc., may at some point result in a report being partially mis-parsed.
The Validation module of **exso** makes it easier to inspect and compare the raw datalake files to the database content.
The validation process creates and launches an excel-file, which conatains:
- The raw lake data, as they originally were
- The timeseries-data, on the same timzone that the raw data lake is
- The timeseries-data on UTC timezone (as they would be stored in the database)
```sh
from exso import Validation
# simplest setup validation (will validate only the first sheet of the raw datalake file)
report_names = 'ISP1ISPResults'
inspect_dates = '2022-2-19'
val = Validation(report_names=report_names, dates = inspect_dates, root_lake=root_lake, fields=None)
val.run()
# multi-val setup
report_name = 'ISP1ISPResults' # str
fields = ['ISP_Activations', 'CCGT_Schedule'] # str|list
inspect_dates = ['2022-2-19', '2022-2-20'] # datetime-like|list
val = Validation(report_name=report_name, dates = inspect_dates, root_lake=root_lake, fields=fields)
val.run()
```
----
## Features under active Development
### Data Documentation
Another aspect that creates difficulties in utilizing the published data (after one overcomes the sparsity of data), is the lack of detailed documentation per report, field, or property.
(e.g. The term "Net Load" may mean System Load minus pumping load, or Consumption minus RES, or Consumption minus RES minus pumping load, etc.)
At this stage, the Data Documentation provided in the ***exso*** package is far from perfect: Data Documentation is currently only on the report-level, providing high-level insights but not detailed disambiguations.
- A custom-made documentation, built as a light non-relational database is currently being developed and will be launched with one of the next versions of ***exso***.
### Analytics API
The current setup is oriented around *reports*. An Analytics API currently under development, will facilitate:
- Seaming properties from different reports of different timeframws (e.g. System Marginal Price to Market Clearing Price)
- Dedicated reporting and visualization (e.g. Daily System Snapshot of market prices, imports, loads, reserves, balancing, generation mix, etc.)
- Advanced analytics methods (e.g. Unit Unavailabilities statistics, comparisons, correlations)
### Support for more Reports
The next version, will contain some improvements on existing reports, and the addition of Water declaration and NTC reports.
### Support for Linux
Support for Linux-based systems is not foreseen at the moment, but feel free to submit a request if needed.
----
## Tests
- ***ExSO*** is fairly tested for the envisaged usage, but since the project is not (at least yet) intended for collaborative development, tests are not published.
- The design philosophy is not to catch all errors imagineable, but rather that basic users will stick to basic/documented usage, and that advanced users know what they're doing
- e.g. attempting to plot a file beyond its available datetime range, will not
- From the user's perspective, the [Validation module](#data-validation) is available to assist in validating/trusting that the database accurately reflects the raw datalaek files.
----
## Issues
- Feel free to submit any issues [here](https://github.com/ThanosGkou/exso/issues) or via e-mail
----
## License
<a rel="license" href="http://creativecommons.org/licenses/by-nc-nd/4.0/"><img alt="Creative Commons License" style="border-width:0" src="https://i.creativecommons.org/l/by-nc-nd/4.0/88x31.png" /></a><br />This work is licensed under a <a rel="license" href="http://creativecommons.org/licenses/by-nc-nd/4.0/">Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (CC BY-NC-ND 4.0)</a>
Briefly (without this description being a substitute for the full license or any of its clauses):
**You are free to**:
- Use — Download/Install/Deploy ***exso***
- Share — copy and redistribute the material in any medium or format
**Under the following terms**:
- Attribution — You must give appropriate credit, provide a link to the license, and indicate if changes were made. You may do so in any reasonable manner, but not in any way that suggests the licensor endorses you or your use.
- NonCommercial — You may not use the material for commercial purposes.
- NoDerivatives — If you remix, transform, or build upon the material, you may not distribute the modified material.
- No additional restrictions — You may not apply legal terms or technological measures that legally restrict others from doing anything the license permits.
----
## Citation
If **exso** assists you in making the "publicly available" data, actually valuable and accessible, consider citing:
#### APA
- Natsikas, T. (2023). ExSO: market Exchange and Sytem Operation analytical framework (Version 0.0.0) [Computer software]. https://github.com/ThanosGkou/exso
#### BibTeX
- @software{Natsikas_ExSO_market_Exchange_2023,
author = {Natsikas, Thanos},
month = apr,
title = {{ExSO: market Exchange and Sytem Operation analytical framework}},
url = {https://github.com/ThanosGkou/exso},
version = {0.0.0},
year = {2023}
}
----
Raw data
{
"_id": null,
"home_page": "",
"name": "exso",
"maintainer": "",
"docs_url": null,
"requires_python": ">=3.10",
"maintainer_email": "Thanos Natsikas <ath.natsikas@example.com>",
"keywords": "admie,energy-markets,greece,henex,ipto,power-exchange,system-operation",
"author": "",
"author_email": "Thanos Natsikas <ath.natsikas@example.com>",
"download_url": "https://files.pythonhosted.org/packages/91/89/f8229409ac9c3fe473d75c0b727c23472190eb7b8a485d61a9de90852ec2/exso-0.0.0.tar.gz",
"platform": null,
"description": "# ExSO\nAn analytical framework for the Greek Power&Gas System Operation (\"SO\") and Market Exchange (\"Ex\") Data.\n\n\n-----\n## <span style=\"color: #cdd613\"> What is it? </span>\n**exso** provides an integrated framework for retrieving, extracting, transforming, loading and analyzing timeseries data for the Greek Power&Gas sector.\n\n- The core of the project is to provide an automated, versatile and robust framework for:\n - Downloading raw files (\"the Datalake\"), as reported by the Publishing Entities (ADMIE/IPTO, HEnEX, Desfa, ...)\n - Parsing/converting raw files to flat, clean, high-quality timeseries\n - Inserting/updating the parsed data to a local, self-maintained database (\"the Database\")\n - Providing an API for accecssing, slicing, transforming, analyzing, and visualizing the local Database.\n\n\n- The local database consists of a tree structure of local directories and .csv files. The resons we opted for csv-based format are aligned with the [Rationale](#rationale) of the project:\n - Anyone can access a csv file without needing programming or SQL skills\n - No local/remote database server required\n - No significant loss of speed\n\n-----\n## Main Features\n\n- Get **info** about implemented reports, their content, their availability periods, metadatata, etc.\n- **Create** a local database of Market and System data (flat, seamless timeseries over the whole availability interval of each report)\n- **Update** (hot/cold-start) the datalake and database for all or some of the implemented reports\n- Interactive **Visualization**\n- **Time-slicing** operations (timezone change, from/to time slicing)\n- **Exporting/Extracting** visualizations and/or time-sliced data to a \"sandbox\" location (data in the database should not be modified in any way)\n\n ### <span style=\"color: #cdd613 \"> Implemented Reports </span> \n\n ([see more here](#implemented-reports))\n\n\n\n| id | Report Name | id | Report Name | id | Report Name | id | Report Name |\n|----:|:-------------------------------------|----:|:-------------------------------------|----:|:-------------------------------------|---:|:-------------------------------------|\n| 1 | AdhocISPResults | 21 | IDM_CRIDA2_AggDemandSupplyCurves | 41 | ISP1DayAheadRESForecast | 61| SystemRealizationSCADA |\n| 2 | BalancingCapacityProduct | 22 | IDM_CRIDA2_MarketCoupling | 42 | ISP1ISPResults |\n| 3 | BalancingEnergyProduct | 23 | IDM_CRIDA2_Results | 43 | ISP1Requirements |\n| 4 | DAM_AggDemandSupplyCurves | 24 | IDM_CRIDA2_ResultsSummary | 44 | ISP1UnitAvailabilities |\n| 5 | DAM_BlockOrders | 25 | IDM_CRIDA3_AggDemandSupplyCurves | 45 | ISP2DayAheadLoadForecast |\n| 6 | DAM_GasVTP | 26 | IDM_CRIDA3_MarketCoupling | 46 | ISP2DayAheadRESForecast |\n| 7 | DAM_MarketCoupling | 27 | IDM_CRIDA3_Results | 47 | ISP2ISPResults |\n| 8 | DAM_PhysicalDeliveriesOfftakes | 28 | IDM_CRIDA3_ResultsSummary | 48 | ISP2Requirements |\n| 9 | DAM_PreMarketSummary | 29 | IDM_LIDA1_AggDemandSupplyCurves | 49 | ISP2UnitAvailabilities |\n| 10 | DAM_Results | 30 | IDM_LIDA1_Results | 50 | ISP3IntraDayLoadForecast |\n| 11 | DAM_ResultsSummary | 31 | IDM_LIDA1_ResultsSummary | 51 | ISP3IntraDayRESForecast |\n| 12 | DayAheadLoadForecast | 32 | IDM_LIDA2_AggDemandSupplyCurves | 52 | ISP3ISPResults |\n| 13 | DayAheadRESForecast | 33 | IDM_LIDA2_Results | 53 | ISP3Requirements |\n| 14 | DAS | 34 | IDM_LIDA2_ResultsSummary | 54 | ISP3UnitAvailabilities |\n| 15 | DayAheadSchedulingUnitAvailabilities | 35 | IDM_LIDA3_AggDemandSupplyCurves | 55 | LTPTRsNominationsSummary |\n| 16 | HVCUSTCONS | 36 | IDM_LIDA3_Results | 56 | RealTimeSCADARES |\n| 17 | IDM_CRIDA1_AggDemandSupplyCurves | 37 | IDM_LIDA3_ResultsSummary | 57 | RealTimeSCADASystemLoad |\n| 18 | IDM_CRIDA1_MarketCoupling | 38 | IDM_XBID_Results | 58 | ReservoirFillingRate |\n| 19 | IDM_CRIDA1_Results | 39 | IMBABE | 59 | RESMV |\n| 20 | IDM_CRIDA1_ResultsSummary | 40 | ISP1DayAheadLoadForecast | 60 | RESMVLVPROD |\n\n\n\n\n-----\n## Rationale\n**Publicly-available does not always mean publicly-accessible** \n- Market players, TSOs, and professionals in the energy sector may or may not already have access to some of the data made accessible by **exso**, through paid or \"mebers only\" subscriptions (e.g. market participants).\n- Individuals, researchers, and in (surprisingly) many cases professionals are either not entitled, or not willing to pay for high-quality data access.\n- Even when an interested party is willing to pay for high-quality, long-term timeseries data, it's not clear where would he/she attend to.\n- To our knowledge, no commercial or \"members-only\" database provides any of the variety, the duration, the reliability and the transparency that **exso** provides. \n- We strongly believe in open access and transparency. **ExSO** is a project aiming to render publicly-available data in the scope of the Greek Power&Gas sector, utilizable and accessible by anyone, expert or not.\n\n\n-----\n## Installation\n\nRequired Python Version >=3.10\n```sh\npip install exso\n```\n\n- Note: If you are connected through your company's access point, and your company has restrictions against [PyPI](https://pypi.org/) and/or [GitHub](https://github.com/), you may be unable to install, not just ***exso*** but any open-source python package.\n### For non-programmers\n- Regardless of whether you have another Python installation in your PC or not, go ahead and install the [latest Python version](https://www.python.org/ftp/python/3.11.3/python-3.11.3-amd64.exe). Opt-in for the \"Add Python to PATH\" option during installation pop-up.\n - ***exso*** is tested for Python >=3.10. Any previous version (<=3.9) will most likely be partially or fully incompatible.\n \n\n- After installation is complete, **open a windows terminal** (In windows search, type \"cmd\" and hit enter)\n\n\n- **Create a virtual environment** ([more on virtual environments (venvs) and how they work](https://docs.python.org/3/library/venv.html#how-venvs-work))\n - First, locate which python versions are installed by typing (in the command-line terminal):\n\n ```sh\n py -0\n ```\n \n - Take note of the latest version (let's assume it is 3.11), and type:\n ```sh\n py -3.11 -m venv \"C:\\Users\\yourUserNameHere\\exso_venv\"\n \n # (Hit Enter) Now the vitrtual environment is created.\n ```\n \n \n \n- #### Activate the Virtual Environment\n ```sh\n # (replace yourUserNameHere with your actual windows username, and hit enter)\n \"C:\\Users\\yourUserNameHere\\exso_venv\\Scripts\\activate.bat\"\n ```\n \n \n \n\n- Install the ***exso*** package by typing (in the **same command line session**)\n ```sh\n pip install exso\n ```\n\n-----\n# ***exso*** API\n ***exso*** can be used either through the **command line interface** (\"CLI-based\" for short), intended for only the core usage, **or** as an importable **python package** through any IDE (\"IDE-based\" for short), intended and allowing more advanced usage.\n- At the moment, the CLI-based API has some namespace inconsistencies compared to the IDE-based API\n - e.g. set_system_formats instead of _modify_system_formats(...), a.o.\n- They can only become annoying if frequently switching from CLI-based to IDE-based usage, which is not really probable, but they will nonetheless be conformed by the next ***exso*** version.\n\n### For non-programmers\n* Both the CLI-based and the IDE-based usages are fairly straight-forward\n* For Command-Line interface, see the [CLI documentation](#command-line-interface-cli)\n* For IDE-based usage:\n * Either install a proper IDE (e.g. [Pycharm Community](https://www.jetbrains.com/pycharm/download/download-thanks.html?platform=windows&code=PCC) is an excellent IDE but there's some learning curve involved), OR \n * Use the python-interface:\n * Open a command-line terminal and [Activate the exso virtual environment](#activate-the-virtual-environment)\n * Then, simply write \"python\" and hit enter\n ![img_1.png](resources/img_1.png)\n * Now, follow the [IDE-based usage documentation](#ide-based-usage)\n \n\n* It is **critical** that you read the [System Formats](#system-formats) section before you get going.\n* If you plan to use the exso API also for querying/time-slicing/visualizing data, you should read the documentation on [Database](#database), [Nodes](#nodes), and the [examples snippets](#examples-use-cases-special-attention)\n\n\n\n-----\n### Command Line Interface (CLI)\n\nOpen a windows terminal, and [activate the exso virtual environment](#activate-the-virtual-environment)\n- List available reports and text descriptions (available in the sense of ***exso***-available, not necessarily already present in datalake/database)\n\n ```sh\n py -m exso info\n ```\n \n- Set system formats:\n ```sh\n # Recommended: The below stand-alone command, permanently informs exso on your system format settings, and does not require to \n # be explicitly specified again\n \n py -m exso set_system_formats --decimal_sep \"your decimal separator\" --list_sep \"your list separator\"\n # Otherwise, you can still specify the --decimal_sep and/or --list_sep arguments, but if the mode argument is not \"set_system_formats\",\n # the modification will be valid just for this run of exso, and will revert to default values afterwards.\n ```\n \n \n \n\n- Database & Datalake update (or first-time setup)\n - **IMPORTANT:** Whenever an argument refers to a path (e.g. path/to/whatever), this should better be placed **inside double quotes** to ensure smooth operation (avoid issues with empty spaces in paths).\n ```sh\n py -m exso update -rl path/to/root/datalake -rb path/to/root/database # or...\n py -m exso update -rl path/to/root/datalake -rb path/to/root/database --which ReportName1 ReportName2 ReportName3\n ```\n\n \n \n - You can find the full path to your specified directories by clicking on the address bar of windows explorer\n\n ![img.png](resources/img.png#center)\n\n\n- Extraction / Timezone Conversion / Timeslicing / Visualizing (See section [Node Locators](#node-locators))\n ```sh\n py -m exso query -rb path/to/root/database -loc NodeLocator -output_dir path/to/some/dir -tz desiredTimezone -from YYYY-MM-DD -until YYYY-MM-DD -extract -plot -stacked\n ```\n\n- **Hint**: To avoid too much copy-pasting, the default values for root-lake and root-base are \".\\datalake\" and \".\\database\"\n - This means, that (after having activated the virtual environment), if you navigate to a desired directory, you can launch the command line without specifying the -rl and -rb arguments\n ```sh\n # Activate venv (indicative name of venv = \"deploy_exso\")\n C:\\Users\\theUser>Desktop\\VENVs\\PythonVENVs\\deploy_exso\\Scripts\\activate.bat\n \n # Navigate to desired directory, that will host / is already hosting both the datalake and the database\n (deploy_exso) C:\\Users\\theUser>cd Desktop\\exso_data\n \n # you can now launch exso without specifying root-lake and/or root-base paths\n (deploy_exso) C:\\Users\\theUser\\Desktop\\exso_data>py -m exso update\n ```\n \n- \n\n #### Notes on CLI arguments\n - **<span style=\"color: #2fb2b6\"> -rl</span>**: path to the (desired or existing) root datalake directory\n - **<span style=\"color: #2fb2b6\"> -rb</span>**: path to the (desired or existing) root databake directory\n - **<span style=\"color: #2fb2b6\"> --which</span>**: if the positional argument is **update**, you can optionally enter only specific report names to update (space separated, case-sensitive, as they are listed in the [Implemented Reports](#implemented-reports) section)\n - the <span style=\"color: #2fb2b6\"> **-extract** </span> argument, if passed, must be accompanied by the -output_dir argument\n - the <span style=\"color: #2fb2b6\"> **-plot** </span> argument is valid only if the node corresponding to the given NodeLocator is a file or property node (not a directory)\n - the <span style=\"color: #2fb2b6\"> **-stacked** </span> argument makes the plot (if given) a stacked-area plot\n - the <span style=\"co[exso_CLI.doc](src%2Fexso_CLI.doc)lor: #2fb2b6\"> **-tz** </span> argument is optional. If given, it will convert the database data (which is in UTC) to the specified timezone (including daylight-saving shifts) \n - Recommendation: Don't enter country-specific timezone names. Prefer broader timezones (e.g. EET, CET, UTC, etc.).\n - the <span style=\"color: #2fb2b6\"> **-from** </span> argument is optional (can be combined with -until): format \"YYYY-MM-DD HH:MM\"\n - the <span style=\"color: #2fb2b6\"> **-until** </span> argument is optional (can be combined with -from): format \"YYYY-MM-DD HH:MM\"\n\n #### CLI formal documentation\n\n ```sh\n py -m exso --help\n ```\n\n ```sh\n usage: py -m exso [-h] [-rl ROOT_LAKE] [-rb ROOT_BASE] [--which WHICH [WHICH ...]] [--val_report VAL_REPORT] [--val_dates VAL_DATES [VAL_DATES ...]] [--val_fields VAL_FIELDS [VAL_FIELDS ...]] [-loc QUERY_LOCATOR]\n [-output_dir QUERY_OUTPUT_DIR] [-tz QUERY_TZ] [-from QUERY_FROM] [-until QUERY_UNTIL] [-extract] [-plot] [-stacked] [--decimal_sep DECIMAL_SEP] [--list_sep LIST_SEP]\n {info,update,validate,query,set_system_formats}\n\n positional arguments:\n {info,update,validate,query,set_system_formats}\n\n options:\n -h, --help show this help message and exit\n -rl ROOT_LAKE, --root_lake ROOT_LAKE\n -rb ROOT_BASE, --root_base ROOT_BASE\n --which WHICH [WHICH ...]\n --which argument can be either 'all' (default), or a list of valid report-names (space-separated)\n --val_report VAL_REPORT\n report name you wish to validate.\n --val_dates VAL_DATES [VAL_DATES ...]\n space separated date(s) to validate. format: YYYY-M-D\n --val_fields VAL_FIELDS [VAL_FIELDS ...]\n \"Field(s)\" are the filenames, as to be found in the database folder of a specific report (space separated).\n -loc QUERY_LOCATOR, --query_locator QUERY_LOCATOR\n 'locator' means a unique identifier of database objects. example: root.admie.isp1ispresults, will extract the whole database of this report and transform it / slice it depending on the rest of the options you\n set.\n -output_dir QUERY_OUTPUT_DIR, --query_output_dir QUERY_OUTPUT_DIR\n If specified, it will be used to save the generated plot (if -plot), and/or the extracted timeslice (if -extract).\n -tz QUERY_TZ, --query_tz QUERY_TZ\n -from QUERY_FROM, --query_from QUERY_FROM\n Start date(time) of query (YYYY-M-D [H:M])\n -until QUERY_UNTIL, --query_until QUERY_UNTIL\n End date(time) of query (YYYY-M-D [H:M])\n -extract, --query_extract\n If added, it means you wish to EXTRACT the specified query (among possible other actions)\n -plot, --query_plot If added, it means you wish to PLOT the upstream query (among possible other actions)\n -stacked, --plot_stacked\n If added, it means you wish the PLOT specified, to be a stacked-area plot\n --decimal_sep DECIMAL_SEP\n --list_sep LIST_SEP\n \n ```\n \n\n-----\n# IDE-based Usage\n\n## Basic update\nThe below script will download and insert to the database all (61) currently supported reports. For more information continue reading.\n\n```sh\nfrom exso import Updater\n# define where you want the datalake and the database to be stored in the disk\nroot_lake = r\"path\\to\\desired\\datalake\\directory\" # e.g. r\"C:\\Users\\your_username\\exsodata\\datalake\"\nroot_base = r\"path\\to\\desired\\database\\directory\" # e.g. r\"C:\\Users\\your_username\\exsodata\\database\"\n\n\n# root_lake and root_base can also be pathlib.Path objects\nupd = Updater(root_lake, root_base, all=True)\nupd.run()\n\n```\n#### Update Progress\nProgress bars will be displayed for every kind of operation for each report, as demonstrated in the figure below, for e.g. report = DAM_Results:\n\n![progress_bar_example.png](resources/progress_bar_example.png)\n\n#### Performance & System Requirements\n\nA **full cold-start** process of all 61 reports, might take from **2 to over 5 hours**, depending on internet speed, processing power, memory and non-***exso*** PC usage. Indicative time requirements:\n\n- Pentium-tier processors + 4GB RAM --> ~6 hours (not a good idea in general)\n- 8th-gen high-performance i5 processor + 16GB --> 2.5 hours\n- 12th-gen high-performance i7 processor + 16GB --> 1.5 hours\n\nAfter the first database update (creation actually), each **daily or weekly update** process takes a matter of **a few minutes**.\n\n***exso*** is not optimized for performance (and probably won't be anytime soon):\n\n- Publishing parties keep on changing formats, contents, adding/removing report sections, modifying string representations, etc. \n- This requires very \"low-level\" (in variable level, not machine-level) control over reading, parsing, datetime-indexing in order to work for all reports\n- Frankly performance is not essential: \n - The blow is taken once, at the database initialization: Daily, weekly, monthly updates are really a matter of minutes.\n - The database querying speeds are fast (speed in data operations is always refering to a specific context). The computational cost lies in parsing the datalake files.\n\nThe combined Database & Datalake takes up approximately **4GB of disk space**\n\n-----\n## Implemented Reports\n***exso*** currently supports a total of 61 reports from ADMIE/IPTO and HEnEx. Some of them are obsolete (but still useful for analytics), and some ongoing.\n\nIn order to review the available reports or decide which ones fit your needs and update only those, a Report.Pool object can be useful.\n```sh\nfrom exso import Report\n\n# Instantiate a report Pool object\nrp = Report.Pool()\ndescription = rp.get_text_description() # returns a dictionary of available reports. e.g. {report_name1: report1_description, report_name2: report2_description}\n\n# to get the description of a specific report, you can print:\nreport_name = \"select a valid report name\"\nprint(description[report_name])\n\n# The .get_available() method, returns dataframe with available reports, and their basic metadata\n# To get only a list of names, set only_names <- False\nmetadata = rp.get_available(only_names = False)\n\n```\n![pool_get_available.png](resources/get_available().png)\n\n\n\n\n\n-----\n## Custom Update\n```sh\n# Now, if you conclude that you want e.g. 3 specific reports (at least for now), you can:\ninteresting_reports = ['reportname1', 'reportname2', 'reportname3']\nupd = Updater(root_lake, root_base, reports_pool = rp, some = interesting_reports) # given that you have already instantiated a Report.Pool object\n\n# or, for future use\nupd = Updater(root_lake, root_base, some = interesting_reports) # with no need of instantiating a Report.Pool object\n\n# Very often, a specific datalake-file (excel file) for a given date, may have multiple versions. (e.g. YYYYMMDD_report_01.xlsx, YYYYMMDD_report_02.xlsx)\n# The use_lake_version argument allows to select which version to use.\n# By default it is set to 'latest'. Other options: 'first' or any natural number\nupd.run(use_lake_version = 'latest')\n\n# The below query, will use datalake-files whose version is 4, given that the file actually has 4 or more versions.\n# If the specific file has only two versions, then version 2 will be used.\n# The integer lake version is interpreted as: \"Use this version, or the most recent version available prior to it\"\nupd.run(use_lake_version = 4)\n```\n-----\n# Datalake\nThe datalake consists of raw excel (.xls, or .xlsx, or .zip of .xls*) reports, as published by the publishing parties. \n- Each report is published (is available) over a specific date range (some reports may be no longer actively updated but still useful for historical analysis)\n- Each report is published on a specific frequency (e.g. each day, each week, each month, etc.)\n- Each report file content, spans over various horizons (e.g. one day-long, one week-long, one month-long, etc.)\n- Each report file consists of one or more excel sheets\n- Each report is expressed in a specific timezone (EET, UTC or CET) and may or may not have well-defined daylight-saving switches.\n\n-----\n# Datalake → Database\n\n***exso*** performs:\n- For each report (report-type)\n - Datalake Update (download raw excel files if there are newly published data) \n - For each report file (i.e. for each date that this report was/is published)\n - For each report file excel-sheet\n - Data parsing, data cleaning, datetime conforming, disambiguation, joining\n - Database \"upsert\" (update / insert)\n\n\n-----\n# Database\nThe database, automatically created and maintained through the ***exso*** package, has a tree-like structure.\n\n→ **After the creation of the local database** ([update mode](#basic-update)), you can access it through the ***exso*** API.\n\n- When the tree is initialized, it only retrieves the structure of the directories and files, and the column-names of each .csv file\n- It does not read and load to memory the whole database. This can be done by the __call__ method of a Node object ([see Data Access section](#data-access))\n\n### Database Tree Visualization\n```sh\nfrom exso import Tree\ntree = Tree(root_base)\n\ntree.make_tree()\n# You can make a quick visualization of the directory structure and contents:\ntree.visualize()\n```\n![database_viz.png](resources/database_Viz.png)\n\n-----\n### Datalake vs Database Comparison\n\n#### Reports (aka report-names, or report-types)\n- ***Report Names are almost always the same as the official report-names*** published by the publishing entities, and that's how they appear both in the datalake and in the database\n- When using the \"report\"-kind to access some data, and having doubts on the correct names/strings to use, the name-to-use can be found:\n - in the [Implemented Reports](#implemented-reports)\n - By actually moving through the database directories (reports are directories)\n - Or by using the [Database Tree Visualization](#database-tree-visualization)\n - Or by accessing a \"publisher\"-kind Node-object's .children attributes\n ```sh\n publisher = Tree['root.some_publisher']\n reports = publisher.children\n ```\n\n#### Excel Sheets / Fields\n- Raw excel sheets may have peculiar naming conventions in the raw files.\n- Excel ***sheet names are referred to as \"fields\" in the database-scope***, and may differ from the actual excel-sheet names of the raw datalake\n- When using the \"field\"-kind to access some data (after the database creation/update), and having doubts on the correct names/strings, to use the name-to-use can be found in the actual database folder structure (Fields are directories in the database, NOT files).\n - You can do this by actually moving through the database directories\n - Or by using the [Database Tree Visualization](#database-tree-visualization)\n - Or by accessing a \"report\"-kind Node-object's .children attributes\n ```sh\n report = Tree['root.some_publisher.some_report']\n fields = report.children\n ```\n\n#### Database files\n- Because each raw excel **sheet** sometimes contains a lot of data that are sometimes heterogenonus:\n - Each raw excel sheet (corresponding to a database-field) is broken down in one or more \"subfields\", which are the actual final .csv files\n - When using the \"file\"-kind to access some data (after the database creation/update), and having doubts on the correct names/strings, the name-to-use can be found in the actual database folder structure (\"Files\" are .csv files in the database).\n - You can do this by actually moving through the database directories\n - Or by using the [Database Tree Visualization](#database-tree-visualization)\n - Or by accessing a \"field\"-kind Node-object's .children attributes\n ```sh\n field = Tree['root.some_publisher.some_report.some_field']\n files = field.children\n ```\n\n#### Database Columns / Properties\n- Each database file has one or more columns or \"properties\" (apart from the datetime column)\n- When using the \"property\"-kind to access some data (after the database creation/update), and having doubts on the correct names/strings, the name-to-use can be found in the actual database folder structure (\"Properties\" are columns in the .csv files in the database).\n - You can do this by manually opening the database .csv files\n - **Not available** through the [Database Tree Visualization](#database-tree-visualization)\n - Or by accessing a \"file\"-kind Node-object's .children attributes\n ```sh\n file = Tree['root.some_publisher.some_report.some_field.some_file']\n properties = file.children\n ``` \n\n-----\n## Nodes\nEach branch (a sub-tree) or leaf (an end-Node) of the Database Tree is a Node object. Each Node has specific [attributes](#node-attributes) (both in the literal, and in programming notion):\nThese attributes assist in accessing, manipulating, visualizing or exporting any combination of required system or market properties.\n\n-----\n### Node Attributes\nThe text below serves both a descriptive and a definitive purpose. (So, from now on, the term \"kind\" will have the meaning defined in this section)\n- .name\n\n - a (descriptive) string. Sometimes, names are automatically given from the raw files, while other times there are some alterations. Names are generally non-unique accross the tree, but unique within the children of one node.\n \n- .path \n - physical path in the disk (directory or file)\n \n- .dna\n - a concatenation of all the node's parents, dot-separated, and **case insensitive** (e.g. \"root.henex.DaM_ReSuLtS\")\n \n- .kind\n - In the ***exso*** database, nodes can be of one of the 6 following kinds:\n - \"root\" (parent of all nodes)\n - \"publisher\" (parent of all reports, published by that publisher)\n - \"report\" (the name of the report, e.g. \"ISP1ISPResults\")\n - \"field\" (the name of the sheet of the original report-excel, e.g. ISP_Schedule. *with some exceptions)\n - \"file\" (a csv file containing some or all of the sheet (field)-data e.g. \"Load\")\n - \"property\" (a column of the csv file, e.g. \"Net Load\")\n \n- .parent\n- .children\n- .siblings\n- .ascendants\n- .descendants\n - These are pretty much self-explanatory. They refer to Node objects, or Groups of Node objects (e.g. node.children returns a Group object, but can be accessed as a Node, e.g. node.children.dna, will return a list of dnas of that node's children)\n\n\n-----\n### Node Locators\n\nNode Locators are unique Node identifications. Nodes can be uniquely accessed in more than one ways. The three main node locator types are:\n- DNA locators\n- Path locators\n- Successive children locators\n\nIn all three cases, nodes are accessed through a succession chain:\n\n- **root > publisher > reportName > fieldName > fileName** [>columnName]\n - root: Literally the database root, which is annotated simply as \"root\"\n - publisher: currently admie/henex\n - For reportName, fieldName, fileName, columnName, consult with sections \n - \n - reportName: \n\nFor better demonstration, we'll use the example of ISP Activations/Redispatch, of a non-schedulued ISP (report_name = \"AdhocISPResults\"), only for Hydroelectric Units (see database visualization above)\n\nThe file is called **\"Hydro.csv\"** and is located in the directory **\"root/admie/AdhocISPResults/ISP_Activations\"**. All three methods below will return the desired Node object.\n\n\n- #### DNA Locator\n\n ```sh\n tree['root.admie.adhocispresults.isp_activations.hydro'] # lower/upper case unimportant\n ```\n \n- #### Path Locator\n ```sh\n tree[\"C:\\path_to_root_database\\admie\\AdhocISPResults\\ISP_Activations\\Hydro.csv\"] # exact path must be provided\n ```\n \n- #### Successive children locators\n ```sh\n tree['root']['admie']['AdhocISPResults']['ISP_Activations']['Hydro'] # case sensitive: it accesses the names of the children of each successive node access\n ```\n\n \n\n\n\n\n\n\n-----\n### Data Access\n\nData can be accessed, visualized, manipulated and extracted through Node operations. **(retrieve, export, plot)**\n\n- Once a node has been called in any way (for export, for plot or just retrieval), the whole node is read and stored in memory.\n- **Any timezone, or time-slicing operations, only affect the returned data.**\n- The node's data remain intact and always in UTC timezone\n\nOptional arguments (**kwargs) are common for all three operations, and they can modify the returned time-range (from/to), and the returned timezone.\n- tz_pipe: None or list\n - If a list is provided: [database_timezone, target_timezone, (None)]\n - The databases timezone is UTC and shall be set to UTC\n - The target_timezone can be any pytz compatible timezone: Try to stick to wider zones (EET, CET, UTC, GMT) and not country-wide timezones (e.g. Europe/Athens, etc.)\n - Last argument: \n - If is None, this means, after tz-conversion, truncate the timezone information (e.g. 2022-1-1 00:00 +02:00 will become --> 2022-1-1 00:00)\n - If it doesn't exist (e.g. the list is [database_tz, target_Tz], the returned data will contain the tz-information) \n\n- start_date / end_date: it can accept pandas Timestamps and datetime.datetime objects, but also strings formatted as YYYY-MM-DD HH:MM\n\n\n\n\n#### Usage\n\nRetrieve a node's data:\n\n ```sh\n node_data = node() # data is actually read from the database files\n \n # retrieve data only for January '22, converted to EET timezone:\n # data is retrieved from memory\n node_data_range = node(tz_pipe = ['utc', 'eet', None], start_date = '2022-1-1', end_date = '2022-1-31')\n\n # retrieve all the node's data, in utc (data is retrieved from memory)\n node_data = node() # node's internal data not affected by an intermediate timezone/timeslicing operation.\n ```\n\n \n \n\nExport a node's data:\n ```sh\n node.export(to_path, **kwargs)\n ```\n \n\n\nPlot a node's data (node must be of file- or property-kind):\n- Plotting whole files of considerable size (i.e. > 5-10 MB) may require considerable time\n\n ```sh\n node.plot() # this will plot the file timeseries in UTC over its whole span\n \n # example: stacked-area plot, in EET timezone, only for the period after 1st-Jan 2023. Also save the plot somewhere specified.\n\n node.plot(tz_pipe = ['utc', 'eet', None], start_date = '2023-1-1', show = True, save_path = \"C:/Users/Desktop/my_plot.html\", area = True)\n ```\n \n\n\n-----\n#### Remarks\n\n- The structure of the ***exso*** database, was built as a \"one-size-fits-all\" solution. This may work intuitively in information-rich reports, and less intuitively in very simple reports:\n ```sh\n # access sequence with DNA locators\n tree['root.<publisher>.<report_name>.<sheet/field>.<filename/subfield>]\n ```\n\n \n\n\n- e.g. report = ISP1ISPResults\n - raw datalake file has **8 sheets**, and each sheet has **multiple subfields** (e.g. thermal dispatch, hydro dispatch, load, reserve requirements, a.o.)\n - in order to access e.g. thermal dispatch: tree['root.admie.isp1ispresults.isp_schedule.thermal]\n - --> intuitive\n\n\n- e.g. report = ISP1DayAheadLoadForecast\n - raw datalake file has a **single sheet**, with a **single subfield** (30-min Load Forecast)\n - in order to access it: tree['root.admie.ISP1DayAheadLoadForecast.LoadForecast.LoadForecast']\n - --> seems repetitive, but that's how it is.\n\n\n\n-----\n\n# Examples, Use-cases, Special Attention\n\n## Database Nodes and NodesLocators\n- Accessing nodes, transforming, and custom-exporting.\n```sh\n# Access a node:\nisp1 = tree['root.admie.isp1ispresults'] # dict-like acess, dot separated, case INsensitive. --> the recommended usage\n\n#Alternative methods\nisp1 = tree.get_node(locator = 'root.admie.isp1ispresults') # get_node method accepts more than DNA locators\nisp1 = tree.get_node(locator = 'path/to/database/admie/isp1ispresults')\nisp1 = tree['root']['admie']['ISP1ISPResults']\n\n# you can access a nodes' children\nprint(isp1.children)\n# and get their dnas, names, paths\nprint(isp1.children.dna)\nprint(isp1.children.name)\n\n\nisp1loadforecast_node = tree['root.admie.isp1ispresults.isp_schedule.load']\n\n# Calling a node, returns the nodes' contents (dataframe, or dict (of dicts) of dataframes\nisp1loadforecast_df = isp1loadforecast_node() # returns all available data in UTC timezone\n\n\n## typical use-case: query some specific date range, convert it to a desired timezone and store it somewhere else, to send it to a colleague or do some excel graphs on it\nisp1 = tree['root.admie.isp1ispresults'] # dot separated, case INsensitive\n\n# export the whole data in utc timezone\nisp1.export(to_path = \"where/to/export/full_data_utc\")\n# export custom range in custom timezone\nisp1.export(to_path = \"where/to/export/sliced_data_eet\", tz_pipe = ['utc', 'eet', None], start_date = '2022-1-1 00:00', end_date = '2022-12-31 23:30')\n\n# Note: the start_date:end_date filter is applied AFTER the timezone conversion (if given) \n\n\n# tz_pipe: a list of timezone operations\n\n#The first argument must ALWAYS be UTC (that's the timezone that the csv files are initially on)\n##e.g. 01/01/2021 00:00 will become 01/01/2021 00:00+00:00\n\n#The second argument (optional) is a timezone to convert the data to. (e.g. EET)\n##e.g. 01/01/2021 00:00+00:00 will become: 2021-01-01 02:00+02:00\n\n#The third argument (optional) can only be None, if provided. It means, truncate the timezone information from the converted data \n##e.g. 2021-01-01 02:00+02:00 will become: 2021-01-01 02:00\n\n\n## IMPORTANT:\n# The optional arguments tz_pipe, start_date, end_date do not persist in memory!!\n# This means that, the returned dataframe or dict of dataframe will have the requested charasteristics, but the node keeps its original information\n# That is, if to be re-called without arguments, it will immediately return its raw content: UTC, tz-unaware, full available range\n\n```\n\n## Visualization\n***exso*** utilizes the (extremely helpful and interactive) package [plotly](https://plotly.com/python) for data visualization.\nThe visualization of a Node object is as simple as calling its .plot() method:\n\nGraphs can be zoomed in/out, rescaled, columns can be toggled-on/off in real time.\n\n→ By default, ***exso*** will **omit to plot any columns that are Zero or NaN** over the whole selected timerange, in order to nake the plot lighter.\n\n```sh\nisp1_thermal_gen = t['root.admie.isp1ispresults.isp_schedule.thermal']\nfig = isp1_thermal_gen.plot(area = True, start_date = '2022-1-1', end_date = '2022-1-10', tz_pipe = ['UTC', 'EET', None], show = True, save_path = None)\n\n# the returned figure is of type \"plotly.graph_objs._figure.Figure\", meaning, you can set \"show\"=False, and update the layout with normal plotly usage before displaying it.\n# Some very basic modification-options (title, x&y labels) will be supported directly through the exso.Node object\n```\n![plotly_viz.png](resources/plotly_viz.png)\n\n#### When calling the .plot() method of a Node:\n- Any columns containing only NaN or only zero & nan values are **dropped**, in order to get a cleaner graph.\n- The area = True/False argument, modifies whether the plot will be a stacked area, or a line plot\n- start/end_dates, tz_pipe work exactly as in the __call__ method (node())\n- show = True/False argument, controls whether to automatically display the graph when its rendered (in both cases, a figure object is returned)\n- The save_path argument accepts a Path-like entry (.html), to locally save the graph (regardless of whether show=True/False)\n\n## Disclaimers\n- Any modification on the datalake and database structure, file and directory names, additions/deletions, will probably cause malfunctions.\n- If you must open a database csv in-place (directly from the database directory, e.g. for quick inspection), which you shouldn't, you should at least not perform any saves (even if apparently, no changes were made)\n- If you accidentally modify e.g. a database file, and the Updater malfunctions, you can manually delete the whole reports folder, and re-run the Updater.\n\n## System Formats\nBy default, ***exso*** uses:\n- \",\" (comma) for list separation \n- \".\" (dot) for decimal point\n\nIf your system settings are different, and don't want to change them, you can:\n\n- Modify ***exso***'s default values just **once**\n- Modify ***exso***'s default values **persistently** → recommended option\n\nIf your default/desired system formats do not comply with ***exso***'s defaults, these lines should be placed **before executing** any ***exso*** operations in order to work properly:\n\n- IDE-based usage\n ```sh\n \n import exso\n # modify persistently\n exso._modify_system_formats(decimal_sep = \"your decimal separator\", list_sep = \"your list separator\")\n \n # modify just once\n exso._list_sep = \"your list separator\"\n exso._decimal_sep = \"you decimal point separator\"\n\n \n ```\n- CLI-based usage\n ```sh\n # modify persistently\n py -m exso set_system_formats --decimal_sep \"your decimal separator\" --list_sep \"your list separator\"\n \n # modify just once\n py -m exso <whatever command and config> --decimal_sep \"your decimal separator\" --list_sep \"your list separator\"\n ```\n \n----\n## Data Validation\nThe frequent changes in formats, reporting properties, periodic or one-time mislabeling, timezone ambiguity etc., may at some point result in a report being partially mis-parsed.\nThe Validation module of **exso** makes it easier to inspect and compare the raw datalake files to the database content.\n\nThe validation process creates and launches an excel-file, which conatains:\n- The raw lake data, as they originally were\n- The timeseries-data, on the same timzone that the raw data lake is\n- The timeseries-data on UTC timezone (as they would be stored in the database)\n```sh\nfrom exso import Validation\n\n# simplest setup validation (will validate only the first sheet of the raw datalake file)\nreport_names = 'ISP1ISPResults'\ninspect_dates = '2022-2-19'\nval = Validation(report_names=report_names, dates = inspect_dates, root_lake=root_lake, fields=None)\nval.run()\n\n\n# multi-val setup\nreport_name = 'ISP1ISPResults' # str\nfields = ['ISP_Activations', 'CCGT_Schedule'] # str|list\ninspect_dates = ['2022-2-19', '2022-2-20'] # datetime-like|list\nval = Validation(report_name=report_name, dates = inspect_dates, root_lake=root_lake, fields=fields)\nval.run()\n```\n\n----\n## Features under active Development\n### Data Documentation\nAnother aspect that creates difficulties in utilizing the published data (after one overcomes the sparsity of data), is the lack of detailed documentation per report, field, or property.\n(e.g. The term \"Net Load\" may mean System Load minus pumping load, or Consumption minus RES, or Consumption minus RES minus pumping load, etc.)\nAt this stage, the Data Documentation provided in the ***exso*** package is far from perfect: Data Documentation is currently only on the report-level, providing high-level insights but not detailed disambiguations.\n\n- A custom-made documentation, built as a light non-relational database is currently being developed and will be launched with one of the next versions of ***exso***.\n\n### Analytics API\nThe current setup is oriented around *reports*. An Analytics API currently under development, will facilitate:\n- Seaming properties from different reports of different timeframws (e.g. System Marginal Price to Market Clearing Price)\n- Dedicated reporting and visualization (e.g. Daily System Snapshot of market prices, imports, loads, reserves, balancing, generation mix, etc.)\n- Advanced analytics methods (e.g. Unit Unavailabilities statistics, comparisons, correlations)\n\n### Support for more Reports\nThe next version, will contain some improvements on existing reports, and the addition of Water declaration and NTC reports.\n\n### Support for Linux\nSupport for Linux-based systems is not foreseen at the moment, but feel free to submit a request if needed.\n\n----\n## Tests\n- ***ExSO*** is fairly tested for the envisaged usage, but since the project is not (at least yet) intended for collaborative development, tests are not published.\n- The design philosophy is not to catch all errors imagineable, but rather that basic users will stick to basic/documented usage, and that advanced users know what they're doing\n - e.g. attempting to plot a file beyond its available datetime range, will not\n- From the user's perspective, the [Validation module](#data-validation) is available to assist in validating/trusting that the database accurately reflects the raw datalaek files. \n----\n## Issues\n- Feel free to submit any issues [here](https://github.com/ThanosGkou/exso/issues) or via e-mail\n\n\n----\n## License\n\n<a rel=\"license\" href=\"http://creativecommons.org/licenses/by-nc-nd/4.0/\"><img alt=\"Creative Commons License\" style=\"border-width:0\" src=\"https://i.creativecommons.org/l/by-nc-nd/4.0/88x31.png\" /></a><br />This work is licensed under a <a rel=\"license\" href=\"http://creativecommons.org/licenses/by-nc-nd/4.0/\">Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (CC BY-NC-ND 4.0)</a>\n\nBriefly (without this description being a substitute for the full license or any of its clauses):\n\n**You are free to**:\n\n- Use \u2014 Download/Install/Deploy ***exso***\n- Share \u2014 copy and redistribute the material in any medium or format \n\n\n**Under the following terms**:\n- Attribution \u2014 You must give appropriate credit, provide a link to the license, and indicate if changes were made. You may do so in any reasonable manner, but not in any way that suggests the licensor endorses you or your use. \n- NonCommercial \u2014 You may not use the material for commercial purposes. \n- NoDerivatives \u2014 If you remix, transform, or build upon the material, you may not distribute the modified material. \n- No additional restrictions \u2014 You may not apply legal terms or technological measures that legally restrict others from doing anything the license permits. \n\n\n----\n## Citation\nIf **exso** assists you in making the \"publicly available\" data, actually valuable and accessible, consider citing:\n\n#### APA\n\n - Natsikas, T. (2023). ExSO: market Exchange and Sytem Operation analytical framework (Version 0.0.0) [Computer software]. https://github.com/ThanosGkou/exso\n\n#### BibTeX\n- @software{Natsikas_ExSO_market_Exchange_2023,\nauthor = {Natsikas, Thanos},\nmonth = apr,\ntitle = {{ExSO: market Exchange and Sytem Operation analytical framework}},\nurl = {https://github.com/ThanosGkou/exso},\nversion = {0.0.0},\nyear = {2023}\n}\n\n\n----\n",
"bugtrack_url": null,
"license": "Attribution-NonCommercial-NoDerivatives 4.0 International ======================================================================= Creative Commons Corporation (\"Creative Commons\") is not a law firm and does not provide legal services or legal advice. Distribution of Creative Commons public licenses does not create a lawyer-client or other relationship. Creative Commons makes its licenses and related information available on an \"as-is\" basis. Creative Commons gives no warranties regarding its licenses, any material licensed under their terms and conditions, or any related information. Creative Commons disclaims all liability for damages resulting from their use to the fullest extent possible. Using Creative Commons Public Licenses Creative Commons public licenses provide a standard set of terms and conditions that creators and other rights holders may use to share original works of authorship and other material subject to copyright and certain other rights specified in the public license below. The following considerations are for informational purposes only, are not exhaustive, and do not form part of our licenses. Considerations for licensors: Our public licenses are intended for use by those authorized to give the public permission to use material in ways otherwise restricted by copyright and certain other rights. Our licenses are irrevocable. Licensors should read and understand the terms and conditions of the license they choose before applying it. Licensors should also secure all rights necessary before applying our licenses so that the public can reuse the material as expected. Licensors should clearly mark any material not subject to the license. This includes other CC- licensed material, or material used under an exception or limitation to copyright. More considerations for licensors: wiki.creativecommons.org/Considerations_for_licensors Considerations for the public: By using one of our public licenses, a licensor grants the public permission to use the licensed material under specified terms and conditions. If the licensor's permission is not necessary for any reason--for example, because of any applicable exception or limitation to copyright--then that use is not regulated by the license. Our licenses grant only permissions under copyright and certain other rights that a licensor has authority to grant. Use of the licensed material may still be restricted for other reasons, including because others have copyright or other rights in the material. A licensor may make special requests, such as asking that all changes be marked or described. Although not required by our licenses, you are encouraged to respect those requests where reasonable. More considerations for the public: wiki.creativecommons.org/Considerations_for_licensees ======================================================================= Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International Public License By exercising the Licensed Rights (defined below), You accept and agree to be bound by the terms and conditions of this Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International Public License (\"Public License\"). To the extent this Public License may be interpreted as a contract, You are granted the Licensed Rights in consideration of Your acceptance of these terms and conditions, and the Licensor grants You such rights in consideration of benefits the Licensor receives from making the Licensed Material available under these terms and conditions. Section 1 -- Definitions. a. Adapted Material means material subject to Copyright and Similar Rights that is derived from or based upon the Licensed Material and in which the Licensed Material is translated, altered, arranged, transformed, or otherwise modified in a manner requiring permission under the Copyright and Similar Rights held by the Licensor. For purposes of this Public License, where the Licensed Material is a musical work, performance, or sound recording, Adapted Material is always produced where the Licensed Material is synched in timed relation with a moving image. b. Copyright and Similar Rights means copyright and/or similar rights closely related to copyright including, without limitation, performance, broadcast, sound recording, and Sui Generis Database Rights, without regard to how the rights are labeled or categorized. For purposes of this Public License, the rights specified in Section 2(b)(1)-(2) are not Copyright and Similar Rights. c. Effective Technological Measures means those measures that, in the absence of proper authority, may not be circumvented under laws fulfilling obligations under Article 11 of the WIPO Copyright Treaty adopted on December 20, 1996, and/or similar international agreements. d. Exceptions and Limitations means fair use, fair dealing, and/or any other exception or limitation to Copyright and Similar Rights that applies to Your use of the Licensed Material. e. Licensed Material means the artistic or literary work, database, or other material to which the Licensor applied this Public License. f. Licensed Rights means the rights granted to You subject to the terms and conditions of this Public License, which are limited to all Copyright and Similar Rights that apply to Your use of the Licensed Material and that the Licensor has authority to license. g. Licensor means the individual(s) or entity(ies) granting rights under this Public License. h. NonCommercial means not primarily intended for or directed towards commercial advantage or monetary compensation. For purposes of this Public License, the exchange of the Licensed Material for other material subject to Copyright and Similar Rights by digital file-sharing or similar means is NonCommercial provided there is no payment of monetary compensation in connection with the exchange. i. Share means to provide material to the public by any means or process that requires permission under the Licensed Rights, such as reproduction, public display, public performance, distribution, dissemination, communication, or importation, and to make material available to the public including in ways that members of the public may access the material from a place and at a time individually chosen by them. j. Sui Generis Database Rights means rights other than copyright resulting from Directive 96/9/EC of the European Parliament and of the Council of 11 March 1996 on the legal protection of databases, as amended and/or succeeded, as well as other essentially equivalent rights anywhere in the world. k. You means the individual or entity exercising the Licensed Rights under this Public License. Your has a corresponding meaning. Section 2 -- Scope. a. License grant. 1. Subject to the terms and conditions of this Public License, the Licensor hereby grants You a worldwide, royalty-free, non-sublicensable, non-exclusive, irrevocable license to exercise the Licensed Rights in the Licensed Material to: a. reproduce and Share the Licensed Material, in whole or in part, for NonCommercial purposes only; and b. produce and reproduce, but not Share, Adapted Material for NonCommercial purposes only. 2. Exceptions and Limitations. For the avoidance of doubt, where Exceptions and Limitations apply to Your use, this Public License does not apply, and You do not need to comply with its terms and conditions. 3. Term. The term of this Public License is specified in Section 6(a). 4. Media and formats; technical modifications allowed. The Licensor authorizes You to exercise the Licensed Rights in all media and formats whether now known or hereafter created, and to make technical modifications necessary to do so. The Licensor waives and/or agrees not to assert any right or authority to forbid You from making technical modifications necessary to exercise the Licensed Rights, including technical modifications necessary to circumvent Effective Technological Measures. For purposes of this Public License, simply making modifications authorized by this Section 2(a) (4) never produces Adapted Material. 5. Downstream recipients. a. Offer from the Licensor -- Licensed Material. Every recipient of the Licensed Material automatically receives an offer from the Licensor to exercise the Licensed Rights under the terms and conditions of this Public License. b. No downstream restrictions. You may not offer or impose any additional or different terms or conditions on, or apply any Effective Technological Measures to, the Licensed Material if doing so restricts exercise of the Licensed Rights by any recipient of the Licensed Material. 6. No endorsement. Nothing in this Public License constitutes or may be construed as permission to assert or imply that You are, or that Your use of the Licensed Material is, connected with, or sponsored, endorsed, or granted official status by, the Licensor or others designated to receive attribution as provided in Section 3(a)(1)(A)(i). b. Other rights. 1. Moral rights, such as the right of integrity, are not licensed under this Public License, nor are publicity, privacy, and/or other similar personality rights; however, to the extent possible, the Licensor waives and/or agrees not to assert any such rights held by the Licensor to the limited extent necessary to allow You to exercise the Licensed Rights, but not otherwise. 2. Patent and trademark rights are not licensed under this Public License. 3. To the extent possible, the Licensor waives any right to collect royalties from You for the exercise of the Licensed Rights, whether directly or through a collecting society under any voluntary or waivable statutory or compulsory licensing scheme. In all other cases the Licensor expressly reserves any right to collect such royalties, including when the Licensed Material is used other than for NonCommercial purposes. Section 3 -- License Conditions. Your exercise of the Licensed Rights is expressly made subject to the following conditions. a. Attribution. 1. If You Share the Licensed Material, You must: a. retain the following if it is supplied by the Licensor with the Licensed Material: i. identification of the creator(s) of the Licensed Material and any others designated to receive attribution, in any reasonable manner requested by the Licensor (including by pseudonym if designated); ii. a copyright notice; iii. a notice that refers to this Public License; iv. a notice that refers to the disclaimer of warranties; v. a URI or hyperlink to the Licensed Material to the extent reasonably practicable; b. indicate if You modified the Licensed Material and retain an indication of any previous modifications; and c. indicate the Licensed Material is licensed under this Public License, and include the text of, or the URI or hyperlink to, this Public License. For the avoidance of doubt, You do not have permission under this Public License to Share Adapted Material. 2. You may satisfy the conditions in Section 3(a)(1) in any reasonable manner based on the medium, means, and context in which You Share the Licensed Material. For example, it may be reasonable to satisfy the conditions by providing a URI or hyperlink to a resource that includes the required information. 3. If requested by the Licensor, You must remove any of the information required by Section 3(a)(1)(A) to the extent reasonably practicable. Section 4 -- Sui Generis Database Rights. Where the Licensed Rights include Sui Generis Database Rights that apply to Your use of the Licensed Material: a. for the avoidance of doubt, Section 2(a)(1) grants You the right to extract, reuse, reproduce, and Share all or a substantial portion of the contents of the database for NonCommercial purposes only and provided You do not Share Adapted Material; b. if You include all or a substantial portion of the database contents in a database in which You have Sui Generis Database Rights, then the database in which You have Sui Generis Database Rights (but not its individual contents) is Adapted Material; and c. You must comply with the conditions in Section 3(a) if You Share all or a substantial portion of the contents of the database. For the avoidance of doubt, this Section 4 supplements and does not replace Your obligations under this Public License where the Licensed Rights include other Copyright and Similar Rights. Section 5 -- Disclaimer of Warranties and Limitation of Liability. a. UNLESS OTHERWISE SEPARATELY UNDERTAKEN BY THE LICENSOR, TO THE EXTENT POSSIBLE, THE LICENSOR OFFERS THE LICENSED MATERIAL AS-IS AND AS-AVAILABLE, AND MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND CONCERNING THE LICENSED MATERIAL, WHETHER EXPRESS, IMPLIED, STATUTORY, OR OTHER. THIS INCLUDES, WITHOUT LIMITATION, WARRANTIES OF TITLE, MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, NON-INFRINGEMENT, ABSENCE OF LATENT OR OTHER DEFECTS, ACCURACY, OR THE PRESENCE OR ABSENCE OF ERRORS, WHETHER OR NOT KNOWN OR DISCOVERABLE. WHERE DISCLAIMERS OF WARRANTIES ARE NOT ALLOWED IN FULL OR IN PART, THIS DISCLAIMER MAY NOT APPLY TO YOU. b. TO THE EXTENT POSSIBLE, IN NO EVENT WILL THE LICENSOR BE LIABLE TO YOU ON ANY LEGAL THEORY (INCLUDING, WITHOUT LIMITATION, NEGLIGENCE) OR OTHERWISE FOR ANY DIRECT, SPECIAL, INDIRECT, INCIDENTAL, CONSEQUENTIAL, PUNITIVE, EXEMPLARY, OR OTHER LOSSES, COSTS, EXPENSES, OR DAMAGES ARISING OUT OF THIS PUBLIC LICENSE OR USE OF THE LICENSED MATERIAL, EVEN IF THE LICENSOR HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH LOSSES, COSTS, EXPENSES, OR DAMAGES. WHERE A LIMITATION OF LIABILITY IS NOT ALLOWED IN FULL OR IN PART, THIS LIMITATION MAY NOT APPLY TO YOU. c. The disclaimer of warranties and limitation of liability provided above shall be interpreted in a manner that, to the extent possible, most closely approximates an absolute disclaimer and waiver of all liability. Section 6 -- Term and Termination. a. This Public License applies for the term of the Copyright and Similar Rights licensed here. However, if You fail to comply with this Public License, then Your rights under this Public License terminate automatically. b. Where Your right to use the Licensed Material has terminated under Section 6(a), it reinstates: 1. automatically as of the date the violation is cured, provided it is cured within 30 days of Your discovery of the violation; or 2. upon express reinstatement by the Licensor. For the avoidance of doubt, this Section 6(b) does not affect any right the Licensor may have to seek remedies for Your violations of this Public License. c. For the avoidance of doubt, the Licensor may also offer the Licensed Material under separate terms or conditions or stop distributing the Licensed Material at any time; however, doing so will not terminate this Public License. d. Sections 1, 5, 6, 7, and 8 survive termination of this Public License. Section 7 -- Other Terms and Conditions. a. The Licensor shall not be bound by any additional or different terms or conditions communicated by You unless expressly agreed. b. Any arrangements, understandings, or agreements regarding the Licensed Material not stated herein are separate from and independent of the terms and conditions of this Public License. Section 8 -- Interpretation. a. For the avoidance of doubt, this Public License does not, and shall not be interpreted to, reduce, limit, restrict, or impose conditions on any use of the Licensed Material that could lawfully be made without permission under this Public License. b. To the extent possible, if any provision of this Public License is deemed unenforceable, it shall be automatically reformed to the minimum extent necessary to make it enforceable. If the provision cannot be reformed, it shall be severed from this Public License without affecting the enforceability of the remaining terms and conditions. c. No term or condition of this Public License will be waived and no failure to comply consented to unless expressly agreed to by the Licensor. d. Nothing in this Public License constitutes or may be interpreted as a limitation upon, or waiver of, any privileges and immunities that apply to the Licensor or You, including from the legal processes of any jurisdiction or authority. ======================================================================= Creative Commons is not a party to its public licenses. Notwithstanding, Creative Commons may elect to apply one of its public licenses to material it publishes and in those instances will be considered the \u201cLicensor.\u201d The text of the Creative Commons public licenses is dedicated to the public domain under the CC0 Public Domain Dedication. Except for the limited purpose of indicating that material is shared under a Creative Commons public license or as otherwise permitted by the Creative Commons policies published at creativecommons.org/policies, Creative Commons does not authorize the use of the trademark \"Creative Commons\" or any other trademark or logo of Creative Commons without its prior written consent including, without limitation, in connection with any unauthorized modifications to any of its public licenses or any other arrangements, understandings, or agreements concerning use of licensed material. For the avoidance of doubt, this paragraph does not form part of the public licenses. Creative Commons may be contacted at creativecommons.org. ",
"summary": "A powerful data colllector",
"version": "0.0.0",
"project_urls": {
"Homepage": "https://github.com/ThanosGkou/exso"
},
"split_keywords": [
"admie",
"energy-markets",
"greece",
"henex",
"ipto",
"power-exchange",
"system-operation"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "1a03ff83da05e795de73b6ec4b4c8e242b1a72d562eb2b309a716f534b59bd29",
"md5": "478418d0a792d9633498cea8206f3a99",
"sha256": "3e82970e041a8d5f6fd1de5f3f2e75cff1f9254c6ea36af26b3a1795f89410f2"
},
"downloads": -1,
"filename": "exso-0.0.0-py3-none-any.whl",
"has_sig": false,
"md5_digest": "478418d0a792d9633498cea8206f3a99",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": ">=3.10",
"size": 209862,
"upload_time": "2023-05-28T16:09:10",
"upload_time_iso_8601": "2023-05-28T16:09:10.085704Z",
"url": "https://files.pythonhosted.org/packages/1a/03/ff83da05e795de73b6ec4b4c8e242b1a72d562eb2b309a716f534b59bd29/exso-0.0.0-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "9189f8229409ac9c3fe473d75c0b727c23472190eb7b8a485d61a9de90852ec2",
"md5": "ad41b99c2a8819db814740ad7ffbcd7e",
"sha256": "f9694aeba345390ace382b22928d3b7ca6a7de990a33ffb64bef9e600a0be315"
},
"downloads": -1,
"filename": "exso-0.0.0.tar.gz",
"has_sig": false,
"md5_digest": "ad41b99c2a8819db814740ad7ffbcd7e",
"packagetype": "sdist",
"python_version": "source",
"requires_python": ">=3.10",
"size": 708830,
"upload_time": "2023-05-28T16:09:21",
"upload_time_iso_8601": "2023-05-28T16:09:21.646693Z",
"url": "https://files.pythonhosted.org/packages/91/89/f8229409ac9c3fe473d75c0b727c23472190eb7b8a485d61a9de90852ec2/exso-0.0.0.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2023-05-28 16:09:21",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "ThanosGkou",
"github_project": "exso",
"travis_ci": false,
"coveralls": false,
"github_actions": false,
"requirements": [],
"lcname": "exso"
}