scholarag


Namescholarag JSON
Version 0.0.6 PyPI version JSON
download
home_pageNone
SummaryPipelines for RAG on scientific papers
upload_time2024-09-19 12:19:02
maintainerNone
docs_urlNone
authorBlue Brain Project, EPFL
requires_python>=3.10
licenseNone
keywords
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            # Scholarag

A Retrieval Augmented Generation (RAG) API meant for scientific literature, which includes data management utilities and relevant endpoints for efficiently showcasing papers to your users.

0. [Quickstart](#quickstart)
1. [Setup](#setup)
2. [Database](#database)
3. [ETL Parsing](#etl-parsing)
4. [Funding and Acknowledgement](#funding-and-acknowledgement)

## Quickstart

Our API requires an Elascticsearch/Opensearch database to work, which can be quickly setup using the utilities we offer.

### Step 1: Download the package and start a simple Opensearch (or Elasticsearch) database using Docker.

The script `manage_index.py` can be used to create an index compatible with `scholarag`'s API requirements.

```bash
pip install scholarag
docker run -p 9200:9200 -it -d opensearchproject/opensearch:2.5.0 -Ediscovery.type=single-node -Eplugins.security.disabled=true
manage-index create pmc_paragraphs http://localhost:9200 -v
```

### Step 2: Start an XML parsing server.

The package `scholaretl` has been specifically made to be compatible with Scholarag. It offers an easy-to-use XML and PDF scientific article parsing (see [ETL Parsing](#etl-parsing)). A docker container can easily be started based as follow.

```bash
docker run -p 8080:8080 -d scholaretl:latest
```
The parsing API is reachable on port 8080.

### Step 3: Fill the database.

Multiple scripts are available to fill the database at different scales. A quick and efficient way to do it leverages the script `pmc_parse_and_upload.py`. It fetches all of the papers that PubMed Central uploaded on their s3 bucket on AWS. It is free and doesn't require an AWS account. Using it jointly with the parsing server deployed previously ensures correct formatting of the documents in the database.

```bash
pmc-parse-and-upload http://localhost:9200 http://localhost:8080/parse/jats_xml --start-date 01-01-1800 -v --batch-size=20
```

Specifying a distant start date prevents the search for matching documents to be too long. More details on the script can be found in the documentation [LINK]

### Step 4: Start the `scholarag-api`.

Create a `.env` file that sets the configuration of the scholarag API (see [Setup](#setup)).

```bash
vim .env
```
Required variables can be set up as such

```bash
SCHOLARAG__DB__DB_TYPE=opensearch
SCHOLARAG__DB__INDEX_PARAGRAPHS=pmc_paragraphs
SCHOLARAG__DB__HOST=http://localhost
SCHOLARAG__DB__PORT=9200
```
Such basic configuration unlocks the `Suggestions` and `Retrieval` endpoints. Simply add your OpenAI key under

```bash
SCHOLARAG__GENERATIVE__OPENAI__TOKEN=...
```
to unlock the `Question Answering` ones.

Finally, start the API.

```bash
scholarag-api
```

By default, it opens the API on port 8000. The swagger ui can be accessed at `http://localhost:8000/docs`

You can now interact with the API using HTTP requests. For example

```bash
curl http://localhost:8000/settings
```


## Setup

The `scholarag` API needs environment variables for configuration. Creating a `.env` file is recommended. A template is available under the file `.env.example`.
`scholarag` requires a database for every endpoint. DB specific variables can be set as follows:

```bash
# required
SCHOLARAG__DB__DB_TYPE= # elasticsearch or opensearch
SCHOLARAG__DB__INDEX_PARAGRAPHS= # name of the index containing the documents
SCHOLARAG__DB__HOST=
SCHOLARAG__DB__PORT=

# optional
SCHOLARAG__DB__INDEX_JOURNALS=
SCHOLARAG__DB__USER=
SCHOLARAG__DB__PASSWORD=
```

The variable `SCHOLARAG__DB__INDEX_JOURNALS` defines the name of the index containing the journals' impact factor information. While the variable is optional, the index can be created using the script `impact_factor.py` (TO BE CREATED).

Only setting the DB related variables unlocks the `Suggestions` and `Retrieval` endpoints. To unlock the `Question Answering` ones, the user must specify its OpenAI API key.

```bash
SCHOLARAG__GENERATIVE__OPENAI__TOKEN=sk-...
```

Our API relies on the `bm25` algorithm to retrieve documents from the database, since it is built for large databases and embedding every document would require a massive amount of storage. A reranking mecanism is implemented to compensate for the lack of transformer based semantic analysis. We leverage Cohere's reranker, which also requires an API key.

```bash
SCHOLARAG__RERANKING__COHERE_TOKEN=
```

While optional, the reranker typically improves greatly the retrieval's quality.

Finally, our API optionally supports `Redis` to cache potentially expensive requests. A redis container can be started locally using Docker

```bash
docker run -p 6379:6379 -d redis:latest
```

Redis related variables can be set as follow.

```bash
SCHOLARAG__REDIS__HOST=
SCHOLARAG__REDIS__PORT=
SCHOLARAG__REDIS__EXPIRY= # Time in days. Float.
```

The remaining environment variables are detailed in the documentation (LINK TO ENV VAR DOC).


## Database

The database is the cornerstone of the Scholarag API. We currently support Elasticsearch and Opensearch databases. The indices containing the documents need to have the following mapping in order to be compatible with the API.

```json
{
    "properties": {
        "article_id": {"type": "keyword"},
        "doi": {"type": "keyword"},
        "pmc_id": {"type": "keyword"},
        "pubmed_id": {"type": "keyword"},
        "arxiv_id": {"type": "keyword"},
        "title": {"type": "text"},
        "authors": {"type": "keyword"},
        "journal": {"type": "keyword"},
        "date": {"type": "date", "format": "yyyy-MM-dd"},
        "section": {"type": "keyword"},
        "paragraph_id": {"type": "short"},
        "text": {"type": "text"},
        "article_type": {"type": "keyword"},
    },
}
```
Multiple utilities are made available to fill de db. The external package `bbs-etl` (RENAME) offers a set of fully compatible parsers that convert raw papers into formatted JSONs ready to be inserted into the database. See [ETL Parsing](#etl-parsing) for more details.
The parsers can be directly used with the parsing and uploading scripts available in this project.

* `parse_and_upload.py`: Parses and uploads to DB documents downloaded locally.
```bash
parse-and-upload path parser_url db_url [OPTIONS]
```
* `pmc_parse_and_upload.py`: Parse and upload to DB documents downloaded from PMC's s3 bucket on AWS.
```bash
pmc-parse-and-upload db_url parser_url [OPTIONS]
```
* `pu_producer.py + pu_comsumer.py`: Producer and consumer logic leveraging AWS's SQS service to parse documents from any s3 bucket (requires access right to the bucket). Made for large scale document scraping.
```bash
pu-producer bucket_name queue_url [OPTIONS]
pu-consumer db_url pqrser_url queue_url [OPTIONS]
```

More info on parsing and uploading scripts available at (LINK TO SCRIPTS DOC).

## ETL Parsing

While the Scholarag API doesn't require any ETL logic per se, part of our package offers utilities to fill a database. `scholaretl` is a fully compatible ETL package offering mutliple parsers for scientific articles accessible through an API interface. The easiest way to get it started is through Docker.

```bash
docker run -p 8080:8080 -d scholaretl:latest
```

Optionally, a grobid server can be used to parse PDFs. It requires a Grobid server to be running. To deploy it, simply run

```bash
docker run -p 8070:8070 -d lfoppiano/grobid:0.7.3
```

and pass the server's URL as an environment variable

```bash
docker run -p 8080:8080 -d -e GROBID_URL=http://host.docker.internal:8070 scholaretl:latest
```

List of available endpoints:
* `/pubmed_xml`: parses XMLs coming from PubMed.
* `/jats_xml`: Parses XMLs coming from PMC.
* `/tei_xml`: Parses XMLs produced by Grobid.
* `/xocs_xml`: Parses XMLs coming from Scopus (Elsevier)
* `/pypdf_pdf`: Parses PDFs without keeping the structure of the document.
* `/grobid_pdf`: Parses PDFs keeping the structure of the document.

## Funding and Acknowledgement

The development of this software was supported by funding to the Blue Brain Project, a research center of the École polytechnique fédérale de Lausanne (EPFL), from the Swiss government’s ETH Board of the Swiss Federal Institutes of Technology.

Copyright (c) 2024 Blue Brain Project/EPFL

            

Raw data

            {
    "_id": null,
    "home_page": null,
    "name": "scholarag",
    "maintainer": null,
    "docs_url": null,
    "requires_python": ">=3.10",
    "maintainer_email": null,
    "keywords": null,
    "author": "Blue Brain Project, EPFL",
    "author_email": null,
    "download_url": "https://files.pythonhosted.org/packages/a2/86/60ef6938ef7b140add83af20047caa36c3bb6aa2cf57964393eadd10894f/scholarag-0.0.6.tar.gz",
    "platform": null,
    "description": "# Scholarag\n\nA Retrieval Augmented Generation (RAG) API meant for scientific literature, which includes data management utilities and relevant endpoints for efficiently showcasing papers to your users.\n\n0. [Quickstart](#quickstart)\n1. [Setup](#setup)\n2. [Database](#database)\n3. [ETL Parsing](#etl-parsing)\n4. [Funding and Acknowledgement](#funding-and-acknowledgement)\n\n## Quickstart\n\nOur API requires an Elascticsearch/Opensearch database to work, which can be quickly setup using the utilities we offer.\n\n### Step 1: Download the package and start a simple Opensearch (or Elasticsearch) database using Docker.\n\nThe script `manage_index.py` can be used to create an index compatible with `scholarag`'s API requirements.\n\n```bash\npip install scholarag\ndocker run -p 9200:9200 -it -d opensearchproject/opensearch:2.5.0 -Ediscovery.type=single-node -Eplugins.security.disabled=true\nmanage-index create pmc_paragraphs http://localhost:9200 -v\n```\n\n### Step 2: Start an XML parsing server.\n\nThe package `scholaretl` has been specifically made to be compatible with Scholarag. It offers an easy-to-use XML and PDF scientific article parsing (see [ETL Parsing](#etl-parsing)). A docker container can easily be started based as follow.\n\n```bash\ndocker run -p 8080:8080 -d scholaretl:latest\n```\nThe parsing API is reachable on port 8080.\n\n### Step 3: Fill the database.\n\nMultiple scripts are available to fill the database at different scales. A quick and efficient way to do it leverages the script `pmc_parse_and_upload.py`. It fetches all of the papers that PubMed Central uploaded on their s3 bucket on AWS. It is free and doesn't require an AWS account. Using it jointly with the parsing server deployed previously ensures correct formatting of the documents in the database.\n\n```bash\npmc-parse-and-upload http://localhost:9200 http://localhost:8080/parse/jats_xml --start-date 01-01-1800 -v --batch-size=20\n```\n\nSpecifying a distant start date prevents the search for matching documents to be too long. More details on the script can be found in the documentation [LINK]\n\n### Step 4: Start the `scholarag-api`.\n\nCreate a `.env` file that sets the configuration of the scholarag API (see [Setup](#setup)).\n\n```bash\nvim .env\n```\nRequired variables can be set up as such\n\n```bash\nSCHOLARAG__DB__DB_TYPE=opensearch\nSCHOLARAG__DB__INDEX_PARAGRAPHS=pmc_paragraphs\nSCHOLARAG__DB__HOST=http://localhost\nSCHOLARAG__DB__PORT=9200\n```\nSuch basic configuration unlocks the `Suggestions` and `Retrieval` endpoints. Simply add your OpenAI key under\n\n```bash\nSCHOLARAG__GENERATIVE__OPENAI__TOKEN=...\n```\nto unlock the `Question Answering` ones.\n\nFinally, start the API.\n\n```bash\nscholarag-api\n```\n\nBy default, it opens the API on port 8000. The swagger ui can be accessed at `http://localhost:8000/docs`\n\nYou can now interact with the API using HTTP requests. For example\n\n```bash\ncurl http://localhost:8000/settings\n```\n\n\n## Setup\n\nThe `scholarag` API needs environment variables for configuration. Creating a `.env` file is recommended. A template is available under the file `.env.example`.\n`scholarag` requires a database for every endpoint. DB specific variables can be set as follows:\n\n```bash\n# required\nSCHOLARAG__DB__DB_TYPE= # elasticsearch or opensearch\nSCHOLARAG__DB__INDEX_PARAGRAPHS= # name of the index containing the documents\nSCHOLARAG__DB__HOST=\nSCHOLARAG__DB__PORT=\n\n# optional\nSCHOLARAG__DB__INDEX_JOURNALS=\nSCHOLARAG__DB__USER=\nSCHOLARAG__DB__PASSWORD=\n```\n\nThe variable `SCHOLARAG__DB__INDEX_JOURNALS` defines the name of the index containing the journals' impact factor information. While the variable is optional, the index can be created using the script `impact_factor.py` (TO BE CREATED).\n\nOnly setting the DB related variables unlocks the `Suggestions` and `Retrieval` endpoints. To unlock the `Question Answering` ones, the user must specify its OpenAI API key.\n\n```bash\nSCHOLARAG__GENERATIVE__OPENAI__TOKEN=sk-...\n```\n\nOur API relies on the `bm25` algorithm to retrieve documents from the database, since it is built for large databases and embedding every document would require a massive amount of storage. A reranking mecanism is implemented to compensate for the lack of transformer based semantic analysis. We leverage Cohere's reranker, which also requires an API key.\n\n```bash\nSCHOLARAG__RERANKING__COHERE_TOKEN=\n```\n\nWhile optional, the reranker typically improves greatly the retrieval's quality.\n\nFinally, our API optionally supports `Redis` to cache potentially expensive requests. A redis container can be started locally using Docker\n\n```bash\ndocker run -p 6379:6379 -d redis:latest\n```\n\nRedis related variables can be set as follow.\n\n```bash\nSCHOLARAG__REDIS__HOST=\nSCHOLARAG__REDIS__PORT=\nSCHOLARAG__REDIS__EXPIRY= # Time in days. Float.\n```\n\nThe remaining environment variables are detailed in the documentation (LINK TO ENV VAR DOC).\n\n\n## Database\n\nThe database is the cornerstone of the Scholarag API. We currently support Elasticsearch and Opensearch databases. The indices containing the documents need to have the following mapping in order to be compatible with the API.\n\n```json\n{\n    \"properties\": {\n        \"article_id\": {\"type\": \"keyword\"},\n        \"doi\": {\"type\": \"keyword\"},\n        \"pmc_id\": {\"type\": \"keyword\"},\n        \"pubmed_id\": {\"type\": \"keyword\"},\n        \"arxiv_id\": {\"type\": \"keyword\"},\n        \"title\": {\"type\": \"text\"},\n        \"authors\": {\"type\": \"keyword\"},\n        \"journal\": {\"type\": \"keyword\"},\n        \"date\": {\"type\": \"date\", \"format\": \"yyyy-MM-dd\"},\n        \"section\": {\"type\": \"keyword\"},\n        \"paragraph_id\": {\"type\": \"short\"},\n        \"text\": {\"type\": \"text\"},\n        \"article_type\": {\"type\": \"keyword\"},\n    },\n}\n```\nMultiple utilities are made available to fill de db. The external package `bbs-etl` (RENAME) offers a set of fully compatible parsers that convert raw papers into formatted JSONs ready to be inserted into the database. See [ETL Parsing](#etl-parsing) for more details.\nThe parsers can be directly used with the parsing and uploading scripts available in this project.\n\n* `parse_and_upload.py`: Parses and uploads to DB documents downloaded locally.\n```bash\nparse-and-upload path parser_url db_url [OPTIONS]\n```\n* `pmc_parse_and_upload.py`: Parse and upload to DB documents downloaded from PMC's s3 bucket on AWS.\n```bash\npmc-parse-and-upload db_url parser_url [OPTIONS]\n```\n* `pu_producer.py + pu_comsumer.py`: Producer and consumer logic leveraging AWS's SQS service to parse documents from any s3 bucket (requires access right to the bucket). Made for large scale document scraping.\n```bash\npu-producer bucket_name queue_url [OPTIONS]\npu-consumer db_url pqrser_url queue_url [OPTIONS]\n```\n\nMore info on parsing and uploading scripts available at (LINK TO SCRIPTS DOC).\n\n## ETL Parsing\n\nWhile the Scholarag API doesn't require any ETL logic per se, part of our package offers utilities to fill a database. `scholaretl` is a fully compatible ETL package offering mutliple parsers for scientific articles accessible through an API interface. The easiest way to get it started is through Docker.\n\n```bash\ndocker run -p 8080:8080 -d scholaretl:latest\n```\n\nOptionally, a grobid server can be used to parse PDFs. It requires a Grobid server to be running. To deploy it, simply run\n\n```bash\ndocker run -p 8070:8070 -d lfoppiano/grobid:0.7.3\n```\n\nand pass the server's URL as an environment variable\n\n```bash\ndocker run -p 8080:8080 -d -e GROBID_URL=http://host.docker.internal:8070 scholaretl:latest\n```\n\nList of available endpoints:\n* `/pubmed_xml`: parses XMLs coming from PubMed.\n* `/jats_xml`: Parses XMLs coming from PMC.\n* `/tei_xml`: Parses XMLs produced by Grobid.\n* `/xocs_xml`: Parses XMLs coming from Scopus (Elsevier)\n* `/pypdf_pdf`: Parses PDFs without keeping the structure of the document.\n* `/grobid_pdf`: Parses PDFs keeping the structure of the document.\n\n## Funding and Acknowledgement\n\nThe development of this software was supported by funding to the Blue Brain Project, a research center of the \u00c9cole polytechnique f\u00e9d\u00e9rale de Lausanne (EPFL), from the Swiss government\u2019s ETH Board of the Swiss Federal Institutes of Technology.\n\nCopyright (c) 2024 Blue Brain Project/EPFL\n",
    "bugtrack_url": null,
    "license": null,
    "summary": "Pipelines for RAG on scientific papers",
    "version": "0.0.6",
    "project_urls": null,
    "split_keywords": [],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "9056787c27edbe8b9ff0a0cb36237e8a4984ef9657606c87513f38e35884c1c2",
                "md5": "25833ca230bbedf120bdade0e571e124",
                "sha256": "75b81cc9df98f60759896783439a7c5f523b4bf099ae52ca60f926c901d22f61"
            },
            "downloads": -1,
            "filename": "scholarag-0.0.6-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "25833ca230bbedf120bdade0e571e124",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.10",
            "size": 76049,
            "upload_time": "2024-09-19T12:19:01",
            "upload_time_iso_8601": "2024-09-19T12:19:01.095926Z",
            "url": "https://files.pythonhosted.org/packages/90/56/787c27edbe8b9ff0a0cb36237e8a4984ef9657606c87513f38e35884c1c2/scholarag-0.0.6-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "a28660ef6938ef7b140add83af20047caa36c3bb6aa2cf57964393eadd10894f",
                "md5": "758f5caee413b5fc947d62afa73ba4e7",
                "sha256": "1cc94663d3b6c631620f88ba7376e4603cf71664f030d1e12557786e9074a45d"
            },
            "downloads": -1,
            "filename": "scholarag-0.0.6.tar.gz",
            "has_sig": false,
            "md5_digest": "758f5caee413b5fc947d62afa73ba4e7",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.10",
            "size": 67457,
            "upload_time": "2024-09-19T12:19:02",
            "upload_time_iso_8601": "2024-09-19T12:19:02.640721Z",
            "url": "https://files.pythonhosted.org/packages/a2/86/60ef6938ef7b140add83af20047caa36c3bb6aa2cf57964393eadd10894f/scholarag-0.0.6.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-09-19 12:19:02",
    "github": false,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "lcname": "scholarag"
}
        
Elapsed time: 0.36338s