# langchain-utils
LangChain Utilities
- [langchain-utils](#langchain-utils)
- [Prompt generation using LangChain document loaders](#prompt-generation-using-langchain-document-loaders)
- [Demos](#demos)
- [`urlprompt`](#urlprompt)
- [`pdfprompt`](#pdfprompt)
- [`ytprompt`](#ytprompt)
- [`textprompt`](#textprompt)
- [`htmlprompt`](#htmlprompt)
- [Installation](#installation)
- [pipx](#pipx)
- [pip](#pip)
- [Develop](#develop)
## Prompt generation using LangChain document loaders
Do you find yourself frequently copy-pasting texts from the web / PDFs / other documents into ChatGPT?
If yes, these tools are for you!
Optimized to feed into a chat interface (like ChatGPT) manually in one or multiple (to get around context length limits) goes.
Basically, the prompts generated look like this:
```python
REPLY_OK_IF_YOU_READ_TEMPLATE = '''
Below is {what}, reply "OK" if you read:
"""
{content}
"""
'''.strip()
```
You can feed it directly to a chat interface like ChatGPT, and ask follow up questions about it.
See [`prompts.py`](./langchain_utils/prompts.py) for other variations.
### Demos
- Loading `https://github.com/tddschn/langchain-utils` and copy to clipboard:
<!-- create a video tag with https://user-images.githubusercontent.com/45612704/231729153-341bd962-28cc-40a3-af8b-91e038ccaf6c.mp4 -->
<video src="https://user-images.githubusercontent.com/45612704/231729153-341bd962-28cc-40a3-af8b-91e038ccaf6c.mp4" controls width="100%"></video>
- Load 3 pages of a pdf file, open each part for inspection before copying, and optionally merge 3 pages into 2 prompts that wouldn't go over the `gpt-3.5-turbo`'s context length limit with langchain's `TokenTextSplitter`.
<!-- for https://user-images.githubusercontent.com/45612704/231731553-63cf3cef-a210-4761-8ca3-dd47bedc3393.mp4 -->
<video src="https://user-images.githubusercontent.com/45612704/231731553-63cf3cef-a210-4761-8ca3-dd47bedc3393.mp4" controls width="100%"></video>
### `urlprompt`
```
$ urlprompt --help
usage: urlprompt [-h] [-V] [-c] [-e] [-m model] [-S] [-s chunk_size]
[-P PARTS [PARTS ...]] [-r] [-R]
[--print-percentage-non-ascii] [-n] [-w WHAT] [-M] [-j] [-g]
[--github-path GITHUB_PATH]
[--github-revision GITHUB_REVISION]
URL
Get a prompt consisting the text content of a webpage
positional arguments:
URL URL to the webpage
options:
-h, --help show this help message and exit
-V, --version show program's version number and exit
-c, --copy Copy the prompt to clipboard (default: False)
-e, --edit Edit the prompt and copy manually (default: False)
-m model, --model model
Model to use (default: gpt-3.5-turbo)
-S, --no-split Do not split the prompt into multiple parts (use this
if the model has a really large context size)
(default: False)
-s chunk_size, --chunk-size chunk_size
Chunk size when splitting transcript, also used to
determine whether to split, defaults to 1/2 of the
context length limit of the model (default: None)
-P PARTS [PARTS ...], --parts PARTS [PARTS ...]
Parts to select in the processes list of Documents
(default: None)
-r, --raw Wraps the content in triple quotes with no extra text
(default: False)
-R, --raw-no-quotes Output the content only (default: False)
--print-percentage-non-ascii
Print percentage of non-ascii characters (default:
False)
-n, --dry-run Dry run (default: False)
-w WHAT, --what WHAT Initial knowledge you want to insert before the PDF
content in the prompt (default: the content of a
webpage)
-M, --merge Merge contents of all pages before processing
(default: False)
-j, --javascript Use JavaScript to render the page (default: False)
-g, --github Load the raw file from a GitHub URL (default: False)
--github-path GITHUB_PATH
Path to the GitHub file (default: README.md)
--github-revision GITHUB_REVISION
Revision for the GitHub file (default: master)
```
### `pdfprompt`
```
$ pdfprompt --help
usage: pdfprompt [-h] [-V] [-c] [-e] [-m model] [-S] [-s chunk_size]
[-P PARTS [PARTS ...]] [-r] [-R]
[--print-percentage-non-ascii] [-n] [-p PAGES [PAGES ...]]
[-l PAGE_SLICE] [-M] [-w WHAT] [-o] [-O] [-L OCR_LANGUAGE]
PDF Path
Get a prompt consisting the text content of a PDF file
positional arguments:
PDF Path Path to the PDF file
options:
-h, --help show this help message and exit
-V, --version show program's version number and exit
-c, --copy Copy the prompt to clipboard (default: False)
-e, --edit Edit the prompt and copy manually (default: False)
-m model, --model model
Model to use (default: gpt-3.5-turbo)
-S, --no-split Do not split the prompt into multiple parts (use this
if the model has a really large context size)
(default: False)
-s chunk_size, --chunk-size chunk_size
Chunk size when splitting transcript, also used to
determine whether to split, defaults to 1/2 of the
context length limit of the model (default: None)
-P PARTS [PARTS ...], --parts PARTS [PARTS ...]
Parts to select in the processes list of Documents
(default: None)
-r, --raw Wraps the content in triple quotes with no extra text
(default: False)
-R, --raw-no-quotes Output the content only (default: False)
--print-percentage-non-ascii
Print percentage of non-ascii characters (default:
False)
-n, --dry-run Dry run (default: False)
-p PAGES [PAGES ...], --pages PAGES [PAGES ...]
Only include specified page numbers (default: None)
-l PAGE_SLICE, --page-slice PAGE_SLICE
Use Python slice syntax to select page numbers (e.g.
1:3, 1:10:2, etc.) (default: None)
-M, --merge Merge contents of all pages before processing
(default: False)
-w WHAT, --what WHAT Initial knowledge you want to insert before the PDF
content in the prompt (default: the content of a PDF
file)
-o, --fallback-ocr Use OCR as fallback if no text detected on page,
please set TESSDATA_PREFIX environment variable to the
path of your tesseract data directory (default: False)
-O, --force-ocr Force OCR on all pages (default: False)
-L OCR_LANGUAGE, --ocr-language OCR_LANGUAGE
Language to use for Tesseract OCR (like eng, chi_sim,
chi_tra, chi_tra_vert etc.)) (default: eng)
```
### `ytprompt`
```
$ ytprompt --help
usage: ytprompt [-h] [-V] [-c] [-e] [-m model] [-S] [-s chunk_size]
[-P PARTS [PARTS ...]] [-r] [-R]
[--print-percentage-non-ascii] [-n]
YouTube URL
Get a prompt consisting Title and Transcript of a YouTube Video
positional arguments:
YouTube URL YouTube URL
options:
-h, --help show this help message and exit
-V, --version show program's version number and exit
-c, --copy Copy the prompt to clipboard (default: False)
-e, --edit Edit the prompt and copy manually (default: False)
-m model, --model model
Model to use (default: gpt-3.5-turbo)
-S, --no-split Do not split the prompt into multiple parts (use this
if the model has a really large context size)
(default: False)
-s chunk_size, --chunk-size chunk_size
Chunk size when splitting transcript, also used to
determine whether to split, defaults to 1/2 of the
context length limit of the model (default: None)
-P PARTS [PARTS ...], --parts PARTS [PARTS ...]
Parts to select in the processes list of Documents
(default: None)
-r, --raw Wraps the content in triple quotes with no extra text
(default: False)
-R, --raw-no-quotes Output the content only (default: False)
--print-percentage-non-ascii
Print percentage of non-ascii characters (default:
False)
-n, --dry-run Dry run (default: False)
```
### `textprompt`
```
$ textprompt --help
usage: textprompt [-h] [-V] [-c] [-e] [-m model] [-S] [-s chunk_size]
[-P PARTS [PARTS ...]] [-r] [-R]
[--print-percentage-non-ascii] [-n] [-C] [-w WHAT] [-M]
[PATH ...]
Get a prompt from text files
positional arguments:
PATH Paths to the text files, or stdin if not provided
(default: None)
options:
-h, --help show this help message and exit
-V, --version show program's version number and exit
-c, --copy Copy the prompt to clipboard (default: False)
-e, --edit Edit the prompt and copy manually (default: False)
-m model, --model model
Model to use (default: gpt-3.5-turbo)
-S, --no-split Do not split the prompt into multiple parts (use this
if the model has a really large context size)
(default: False)
-s chunk_size, --chunk-size chunk_size
Chunk size when splitting transcript, also used to
determine whether to split, defaults to 1/2 of the
context length limit of the model (default: None)
-P PARTS [PARTS ...], --parts PARTS [PARTS ...]
Parts to select in the processes list of Documents
(default: None)
-r, --raw Wraps the content in triple quotes with no extra text
(default: False)
-R, --raw-no-quotes Output the content only (default: False)
--print-percentage-non-ascii
Print percentage of non-ascii characters (default:
False)
-n, --dry-run Dry run (default: False)
-C, --from-clipboard Load text from clipboard (default: False)
-w WHAT, --what WHAT Initial knowledge you want to insert before the PDF
content in the prompt (default: the content of a
document)
-M, --merge Merge contents of all pages before processing
(default: False)
```
### `htmlprompt`
```
$ htmlprompt --help
usage: htmlprompt [-h] [-V] [-c] [-e] [-m model] [-S] [-s chunk_size]
[-P PARTS [PARTS ...]] [-r] [-R]
[--print-percentage-non-ascii] [-n] [-C] [-w WHAT] [-M]
[PATH ...]
Get a prompt from html files
positional arguments:
PATH Paths to the html files, or stdin if not provided
(default: None)
options:
-h, --help show this help message and exit
-V, --version show program's version number and exit
-c, --copy Copy the prompt to clipboard (default: False)
-e, --edit Edit the prompt and copy manually (default: False)
-m model, --model model
Model to use (default: gpt-3.5-turbo)
-S, --no-split Do not split the prompt into multiple parts (use this
if the model has a really large context size)
(default: False)
-s chunk_size, --chunk-size chunk_size
Chunk size when splitting transcript, also used to
determine whether to split, defaults to 1/2 of the
context length limit of the model (default: None)
-P PARTS [PARTS ...], --parts PARTS [PARTS ...]
Parts to select in the processes list of Documents
(default: None)
-r, --raw Wraps the content in triple quotes with no extra text
(default: False)
-R, --raw-no-quotes Output the content only (default: False)
--print-percentage-non-ascii
Print percentage of non-ascii characters (default:
False)
-n, --dry-run Dry run (default: False)
-C, --from-clipboard Load text from clipboard (default: False)
-w WHAT, --what WHAT Initial knowledge you want to insert before the PDF
content in the prompt (default: the text content of a
html file)
-M, --merge Merge contents of all pages before processing
(default: False)
```
## Installation
### pipx
This is the recommended installation method.
```
$ pipx install langchain-utils
```
### [pip](https://pypi.org/project/langchain-utils/)
```
$ pip install langchain-utils
```
## Develop
```
$ git clone https://github.com/tddschn/langchain-utils.git
$ cd langchain-utils
$ poetry install
```
Raw data
{
"_id": null,
"home_page": "https://github.com/tddschn/langchain-utils",
"name": "langchain-utils",
"maintainer": "",
"docs_url": null,
"requires_python": ">=3.11,<4.0",
"maintainer_email": "",
"keywords": "langchain,utils,LLM,prompts,CLI",
"author": "Teddy Xinyuan Chen",
"author_email": "45612704+tddschn@users.noreply.github.com",
"download_url": "https://files.pythonhosted.org/packages/ed/ed/d92fc27849f910f6cc4474b0aabe6db5e79dd32a5006373c773650ba9952/langchain_utils-0.5.6.tar.gz",
"platform": null,
"description": "# langchain-utils\n\nLangChain Utilities\n\n- [langchain-utils](#langchain-utils)\n - [Prompt generation using LangChain document loaders](#prompt-generation-using-langchain-document-loaders)\n - [Demos](#demos)\n - [`urlprompt`](#urlprompt)\n - [`pdfprompt`](#pdfprompt)\n - [`ytprompt`](#ytprompt)\n - [`textprompt`](#textprompt)\n - [`htmlprompt`](#htmlprompt)\n - [Installation](#installation)\n - [pipx](#pipx)\n - [pip](#pip)\n - [Develop](#develop)\n\n## Prompt generation using LangChain document loaders\n\nDo you find yourself frequently copy-pasting texts from the web / PDFs / other documents into ChatGPT?\n\nIf yes, these tools are for you!\n\nOptimized to feed into a chat interface (like ChatGPT) manually in one or multiple (to get around context length limits) goes.\n\nBasically, the prompts generated look like this:\n\n```python\nREPLY_OK_IF_YOU_READ_TEMPLATE = '''\nBelow is {what}, reply \"OK\" if you read:\n\n\"\"\"\n{content}\n\"\"\"\n'''.strip()\n```\n\nYou can feed it directly to a chat interface like ChatGPT, and ask follow up questions about it.\n\nSee [`prompts.py`](./langchain_utils/prompts.py) for other variations.\n\n### Demos\n\n- Loading `https://github.com/tddschn/langchain-utils` and copy to clipboard:\n\n<!-- create a video tag with https://user-images.githubusercontent.com/45612704/231729153-341bd962-28cc-40a3-af8b-91e038ccaf6c.mp4 -->\n\n<video src=\"https://user-images.githubusercontent.com/45612704/231729153-341bd962-28cc-40a3-af8b-91e038ccaf6c.mp4\" controls width=\"100%\"></video>\n\n- Load 3 pages of a pdf file, open each part for inspection before copying, and optionally merge 3 pages into 2 prompts that wouldn't go over the `gpt-3.5-turbo`'s context length limit with langchain's `TokenTextSplitter`.\n\n<!-- for https://user-images.githubusercontent.com/45612704/231731553-63cf3cef-a210-4761-8ca3-dd47bedc3393.mp4 -->\n\n<video src=\"https://user-images.githubusercontent.com/45612704/231731553-63cf3cef-a210-4761-8ca3-dd47bedc3393.mp4\" controls width=\"100%\"></video>\n\n### `urlprompt`\n\n```\n$ urlprompt --help\n\nusage: urlprompt [-h] [-V] [-c] [-e] [-m model] [-S] [-s chunk_size]\n [-P PARTS [PARTS ...]] [-r] [-R]\n [--print-percentage-non-ascii] [-n] [-w WHAT] [-M] [-j] [-g]\n [--github-path GITHUB_PATH]\n [--github-revision GITHUB_REVISION]\n URL\n\nGet a prompt consisting the text content of a webpage\n\npositional arguments:\n URL URL to the webpage\n\noptions:\n -h, --help show this help message and exit\n -V, --version show program's version number and exit\n -c, --copy Copy the prompt to clipboard (default: False)\n -e, --edit Edit the prompt and copy manually (default: False)\n -m model, --model model\n Model to use (default: gpt-3.5-turbo)\n -S, --no-split Do not split the prompt into multiple parts (use this\n if the model has a really large context size)\n (default: False)\n -s chunk_size, --chunk-size chunk_size\n Chunk size when splitting transcript, also used to\n determine whether to split, defaults to 1/2 of the\n context length limit of the model (default: None)\n -P PARTS [PARTS ...], --parts PARTS [PARTS ...]\n Parts to select in the processes list of Documents\n (default: None)\n -r, --raw Wraps the content in triple quotes with no extra text\n (default: False)\n -R, --raw-no-quotes Output the content only (default: False)\n --print-percentage-non-ascii\n Print percentage of non-ascii characters (default:\n False)\n -n, --dry-run Dry run (default: False)\n -w WHAT, --what WHAT Initial knowledge you want to insert before the PDF\n content in the prompt (default: the content of a\n webpage)\n -M, --merge Merge contents of all pages before processing\n (default: False)\n -j, --javascript Use JavaScript to render the page (default: False)\n -g, --github Load the raw file from a GitHub URL (default: False)\n --github-path GITHUB_PATH\n Path to the GitHub file (default: README.md)\n --github-revision GITHUB_REVISION\n Revision for the GitHub file (default: master)\n\n```\n### `pdfprompt`\n\n```\n$ pdfprompt --help\n\nusage: pdfprompt [-h] [-V] [-c] [-e] [-m model] [-S] [-s chunk_size]\n [-P PARTS [PARTS ...]] [-r] [-R]\n [--print-percentage-non-ascii] [-n] [-p PAGES [PAGES ...]]\n [-l PAGE_SLICE] [-M] [-w WHAT] [-o] [-O] [-L OCR_LANGUAGE]\n PDF Path\n\nGet a prompt consisting the text content of a PDF file\n\npositional arguments:\n PDF Path Path to the PDF file\n\noptions:\n -h, --help show this help message and exit\n -V, --version show program's version number and exit\n -c, --copy Copy the prompt to clipboard (default: False)\n -e, --edit Edit the prompt and copy manually (default: False)\n -m model, --model model\n Model to use (default: gpt-3.5-turbo)\n -S, --no-split Do not split the prompt into multiple parts (use this\n if the model has a really large context size)\n (default: False)\n -s chunk_size, --chunk-size chunk_size\n Chunk size when splitting transcript, also used to\n determine whether to split, defaults to 1/2 of the\n context length limit of the model (default: None)\n -P PARTS [PARTS ...], --parts PARTS [PARTS ...]\n Parts to select in the processes list of Documents\n (default: None)\n -r, --raw Wraps the content in triple quotes with no extra text\n (default: False)\n -R, --raw-no-quotes Output the content only (default: False)\n --print-percentage-non-ascii\n Print percentage of non-ascii characters (default:\n False)\n -n, --dry-run Dry run (default: False)\n -p PAGES [PAGES ...], --pages PAGES [PAGES ...]\n Only include specified page numbers (default: None)\n -l PAGE_SLICE, --page-slice PAGE_SLICE\n Use Python slice syntax to select page numbers (e.g.\n 1:3, 1:10:2, etc.) (default: None)\n -M, --merge Merge contents of all pages before processing\n (default: False)\n -w WHAT, --what WHAT Initial knowledge you want to insert before the PDF\n content in the prompt (default: the content of a PDF\n file)\n -o, --fallback-ocr Use OCR as fallback if no text detected on page,\n please set TESSDATA_PREFIX environment variable to the\n path of your tesseract data directory (default: False)\n -O, --force-ocr Force OCR on all pages (default: False)\n -L OCR_LANGUAGE, --ocr-language OCR_LANGUAGE\n Language to use for Tesseract OCR (like eng, chi_sim,\n chi_tra, chi_tra_vert etc.)) (default: eng)\n\n```\n### `ytprompt`\n\n```\n$ ytprompt --help\n\nusage: ytprompt [-h] [-V] [-c] [-e] [-m model] [-S] [-s chunk_size]\n [-P PARTS [PARTS ...]] [-r] [-R]\n [--print-percentage-non-ascii] [-n]\n YouTube URL\n\nGet a prompt consisting Title and Transcript of a YouTube Video\n\npositional arguments:\n YouTube URL YouTube URL\n\noptions:\n -h, --help show this help message and exit\n -V, --version show program's version number and exit\n -c, --copy Copy the prompt to clipboard (default: False)\n -e, --edit Edit the prompt and copy manually (default: False)\n -m model, --model model\n Model to use (default: gpt-3.5-turbo)\n -S, --no-split Do not split the prompt into multiple parts (use this\n if the model has a really large context size)\n (default: False)\n -s chunk_size, --chunk-size chunk_size\n Chunk size when splitting transcript, also used to\n determine whether to split, defaults to 1/2 of the\n context length limit of the model (default: None)\n -P PARTS [PARTS ...], --parts PARTS [PARTS ...]\n Parts to select in the processes list of Documents\n (default: None)\n -r, --raw Wraps the content in triple quotes with no extra text\n (default: False)\n -R, --raw-no-quotes Output the content only (default: False)\n --print-percentage-non-ascii\n Print percentage of non-ascii characters (default:\n False)\n -n, --dry-run Dry run (default: False)\n\n```\n### `textprompt`\n\n```\n$ textprompt --help\n\nusage: textprompt [-h] [-V] [-c] [-e] [-m model] [-S] [-s chunk_size]\n [-P PARTS [PARTS ...]] [-r] [-R]\n [--print-percentage-non-ascii] [-n] [-C] [-w WHAT] [-M]\n [PATH ...]\n\nGet a prompt from text files\n\npositional arguments:\n PATH Paths to the text files, or stdin if not provided\n (default: None)\n\noptions:\n -h, --help show this help message and exit\n -V, --version show program's version number and exit\n -c, --copy Copy the prompt to clipboard (default: False)\n -e, --edit Edit the prompt and copy manually (default: False)\n -m model, --model model\n Model to use (default: gpt-3.5-turbo)\n -S, --no-split Do not split the prompt into multiple parts (use this\n if the model has a really large context size)\n (default: False)\n -s chunk_size, --chunk-size chunk_size\n Chunk size when splitting transcript, also used to\n determine whether to split, defaults to 1/2 of the\n context length limit of the model (default: None)\n -P PARTS [PARTS ...], --parts PARTS [PARTS ...]\n Parts to select in the processes list of Documents\n (default: None)\n -r, --raw Wraps the content in triple quotes with no extra text\n (default: False)\n -R, --raw-no-quotes Output the content only (default: False)\n --print-percentage-non-ascii\n Print percentage of non-ascii characters (default:\n False)\n -n, --dry-run Dry run (default: False)\n -C, --from-clipboard Load text from clipboard (default: False)\n -w WHAT, --what WHAT Initial knowledge you want to insert before the PDF\n content in the prompt (default: the content of a\n document)\n -M, --merge Merge contents of all pages before processing\n (default: False)\n\n```\n### `htmlprompt`\n\n```\n$ htmlprompt --help\n\nusage: htmlprompt [-h] [-V] [-c] [-e] [-m model] [-S] [-s chunk_size]\n [-P PARTS [PARTS ...]] [-r] [-R]\n [--print-percentage-non-ascii] [-n] [-C] [-w WHAT] [-M]\n [PATH ...]\n\nGet a prompt from html files\n\npositional arguments:\n PATH Paths to the html files, or stdin if not provided\n (default: None)\n\noptions:\n -h, --help show this help message and exit\n -V, --version show program's version number and exit\n -c, --copy Copy the prompt to clipboard (default: False)\n -e, --edit Edit the prompt and copy manually (default: False)\n -m model, --model model\n Model to use (default: gpt-3.5-turbo)\n -S, --no-split Do not split the prompt into multiple parts (use this\n if the model has a really large context size)\n (default: False)\n -s chunk_size, --chunk-size chunk_size\n Chunk size when splitting transcript, also used to\n determine whether to split, defaults to 1/2 of the\n context length limit of the model (default: None)\n -P PARTS [PARTS ...], --parts PARTS [PARTS ...]\n Parts to select in the processes list of Documents\n (default: None)\n -r, --raw Wraps the content in triple quotes with no extra text\n (default: False)\n -R, --raw-no-quotes Output the content only (default: False)\n --print-percentage-non-ascii\n Print percentage of non-ascii characters (default:\n False)\n -n, --dry-run Dry run (default: False)\n -C, --from-clipboard Load text from clipboard (default: False)\n -w WHAT, --what WHAT Initial knowledge you want to insert before the PDF\n content in the prompt (default: the text content of a\n html file)\n -M, --merge Merge contents of all pages before processing\n (default: False)\n\n```\n\n## Installation\n\n### pipx\n\nThis is the recommended installation method.\n\n```\n$ pipx install langchain-utils\n```\n\n### [pip](https://pypi.org/project/langchain-utils/)\n\n```\n$ pip install langchain-utils\n```\n\n## Develop\n\n```\n$ git clone https://github.com/tddschn/langchain-utils.git\n$ cd langchain-utils\n$ poetry install\n```",
"bugtrack_url": null,
"license": "MIT",
"summary": "Utilities built upon the langchain library",
"version": "0.5.6",
"project_urls": {
"Bug Tracker": "https://github.com/tddschn/langchain-utils/issues",
"Homepage": "https://github.com/tddschn/langchain-utils",
"Repository": "https://github.com/tddschn/langchain-utils"
},
"split_keywords": [
"langchain",
"utils",
"llm",
"prompts",
"cli"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "1c3618b9fd521c1ee30912b11744e1553fa11381817844d51baf3908d3e37818",
"md5": "d9215ebe5680d14813a5dec82d2f3231",
"sha256": "46fa4a57d5c58bd6fab4eb362776a22beb494cf87b3a54953bbb433528f3a9b2"
},
"downloads": -1,
"filename": "langchain_utils-0.5.6-py3-none-any.whl",
"has_sig": false,
"md5_digest": "d9215ebe5680d14813a5dec82d2f3231",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": ">=3.11,<4.0",
"size": 27734,
"upload_time": "2023-12-09T18:17:26",
"upload_time_iso_8601": "2023-12-09T18:17:26.013386Z",
"url": "https://files.pythonhosted.org/packages/1c/36/18b9fd521c1ee30912b11744e1553fa11381817844d51baf3908d3e37818/langchain_utils-0.5.6-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "ededd92fc27849f910f6cc4474b0aabe6db5e79dd32a5006373c773650ba9952",
"md5": "cf473a37dc15488444a54deaf7a88125",
"sha256": "e35d6fe95de50879a3a9a6a61c1b1e465e335a8041224274c3b3358f4d958f99"
},
"downloads": -1,
"filename": "langchain_utils-0.5.6.tar.gz",
"has_sig": false,
"md5_digest": "cf473a37dc15488444a54deaf7a88125",
"packagetype": "sdist",
"python_version": "source",
"requires_python": ">=3.11,<4.0",
"size": 19467,
"upload_time": "2023-12-09T18:17:27",
"upload_time_iso_8601": "2023-12-09T18:17:27.823476Z",
"url": "https://files.pythonhosted.org/packages/ed/ed/d92fc27849f910f6cc4474b0aabe6db5e79dd32a5006373c773650ba9952/langchain_utils-0.5.6.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2023-12-09 18:17:27",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "tddschn",
"github_project": "langchain-utils",
"travis_ci": false,
"coveralls": false,
"github_actions": false,
"lcname": "langchain-utils"
}