langchain-runpod-llm


Namelangchain-runpod-llm JSON
Version 0.0.4 PyPI version JSON
download
home_pagehttps://github.com/tsangwailam/langchain-runpod-llm
Summary🐍 | Python library for langchain using RunPod API endpoint as LLM.
upload_time2024-02-22 21:28:58
maintainer
docs_urlNone
authorWilliam Tsang
requires_python>=3.9
licenseMIT License
keywords runpod ai langchain llm llama2 sdk api python library
VCS
bugtrack_url
requirements emoji langchain_core Requests setuptools
Travis-CI No Travis.
coveralls test coverage No coveralls.
            # Runpod LLM API Endpoint Lib for LangChain
[![PyPI version](https://badge.fury.io/py/langchain-runpod-llm.svg)](https://badge.fury.io/py/langchain-runpod-llm)

## Installation

```
# Install the latest release version
pip install runpod-llm

# or

# Install the latest development version (main branch)
pip install git+https://https://github.com/tsangwailam/langchain-runpod-llm
```

## Get Runpod API key

1. Goto www.runpod.io. Create a RunPod account.
2. From the portal, goto Settings>APIKeys
3. Create a new API key by click the "+ API Key" button.

## Usage

```python
from runpod_llm import RunpodLlama2

llm = RunpodLlama2(
        apikey="YOU_RUNPOD_API_KEY",
        llm_type="7b|13b",
        config={
            "max_tokens": 500, 
            #Maximum number of tokens to generate per output sequence.
            "n": 1,  # Number of output sequences to return for the given prompt.
            "best_of": 1,  # Number of output sequences that are generated from the prompt. From these best_of sequences, the top n sequences are returned. best_of must be greater than or equal to n. This is treated as the beam width when use_beam_search is True. By default, best_of is set to n.
            "Presence penalty": 0.2,  # Float that penalizes new tokens based on whether they appear in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.
            "Frequency penalty": 0.5,  # Float that penalizes new tokens based on their frequency in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.
            "temperature": 0.3,  # Float that controls the randomness of the sampling. Lower values make the model more deterministic, while higher values make the model more random. Zero means greedy sampling.
            "top_p": 1,  # Float that controls the cumulative probability of the top tokens to consider. Must be in (0, 1]. Set to 1 to consider all tokens.
            "top_k": -1,  # Integer that controls the number of top tokens to consider. Set to -1 to consider all tokens.
            "use_beam_search": False,  # Whether to use beam search instead of sampling.
        },
        verbose=True, # verbose output
    )

    some_prompt_template = xxxxx
    output_chain = some_prompt_template | llm
    output_chain.invoke({"input":"some input to prompt template"})
```

            

Raw data

            {
    "_id": null,
    "home_page": "https://github.com/tsangwailam/langchain-runpod-llm",
    "name": "langchain-runpod-llm",
    "maintainer": "",
    "docs_url": null,
    "requires_python": ">=3.9",
    "maintainer_email": "",
    "keywords": "runpod,ai,langchain,llm,llama2,SDK,API,python,library",
    "author": "William Tsang",
    "author_email": "William Tsang <contact@williamtsang.me>",
    "download_url": "https://files.pythonhosted.org/packages/f9/85/db36c7dda297f34ae96d7ed45f4866a21e0340aef94d9ca0b1f53557ba6e/langchain-runpod-llm-0.0.4.tar.gz",
    "platform": null,
    "description": "# Runpod LLM API Endpoint Lib for LangChain\n[![PyPI version](https://badge.fury.io/py/langchain-runpod-llm.svg)](https://badge.fury.io/py/langchain-runpod-llm)\n\n## Installation\n\n```\n# Install the latest release version\npip install runpod-llm\n\n# or\n\n# Install the latest development version (main branch)\npip install git+https://https://github.com/tsangwailam/langchain-runpod-llm\n```\n\n## Get Runpod API key\n\n1. Goto www.runpod.io. Create a RunPod account.\n2. From the portal, goto Settings>APIKeys\n3. Create a new API key by click the \"+ API Key\" button.\n\n## Usage\n\n```python\nfrom runpod_llm import RunpodLlama2\n\nllm = RunpodLlama2(\n        apikey=\"YOU_RUNPOD_API_KEY\",\n        llm_type=\"7b|13b\",\n        config={\n            \"max_tokens\": 500, \n            #Maximum number of tokens to generate per output sequence.\n            \"n\": 1,  # Number of output sequences to return for the given prompt.\n            \"best_of\": 1,  # Number of output sequences that are generated from the prompt. From these best_of sequences, the top n sequences are returned. best_of must be greater than or equal to n. This is treated as the beam width when use_beam_search is True. By default, best_of is set to n.\n            \"Presence penalty\": 0.2,  # Float that penalizes new tokens based on whether they appear in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.\n            \"Frequency penalty\": 0.5,  # Float that penalizes new tokens based on their frequency in the generated text so far. Values > 0 encourage the model to use new tokens, while values < 0 encourage the model to repeat tokens.\n            \"temperature\": 0.3,  # Float that controls the randomness of the sampling. Lower values make the model more deterministic, while higher values make the model more random. Zero means greedy sampling.\n            \"top_p\": 1,  # Float that controls the cumulative probability of the top tokens to consider. Must be in (0, 1]. Set to 1 to consider all tokens.\n            \"top_k\": -1,  # Integer that controls the number of top tokens to consider. Set to -1 to consider all tokens.\n            \"use_beam_search\": False,  # Whether to use beam search instead of sampling.\n        },\n        verbose=True, # verbose output\n    )\n\n    some_prompt_template = xxxxx\n    output_chain = some_prompt_template | llm\n    output_chain.invoke({\"input\":\"some input to prompt template\"})\n```\n",
    "bugtrack_url": null,
    "license": "MIT License",
    "summary": "\ud83d\udc0d | Python library for langchain using RunPod API endpoint as LLM.",
    "version": "0.0.4",
    "project_urls": {
        "Bug Tracker": "https://github.com/tsangwailam/langchain-runpod-llm/issues",
        "Changelog": "https://github.com/tsangwailam/langchain-runpod-llm/blob/main/CHANGELOG.md",
        "Documentation": "https://github.com/tsangwailam/langchain-runpod-llm/blob/main/README.md",
        "Homepage": "https://github.com/tsangwailam/langchain-runpod-llm",
        "Repository": "https://github.com/tsangwailam/langchain-runpod-llm"
    },
    "split_keywords": [
        "runpod",
        "ai",
        "langchain",
        "llm",
        "llama2",
        "sdk",
        "api",
        "python",
        "library"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "48d05e61a6cfd27666880b5cf760e459dd8e12a84917ef0156d4e7acb62d2315",
                "md5": "68f839f51e38fa4fc2bf3387eaef8eaa",
                "sha256": "f890ea51ab0feb88c2a1c29ad476c6ac6df729aba3d882daad251cde0b4d8750"
            },
            "downloads": -1,
            "filename": "langchain_runpod_llm-0.0.4-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "68f839f51e38fa4fc2bf3387eaef8eaa",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.9",
            "size": 10079,
            "upload_time": "2024-02-22T21:28:57",
            "upload_time_iso_8601": "2024-02-22T21:28:57.043541Z",
            "url": "https://files.pythonhosted.org/packages/48/d0/5e61a6cfd27666880b5cf760e459dd8e12a84917ef0156d4e7acb62d2315/langchain_runpod_llm-0.0.4-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "f985db36c7dda297f34ae96d7ed45f4866a21e0340aef94d9ca0b1f53557ba6e",
                "md5": "5c490fd6c323410757e3fc25ed0ba17b",
                "sha256": "695f8291a46d3c23640c6a3f96b52f7a8537a196e9309b281dca2595f59f0b33"
            },
            "downloads": -1,
            "filename": "langchain-runpod-llm-0.0.4.tar.gz",
            "has_sig": false,
            "md5_digest": "5c490fd6c323410757e3fc25ed0ba17b",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.9",
            "size": 8883,
            "upload_time": "2024-02-22T21:28:58",
            "upload_time_iso_8601": "2024-02-22T21:28:58.087106Z",
            "url": "https://files.pythonhosted.org/packages/f9/85/db36c7dda297f34ae96d7ed45f4866a21e0340aef94d9ca0b1f53557ba6e/langchain-runpod-llm-0.0.4.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-02-22 21:28:58",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "tsangwailam",
    "github_project": "langchain-runpod-llm",
    "travis_ci": false,
    "coveralls": false,
    "github_actions": true,
    "requirements": [
        {
            "name": "emoji",
            "specs": [
                [
                    ">=",
                    "2.9.0"
                ]
            ]
        },
        {
            "name": "langchain_core",
            "specs": [
                [
                    ">=",
                    "0.1.18"
                ]
            ]
        },
        {
            "name": "Requests",
            "specs": [
                [
                    ">=",
                    "2.31.0"
                ]
            ]
        },
        {
            "name": "setuptools",
            "specs": [
                [
                    "==",
                    "68.2.2"
                ]
            ]
        }
    ],
    "lcname": "langchain-runpod-llm"
}
        
Elapsed time: 0.19276s