MyLLM


NameMyLLM JSON
Version 4.13.12 PyPI version JSON
download
home_pageNone
SummaryA python package to interact with llm model supported by g4f and langchain.
upload_time2024-11-23 15:02:59
maintainerNone
docs_urlNone
authormraniki
requires_python<4.0,>=3.10
licenseMIT
keywords chatgpt llm ai llama ai g4f freegpt
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            <table style="border: 1px solid transparent">
  <tr>
    <td>
<a href="http://talky.readthedocs.io"><img src="https://img.shields.io/badge/Wiki-%23000000.svg?style=for-the-badge&logo=wikipedia&logoColor=white"></a>
<a href="https://github.com/mraniki/tt/"><img src="https://img.shields.io/badge/github-%23000000.svg?style=for-the-badge&logo=github&logoColor=white"></a><br>
<a href="https://hub.docker.com/r/mraniki/tt"><img src="https://img.shields.io/docker/pulls/mraniki/tt?style=for-the-badge"></a><br>
       </td>
    <td align="center"><img width="200" alt="Logo" src="/docs/_static/logo-full.png"></td>
  </tr>
  <tr>
    <td>
      <a href="https://pypi.org/project/myllm/"><img src="https://img.shields.io/pypi/v/myllm?style=for-the-badge&logo=PyPI&logoColor=white"></a><br>
      <a href="https://pypi.org/project/myllm/"><img src="https://img.shields.io/pypi/dm/myllm?style=for-the-badge&logo=PyPI&logoColor=white"></a><br>
      <a href="https://github.com/mraniki/myllm/"><img src="https://img.shields.io/github/actions/workflow/status/mraniki/myllm/%F0%9F%91%B7Flow.yml?style=for-the-badge&logo=GitHub&logoColor=white"></a><br>
      <a href="https://talky.readthedocs.io/projects/myllm/"><img src="https://readthedocs.org/projects/myllm/badge/?version=latest&style=for-the-badge"></a><br>
      <a href="https://codebeat.co/projects/github-com-mraniki-myllm-main"><img src="https://codebeat.co/badges/0567b9d9-3cbb-4263-80ec-8ac8043332ea"/></a> <br>
      <a href="https://app.codacy.com/gh/mraniki/MyLLM/dashboard"><img src="https://app.codacy.com/project/badge/Grade/75f2650d58044355957fe667aeed50a1"/></a> <br>
      <a href="https://codecov.io/gh/mraniki/myllm"><img src="https://codecov.io/gh/mraniki/myllm/branch/main/graph/badge.svg?token=WAHUEMAJN6"/></a><br>
    </td>
    <td align="left"> 
Interact with LLM in simple way,<br>via  <a href="https://github.com/xtekky/gpt4free">G4F</a>, <a href="https://github.com/openai/openai-python">OPENAI</a>/<a href="https://localai.io/basics/getting_started/">LOCALAI</a><br>
  </td>   
  </tr>
</table>


<h5>How to use it</h5>
<pre>
<code>
    talky = MyLLM()

    chat = await talky.chat("My name is Jack")
    print(chat)
    # 🐻 Hello Jack, this is Bing. I'm happy to help you with general tasks. 😊

    chat = await talky.chat("tell me who is president of the united states?")
    print(chat)
    # The current president of the United States is **Joe Biden**[^1^][1] [^2^][2].
    # He took office on **January 20, 2021**[^1^][1] [^2^][2].
    # He is the **46th** president of the United States[^2^][2] [^3^][5].

    chat = await talky.chat("what is my name")
    print(chat)
    # 🐻 You told me your name is Jack. Is that correct?

    await talky.export_chat_history()
    # talky.clear_chat_history()
</code>
</pre>

<h5>Documentation</h5>
<a href="https://talky.readthedocs.io/projects/myllm/en/latest/"><img src="https://img.shields.io/badge/Documentation-000000?style=for-the-badge&logo=readthedocs&logoColor=white"></a><br>


            

Raw data

            {
    "_id": null,
    "home_page": null,
    "name": "MyLLM",
    "maintainer": null,
    "docs_url": null,
    "requires_python": "<4.0,>=3.10",
    "maintainer_email": null,
    "keywords": "chatgpt, llm, ai, llama, ai, g4f, freegpt",
    "author": "mraniki",
    "author_email": "8766259+mraniki@users.noreply.github.com",
    "download_url": "https://files.pythonhosted.org/packages/b8/b6/d150214ce42054fb65147a49480a8779bc0f72dd3cb3ab1fe85e224e0d78/myllm-4.13.12.tar.gz",
    "platform": null,
    "description": "<table style=\"border: 1px solid transparent\">\n  <tr>\n    <td>\n<a href=\"http://talky.readthedocs.io\"><img src=\"https://img.shields.io/badge/Wiki-%23000000.svg?style=for-the-badge&logo=wikipedia&logoColor=white\"></a>\n<a href=\"https://github.com/mraniki/tt/\"><img src=\"https://img.shields.io/badge/github-%23000000.svg?style=for-the-badge&logo=github&logoColor=white\"></a><br>\n<a href=\"https://hub.docker.com/r/mraniki/tt\"><img src=\"https://img.shields.io/docker/pulls/mraniki/tt?style=for-the-badge\"></a><br>\n       </td>\n    <td align=\"center\"><img width=\"200\" alt=\"Logo\" src=\"/docs/_static/logo-full.png\"></td>\n  </tr>\n  <tr>\n    <td>\n      <a href=\"https://pypi.org/project/myllm/\"><img src=\"https://img.shields.io/pypi/v/myllm?style=for-the-badge&logo=PyPI&logoColor=white\"></a><br>\n      <a href=\"https://pypi.org/project/myllm/\"><img src=\"https://img.shields.io/pypi/dm/myllm?style=for-the-badge&logo=PyPI&logoColor=white\"></a><br>\n      <a href=\"https://github.com/mraniki/myllm/\"><img src=\"https://img.shields.io/github/actions/workflow/status/mraniki/myllm/%F0%9F%91%B7Flow.yml?style=for-the-badge&logo=GitHub&logoColor=white\"></a><br>\n      <a href=\"https://talky.readthedocs.io/projects/myllm/\"><img src=\"https://readthedocs.org/projects/myllm/badge/?version=latest&style=for-the-badge\"></a><br>\n      <a href=\"https://codebeat.co/projects/github-com-mraniki-myllm-main\"><img src=\"https://codebeat.co/badges/0567b9d9-3cbb-4263-80ec-8ac8043332ea\"/></a> <br>\n      <a href=\"https://app.codacy.com/gh/mraniki/MyLLM/dashboard\"><img src=\"https://app.codacy.com/project/badge/Grade/75f2650d58044355957fe667aeed50a1\"/></a> <br>\n      <a href=\"https://codecov.io/gh/mraniki/myllm\"><img src=\"https://codecov.io/gh/mraniki/myllm/branch/main/graph/badge.svg?token=WAHUEMAJN6\"/></a><br>\n    </td>\n    <td align=\"left\"> \nInteract with LLM in simple way,<br>via  <a href=\"https://github.com/xtekky/gpt4free\">G4F</a>, <a href=\"https://github.com/openai/openai-python\">OPENAI</a>/<a href=\"https://localai.io/basics/getting_started/\">LOCALAI</a><br>\n  </td>   \n  </tr>\n</table>\n\n\n<h5>How to use it</h5>\n<pre>\n<code>\n    talky = MyLLM()\n\n    chat = await talky.chat(\"My name is Jack\")\n    print(chat)\n    # \ud83d\udc3b Hello Jack, this is Bing. I'm happy to help you with general tasks. \ud83d\ude0a\n\n    chat = await talky.chat(\"tell me who is president of the united states?\")\n    print(chat)\n    # The current president of the United States is **Joe Biden**[^1^][1] [^2^][2].\n    # He took office on **January 20, 2021**[^1^][1] [^2^][2].\n    # He is the **46th** president of the United States[^2^][2] [^3^][5].\n\n    chat = await talky.chat(\"what is my name\")\n    print(chat)\n    # \ud83d\udc3b You told me your name is Jack. Is that correct?\n\n    await talky.export_chat_history()\n    # talky.clear_chat_history()\n</code>\n</pre>\n\n<h5>Documentation</h5>\n<a href=\"https://talky.readthedocs.io/projects/myllm/en/latest/\"><img src=\"https://img.shields.io/badge/Documentation-000000?style=for-the-badge&logo=readthedocs&logoColor=white\"></a><br>\n\n",
    "bugtrack_url": null,
    "license": "MIT",
    "summary": "A python package to interact with llm model supported by g4f and langchain.",
    "version": "4.13.12",
    "project_urls": {
        "Changelog": "https://github.com/mraniki/MyLLM/blob/dev/CHANGELOG.rst",
        "Issues": "https://github.com/mraniki/MyLLM/issues",
        "Support": "https://github.com/mraniki/MyLLM/discussions"
    },
    "split_keywords": [
        "chatgpt",
        " llm",
        " ai",
        " llama",
        " ai",
        " g4f",
        " freegpt"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "12ff38778c8e87f89d98362b4034c4db5b405116ad1bc97af13d23af2159c196",
                "md5": "22decd95b5a4d97f6cad2b55edf865fe",
                "sha256": "74fc15f6b561b1229044da30757be67de3cd270522e324332ece7eb2a6cc4cb5"
            },
            "downloads": -1,
            "filename": "myllm-4.13.12-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "22decd95b5a4d97f6cad2b55edf865fe",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": "<4.0,>=3.10",
            "size": 13676,
            "upload_time": "2024-11-23T15:02:58",
            "upload_time_iso_8601": "2024-11-23T15:02:58.377429Z",
            "url": "https://files.pythonhosted.org/packages/12/ff/38778c8e87f89d98362b4034c4db5b405116ad1bc97af13d23af2159c196/myllm-4.13.12-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "b8b6d150214ce42054fb65147a49480a8779bc0f72dd3cb3ab1fe85e224e0d78",
                "md5": "f9d65453e073029ef378012eba14c96d",
                "sha256": "5a485abb9000d87ae10d5e2a4efee0b72f637554c158cb7a46ada91061935d42"
            },
            "downloads": -1,
            "filename": "myllm-4.13.12.tar.gz",
            "has_sig": false,
            "md5_digest": "f9d65453e073029ef378012eba14c96d",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": "<4.0,>=3.10",
            "size": 12268,
            "upload_time": "2024-11-23T15:02:59",
            "upload_time_iso_8601": "2024-11-23T15:02:59.436312Z",
            "url": "https://files.pythonhosted.org/packages/b8/b6/d150214ce42054fb65147a49480a8779bc0f72dd3cb3ab1fe85e224e0d78/myllm-4.13.12.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-11-23 15:02:59",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "mraniki",
    "github_project": "MyLLM",
    "travis_ci": false,
    "coveralls": false,
    "github_actions": true,
    "lcname": "myllm"
}
        
Elapsed time: 0.43385s