Name | MyLLM JSON |
Version |
4.13.53
JSON |
| download |
home_page | None |
Summary | A python package to interact with llm model supported by g4f and langchain. |
upload_time | 2025-03-21 20:32:50 |
maintainer | None |
docs_url | None |
author | mraniki |
requires_python | <4.0,>=3.10 |
license | MIT |
keywords |
chatgpt
llm
ai
llama
ai
g4f
freegpt
|
VCS |
 |
bugtrack_url |
|
requirements |
No requirements were recorded.
|
Travis-CI |
No Travis.
|
coveralls test coverage |
No coveralls.
|
<table style="border: 1px solid transparent">
<tr>
<td>
<a href="http://talky.readthedocs.io"><img src="https://img.shields.io/badge/Wiki-%23000000.svg?style=for-the-badge&logo=wikipedia&logoColor=white"></a>
<a href="https://github.com/mraniki/tt/"><img src="https://img.shields.io/badge/github-%23000000.svg?style=for-the-badge&logo=github&logoColor=white"></a><br>
<a href="https://hub.docker.com/r/mraniki/tt"><img src="https://img.shields.io/docker/pulls/mraniki/tt?style=for-the-badge"></a><br>
</td>
<td align="center"><img width="200" alt="Logo" src="/docs/_static/logo-full.png"></td>
</tr>
<tr>
<td>
<a href="https://pypi.org/project/myllm/"><img src="https://img.shields.io/pypi/v/myllm?style=for-the-badge&logo=PyPI&logoColor=white"></a><br>
<a href="https://pypi.org/project/myllm/"><img src="https://img.shields.io/pypi/dm/myllm?style=for-the-badge&logo=PyPI&logoColor=white"></a><br>
<a href="https://github.com/mraniki/myllm/"><img src="https://img.shields.io/github/actions/workflow/status/mraniki/myllm/%F0%9F%91%B7Flow.yml?style=for-the-badge&logo=GitHub&logoColor=white"></a><br>
<a href="https://talky.readthedocs.io/projects/myllm/"><img src="https://readthedocs.org/projects/myllm/badge/?version=latest&style=for-the-badge"></a><br>
<a href="https://codebeat.co/projects/github-com-mraniki-myllm-main"><img src="https://codebeat.co/badges/0567b9d9-3cbb-4263-80ec-8ac8043332ea"/></a> <br>
<a href="https://app.codacy.com/gh/mraniki/MyLLM/dashboard"><img src="https://app.codacy.com/project/badge/Grade/75f2650d58044355957fe667aeed50a1"/></a> <br>
<a href="https://codecov.io/gh/mraniki/myllm"><img src="https://codecov.io/gh/mraniki/myllm/branch/main/graph/badge.svg?token=WAHUEMAJN6"/></a><br>
</td>
<td align="left">
Interact with LLM in simple way,<br>via <a href="https://github.com/xtekky/gpt4free">G4F</a>, <a href="https://github.com/openai/openai-python">OPENAI</a>/<a href="https://localai.io/basics/getting_started/">LOCALAI</a><br>
</td>
</tr>
</table>
<h5>How to use it</h5>
<pre>
<code>
talky = MyLLM()
chat = await talky.chat("My name is Jack")
print(chat)
# 🐻 Hello Jack, this is Bing. I'm happy to help you with general tasks. 😊
chat = await talky.chat("tell me who is president of the united states?")
print(chat)
# The current president of the United States is **Joe Biden**[^1^][1] [^2^][2].
# He took office on **January 20, 2021**[^1^][1] [^2^][2].
# He is the **46th** president of the United States[^2^][2] [^3^][5].
chat = await talky.chat("what is my name")
print(chat)
# 🐻 You told me your name is Jack. Is that correct?
await talky.export_chat_history()
# talky.clear_chat_history()
</code>
</pre>
<h5>Documentation</h5>
<a href="https://talky.readthedocs.io/projects/myllm/en/latest/"><img src="https://img.shields.io/badge/Documentation-000000?style=for-the-badge&logo=readthedocs&logoColor=white"></a><br>
Raw data
{
"_id": null,
"home_page": null,
"name": "MyLLM",
"maintainer": null,
"docs_url": null,
"requires_python": "<4.0,>=3.10",
"maintainer_email": null,
"keywords": "chatgpt, llm, ai, llama, ai, g4f, freegpt",
"author": "mraniki",
"author_email": "8766259+mraniki@users.noreply.github.com",
"download_url": "https://files.pythonhosted.org/packages/cd/2c/994610e5be9dfe312c2969a0487e998bfb94acbce6cad151bffa603421b4/myllm-4.13.53.tar.gz",
"platform": null,
"description": "<table style=\"border: 1px solid transparent\">\n <tr>\n <td>\n<a href=\"http://talky.readthedocs.io\"><img src=\"https://img.shields.io/badge/Wiki-%23000000.svg?style=for-the-badge&logo=wikipedia&logoColor=white\"></a>\n<a href=\"https://github.com/mraniki/tt/\"><img src=\"https://img.shields.io/badge/github-%23000000.svg?style=for-the-badge&logo=github&logoColor=white\"></a><br>\n<a href=\"https://hub.docker.com/r/mraniki/tt\"><img src=\"https://img.shields.io/docker/pulls/mraniki/tt?style=for-the-badge\"></a><br>\n </td>\n <td align=\"center\"><img width=\"200\" alt=\"Logo\" src=\"/docs/_static/logo-full.png\"></td>\n </tr>\n <tr>\n <td>\n <a href=\"https://pypi.org/project/myllm/\"><img src=\"https://img.shields.io/pypi/v/myllm?style=for-the-badge&logo=PyPI&logoColor=white\"></a><br>\n <a href=\"https://pypi.org/project/myllm/\"><img src=\"https://img.shields.io/pypi/dm/myllm?style=for-the-badge&logo=PyPI&logoColor=white\"></a><br>\n <a href=\"https://github.com/mraniki/myllm/\"><img src=\"https://img.shields.io/github/actions/workflow/status/mraniki/myllm/%F0%9F%91%B7Flow.yml?style=for-the-badge&logo=GitHub&logoColor=white\"></a><br>\n <a href=\"https://talky.readthedocs.io/projects/myllm/\"><img src=\"https://readthedocs.org/projects/myllm/badge/?version=latest&style=for-the-badge\"></a><br>\n <a href=\"https://codebeat.co/projects/github-com-mraniki-myllm-main\"><img src=\"https://codebeat.co/badges/0567b9d9-3cbb-4263-80ec-8ac8043332ea\"/></a> <br>\n <a href=\"https://app.codacy.com/gh/mraniki/MyLLM/dashboard\"><img src=\"https://app.codacy.com/project/badge/Grade/75f2650d58044355957fe667aeed50a1\"/></a> <br>\n <a href=\"https://codecov.io/gh/mraniki/myllm\"><img src=\"https://codecov.io/gh/mraniki/myllm/branch/main/graph/badge.svg?token=WAHUEMAJN6\"/></a><br>\n </td>\n <td align=\"left\"> \nInteract with LLM in simple way,<br>via <a href=\"https://github.com/xtekky/gpt4free\">G4F</a>, <a href=\"https://github.com/openai/openai-python\">OPENAI</a>/<a href=\"https://localai.io/basics/getting_started/\">LOCALAI</a><br>\n </td> \n </tr>\n</table>\n\n\n<h5>How to use it</h5>\n<pre>\n<code>\n talky = MyLLM()\n\n chat = await talky.chat(\"My name is Jack\")\n print(chat)\n # \ud83d\udc3b Hello Jack, this is Bing. I'm happy to help you with general tasks. \ud83d\ude0a\n\n chat = await talky.chat(\"tell me who is president of the united states?\")\n print(chat)\n # The current president of the United States is **Joe Biden**[^1^][1] [^2^][2].\n # He took office on **January 20, 2021**[^1^][1] [^2^][2].\n # He is the **46th** president of the United States[^2^][2] [^3^][5].\n\n chat = await talky.chat(\"what is my name\")\n print(chat)\n # \ud83d\udc3b You told me your name is Jack. Is that correct?\n\n await talky.export_chat_history()\n # talky.clear_chat_history()\n</code>\n</pre>\n\n<h5>Documentation</h5>\n<a href=\"https://talky.readthedocs.io/projects/myllm/en/latest/\"><img src=\"https://img.shields.io/badge/Documentation-000000?style=for-the-badge&logo=readthedocs&logoColor=white\"></a><br>\n\n",
"bugtrack_url": null,
"license": "MIT",
"summary": "A python package to interact with llm model supported by g4f and langchain.",
"version": "4.13.53",
"project_urls": {
"Changelog": "https://github.com/mraniki/MyLLM/blob/dev/CHANGELOG.rst",
"Issues": "https://github.com/mraniki/MyLLM/issues",
"Support": "https://github.com/mraniki/MyLLM/discussions"
},
"split_keywords": [
"chatgpt",
" llm",
" ai",
" llama",
" ai",
" g4f",
" freegpt"
],
"urls": [
{
"comment_text": null,
"digests": {
"blake2b_256": "447df094b807ceaf6d115243c3884ce50c9e63d8ed2baf402821b24e5c804055",
"md5": "f9fc887c84ca9b9b13c383283e9e8a41",
"sha256": "95bd047d52822f21ee5fa636b30bf181728b4cea40a87fd275dcbf39a3609c07"
},
"downloads": -1,
"filename": "myllm-4.13.53-py3-none-any.whl",
"has_sig": false,
"md5_digest": "f9fc887c84ca9b9b13c383283e9e8a41",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": "<4.0,>=3.10",
"size": 13678,
"upload_time": "2025-03-21T20:32:49",
"upload_time_iso_8601": "2025-03-21T20:32:49.443384Z",
"url": "https://files.pythonhosted.org/packages/44/7d/f094b807ceaf6d115243c3884ce50c9e63d8ed2baf402821b24e5c804055/myllm-4.13.53-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": null,
"digests": {
"blake2b_256": "cd2c994610e5be9dfe312c2969a0487e998bfb94acbce6cad151bffa603421b4",
"md5": "895336dc2b6a83d9a528c9c3f50c7ba9",
"sha256": "482926ee1fbd8df8a670b96e1629f8492407dff4d43ebc0c494f4d193de15235"
},
"downloads": -1,
"filename": "myllm-4.13.53.tar.gz",
"has_sig": false,
"md5_digest": "895336dc2b6a83d9a528c9c3f50c7ba9",
"packagetype": "sdist",
"python_version": "source",
"requires_python": "<4.0,>=3.10",
"size": 11983,
"upload_time": "2025-03-21T20:32:50",
"upload_time_iso_8601": "2025-03-21T20:32:50.760000Z",
"url": "https://files.pythonhosted.org/packages/cd/2c/994610e5be9dfe312c2969a0487e998bfb94acbce6cad151bffa603421b4/myllm-4.13.53.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2025-03-21 20:32:50",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "mraniki",
"github_project": "MyLLM",
"travis_ci": false,
"coveralls": false,
"github_actions": true,
"lcname": "myllm"
}