Name | MyLLM JSON |
Version |
4.13.12
JSON |
| download |
home_page | None |
Summary | A python package to interact with llm model supported by g4f and langchain. |
upload_time | 2024-11-23 15:02:59 |
maintainer | None |
docs_url | None |
author | mraniki |
requires_python | <4.0,>=3.10 |
license | MIT |
keywords |
chatgpt
llm
ai
llama
ai
g4f
freegpt
|
VCS |
|
bugtrack_url |
|
requirements |
No requirements were recorded.
|
Travis-CI |
No Travis.
|
coveralls test coverage |
No coveralls.
|
<table style="border: 1px solid transparent">
<tr>
<td>
<a href="http://talky.readthedocs.io"><img src="https://img.shields.io/badge/Wiki-%23000000.svg?style=for-the-badge&logo=wikipedia&logoColor=white"></a>
<a href="https://github.com/mraniki/tt/"><img src="https://img.shields.io/badge/github-%23000000.svg?style=for-the-badge&logo=github&logoColor=white"></a><br>
<a href="https://hub.docker.com/r/mraniki/tt"><img src="https://img.shields.io/docker/pulls/mraniki/tt?style=for-the-badge"></a><br>
</td>
<td align="center"><img width="200" alt="Logo" src="/docs/_static/logo-full.png"></td>
</tr>
<tr>
<td>
<a href="https://pypi.org/project/myllm/"><img src="https://img.shields.io/pypi/v/myllm?style=for-the-badge&logo=PyPI&logoColor=white"></a><br>
<a href="https://pypi.org/project/myllm/"><img src="https://img.shields.io/pypi/dm/myllm?style=for-the-badge&logo=PyPI&logoColor=white"></a><br>
<a href="https://github.com/mraniki/myllm/"><img src="https://img.shields.io/github/actions/workflow/status/mraniki/myllm/%F0%9F%91%B7Flow.yml?style=for-the-badge&logo=GitHub&logoColor=white"></a><br>
<a href="https://talky.readthedocs.io/projects/myllm/"><img src="https://readthedocs.org/projects/myllm/badge/?version=latest&style=for-the-badge"></a><br>
<a href="https://codebeat.co/projects/github-com-mraniki-myllm-main"><img src="https://codebeat.co/badges/0567b9d9-3cbb-4263-80ec-8ac8043332ea"/></a> <br>
<a href="https://app.codacy.com/gh/mraniki/MyLLM/dashboard"><img src="https://app.codacy.com/project/badge/Grade/75f2650d58044355957fe667aeed50a1"/></a> <br>
<a href="https://codecov.io/gh/mraniki/myllm"><img src="https://codecov.io/gh/mraniki/myllm/branch/main/graph/badge.svg?token=WAHUEMAJN6"/></a><br>
</td>
<td align="left">
Interact with LLM in simple way,<br>via <a href="https://github.com/xtekky/gpt4free">G4F</a>, <a href="https://github.com/openai/openai-python">OPENAI</a>/<a href="https://localai.io/basics/getting_started/">LOCALAI</a><br>
</td>
</tr>
</table>
<h5>How to use it</h5>
<pre>
<code>
talky = MyLLM()
chat = await talky.chat("My name is Jack")
print(chat)
# 🐻 Hello Jack, this is Bing. I'm happy to help you with general tasks. 😊
chat = await talky.chat("tell me who is president of the united states?")
print(chat)
# The current president of the United States is **Joe Biden**[^1^][1] [^2^][2].
# He took office on **January 20, 2021**[^1^][1] [^2^][2].
# He is the **46th** president of the United States[^2^][2] [^3^][5].
chat = await talky.chat("what is my name")
print(chat)
# 🐻 You told me your name is Jack. Is that correct?
await talky.export_chat_history()
# talky.clear_chat_history()
</code>
</pre>
<h5>Documentation</h5>
<a href="https://talky.readthedocs.io/projects/myllm/en/latest/"><img src="https://img.shields.io/badge/Documentation-000000?style=for-the-badge&logo=readthedocs&logoColor=white"></a><br>
Raw data
{
"_id": null,
"home_page": null,
"name": "MyLLM",
"maintainer": null,
"docs_url": null,
"requires_python": "<4.0,>=3.10",
"maintainer_email": null,
"keywords": "chatgpt, llm, ai, llama, ai, g4f, freegpt",
"author": "mraniki",
"author_email": "8766259+mraniki@users.noreply.github.com",
"download_url": "https://files.pythonhosted.org/packages/b8/b6/d150214ce42054fb65147a49480a8779bc0f72dd3cb3ab1fe85e224e0d78/myllm-4.13.12.tar.gz",
"platform": null,
"description": "<table style=\"border: 1px solid transparent\">\n <tr>\n <td>\n<a href=\"http://talky.readthedocs.io\"><img src=\"https://img.shields.io/badge/Wiki-%23000000.svg?style=for-the-badge&logo=wikipedia&logoColor=white\"></a>\n<a href=\"https://github.com/mraniki/tt/\"><img src=\"https://img.shields.io/badge/github-%23000000.svg?style=for-the-badge&logo=github&logoColor=white\"></a><br>\n<a href=\"https://hub.docker.com/r/mraniki/tt\"><img src=\"https://img.shields.io/docker/pulls/mraniki/tt?style=for-the-badge\"></a><br>\n </td>\n <td align=\"center\"><img width=\"200\" alt=\"Logo\" src=\"/docs/_static/logo-full.png\"></td>\n </tr>\n <tr>\n <td>\n <a href=\"https://pypi.org/project/myllm/\"><img src=\"https://img.shields.io/pypi/v/myllm?style=for-the-badge&logo=PyPI&logoColor=white\"></a><br>\n <a href=\"https://pypi.org/project/myllm/\"><img src=\"https://img.shields.io/pypi/dm/myllm?style=for-the-badge&logo=PyPI&logoColor=white\"></a><br>\n <a href=\"https://github.com/mraniki/myllm/\"><img src=\"https://img.shields.io/github/actions/workflow/status/mraniki/myllm/%F0%9F%91%B7Flow.yml?style=for-the-badge&logo=GitHub&logoColor=white\"></a><br>\n <a href=\"https://talky.readthedocs.io/projects/myllm/\"><img src=\"https://readthedocs.org/projects/myllm/badge/?version=latest&style=for-the-badge\"></a><br>\n <a href=\"https://codebeat.co/projects/github-com-mraniki-myllm-main\"><img src=\"https://codebeat.co/badges/0567b9d9-3cbb-4263-80ec-8ac8043332ea\"/></a> <br>\n <a href=\"https://app.codacy.com/gh/mraniki/MyLLM/dashboard\"><img src=\"https://app.codacy.com/project/badge/Grade/75f2650d58044355957fe667aeed50a1\"/></a> <br>\n <a href=\"https://codecov.io/gh/mraniki/myllm\"><img src=\"https://codecov.io/gh/mraniki/myllm/branch/main/graph/badge.svg?token=WAHUEMAJN6\"/></a><br>\n </td>\n <td align=\"left\"> \nInteract with LLM in simple way,<br>via <a href=\"https://github.com/xtekky/gpt4free\">G4F</a>, <a href=\"https://github.com/openai/openai-python\">OPENAI</a>/<a href=\"https://localai.io/basics/getting_started/\">LOCALAI</a><br>\n </td> \n </tr>\n</table>\n\n\n<h5>How to use it</h5>\n<pre>\n<code>\n talky = MyLLM()\n\n chat = await talky.chat(\"My name is Jack\")\n print(chat)\n # \ud83d\udc3b Hello Jack, this is Bing. I'm happy to help you with general tasks. \ud83d\ude0a\n\n chat = await talky.chat(\"tell me who is president of the united states?\")\n print(chat)\n # The current president of the United States is **Joe Biden**[^1^][1] [^2^][2].\n # He took office on **January 20, 2021**[^1^][1] [^2^][2].\n # He is the **46th** president of the United States[^2^][2] [^3^][5].\n\n chat = await talky.chat(\"what is my name\")\n print(chat)\n # \ud83d\udc3b You told me your name is Jack. Is that correct?\n\n await talky.export_chat_history()\n # talky.clear_chat_history()\n</code>\n</pre>\n\n<h5>Documentation</h5>\n<a href=\"https://talky.readthedocs.io/projects/myllm/en/latest/\"><img src=\"https://img.shields.io/badge/Documentation-000000?style=for-the-badge&logo=readthedocs&logoColor=white\"></a><br>\n\n",
"bugtrack_url": null,
"license": "MIT",
"summary": "A python package to interact with llm model supported by g4f and langchain.",
"version": "4.13.12",
"project_urls": {
"Changelog": "https://github.com/mraniki/MyLLM/blob/dev/CHANGELOG.rst",
"Issues": "https://github.com/mraniki/MyLLM/issues",
"Support": "https://github.com/mraniki/MyLLM/discussions"
},
"split_keywords": [
"chatgpt",
" llm",
" ai",
" llama",
" ai",
" g4f",
" freegpt"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "12ff38778c8e87f89d98362b4034c4db5b405116ad1bc97af13d23af2159c196",
"md5": "22decd95b5a4d97f6cad2b55edf865fe",
"sha256": "74fc15f6b561b1229044da30757be67de3cd270522e324332ece7eb2a6cc4cb5"
},
"downloads": -1,
"filename": "myllm-4.13.12-py3-none-any.whl",
"has_sig": false,
"md5_digest": "22decd95b5a4d97f6cad2b55edf865fe",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": "<4.0,>=3.10",
"size": 13676,
"upload_time": "2024-11-23T15:02:58",
"upload_time_iso_8601": "2024-11-23T15:02:58.377429Z",
"url": "https://files.pythonhosted.org/packages/12/ff/38778c8e87f89d98362b4034c4db5b405116ad1bc97af13d23af2159c196/myllm-4.13.12-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "b8b6d150214ce42054fb65147a49480a8779bc0f72dd3cb3ab1fe85e224e0d78",
"md5": "f9d65453e073029ef378012eba14c96d",
"sha256": "5a485abb9000d87ae10d5e2a4efee0b72f637554c158cb7a46ada91061935d42"
},
"downloads": -1,
"filename": "myllm-4.13.12.tar.gz",
"has_sig": false,
"md5_digest": "f9d65453e073029ef378012eba14c96d",
"packagetype": "sdist",
"python_version": "source",
"requires_python": "<4.0,>=3.10",
"size": 12268,
"upload_time": "2024-11-23T15:02:59",
"upload_time_iso_8601": "2024-11-23T15:02:59.436312Z",
"url": "https://files.pythonhosted.org/packages/b8/b6/d150214ce42054fb65147a49480a8779bc0f72dd3cb3ab1fe85e224e0d78/myllm-4.13.12.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2024-11-23 15:02:59",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "mraniki",
"github_project": "MyLLM",
"travis_ci": false,
"coveralls": false,
"github_actions": true,
"lcname": "myllm"
}