Name | llmstudio-core JSON |
Version |
1.0.3
JSON |
| download |
home_page | None |
Summary | LLMStudio core capabilities for routing llm calls for any vendor. No proxy server required. For that use llmstudio[proxy] |
upload_time | 2025-02-12 10:14:51 |
maintainer | None |
docs_url | None |
author | Cláudio Lemos |
requires_python | <4.0,>=3.9 |
license | MIT |
keywords |
|
VCS |
|
bugtrack_url |
|
requirements |
No requirements were recorded.
|
Travis-CI |
No Travis.
|
coveralls test coverage |
No coveralls.
|
# LLMstudio-core by [TensorOps](http://tensorops.ai "TensorOps")
Prompt Engineering at your fingertips

## 🌟 Features
- **Custom and Local LLM Support**: Use custom or local open-source LLMs through Ollama.
- **Python SDK**: Easily integrate LLMstudio into your existing workflows.
- **LangChain Integration**: LLMstudio integrates with your already existing LangChain projects.
## 🚀 Quickstart
Don't forget to check out [https://docs.llmstudio.ai](docs) page.
## Installation
Install the latest version of **LLMstudio** using `pip`. We suggest that you create and activate a new environment using `conda`
```bash
pip install llmstudio-core
```
Create a `.env` file at the same path you'll run **LLMstudio**
```bash
OPENAI_API_KEY="sk-api_key"
GOOGLE_API_KEY="sk-api_key"
```
Now you should be able to run **LLMstudio** Providers using the following code:
```
# You can set OPENAI_API_KEY environment variable, add it to .env, or pass directly as api_key
import os
from llmstudio_core.providers import LLMCore as LLM
llm = LLM("vertexai", api_key=os.environ["GOOGLE_API_KEY"])
response = llm.chat("How are you", model="gemini-1.5-pro-latest")
print(response.chat_output, response.metrics)
```
## 📖 Documentation
- [Visit our docs to learn how the SDK works](https://docs.LLMstudio.ai) (coming soon)
- Checkout our [notebook examples](https://github.com/TensorOpsAI/LLMstudio/tree/main/examples) to follow along with interactive tutorials
## 👨💻 Contributing
- Head on to our [Contribution Guide](https://github.com/TensorOpsAI/LLMstudio/tree/main/CONTRIBUTING.md) to see how you can help LLMstudio.
- Join our [Discord](https://discord.gg/GkAfPZR9wy) to talk with other LLMstudio enthusiasts.
## Training
[](https://www.tensorops.ai/llm-studio-workshop)
---
Thank you for choosing LLMstudio. Your journey to perfecting AI interactions starts here.
Raw data
{
"_id": null,
"home_page": null,
"name": "llmstudio-core",
"maintainer": null,
"docs_url": null,
"requires_python": "<4.0,>=3.9",
"maintainer_email": null,
"keywords": null,
"author": "Cl\u00e1udio Lemos",
"author_email": "claudio.lemos@tensorops.ai",
"download_url": "https://files.pythonhosted.org/packages/0e/ed/1503f288f94914850662af34eae9ca8643c48e958b08205cb5a2f1fe92c5/llmstudio_core-1.0.3.tar.gz",
"platform": null,
"description": "# LLMstudio-core by [TensorOps](http://tensorops.ai \"TensorOps\")\n\nPrompt Engineering at your fingertips\n\n\n\n## \ud83c\udf1f Features\n- **Custom and Local LLM Support**: Use custom or local open-source LLMs through Ollama.\n- **Python SDK**: Easily integrate LLMstudio into your existing workflows.\n- **LangChain Integration**: LLMstudio integrates with your already existing LangChain projects.\n\n## \ud83d\ude80 Quickstart\n\nDon't forget to check out [https://docs.llmstudio.ai](docs) page.\n\n## Installation\n\nInstall the latest version of **LLMstudio** using `pip`. We suggest that you create and activate a new environment using `conda`\n\n```bash\npip install llmstudio-core\n```\n\nCreate a `.env` file at the same path you'll run **LLMstudio**\n\n```bash\nOPENAI_API_KEY=\"sk-api_key\"\nGOOGLE_API_KEY=\"sk-api_key\"\n```\n\nNow you should be able to run **LLMstudio** Providers using the following code:\n\n```\n# You can set OPENAI_API_KEY environment variable, add it to .env, or pass directly as api_key\nimport os\nfrom llmstudio_core.providers import LLMCore as LLM\nllm = LLM(\"vertexai\", api_key=os.environ[\"GOOGLE_API_KEY\"])\nresponse = llm.chat(\"How are you\", model=\"gemini-1.5-pro-latest\")\nprint(response.chat_output, response.metrics)\n```\n## \ud83d\udcd6 Documentation\n\n- [Visit our docs to learn how the SDK works](https://docs.LLMstudio.ai) (coming soon)\n- Checkout our [notebook examples](https://github.com/TensorOpsAI/LLMstudio/tree/main/examples) to follow along with interactive tutorials\n\n## \ud83d\udc68\u200d\ud83d\udcbb Contributing\n\n- Head on to our [Contribution Guide](https://github.com/TensorOpsAI/LLMstudio/tree/main/CONTRIBUTING.md) to see how you can help LLMstudio.\n- Join our [Discord](https://discord.gg/GkAfPZR9wy) to talk with other LLMstudio enthusiasts.\n\n## Training\n\n[](https://www.tensorops.ai/llm-studio-workshop)\n\n---\n\nThank you for choosing LLMstudio. Your journey to perfecting AI interactions starts here.\n",
"bugtrack_url": null,
"license": "MIT",
"summary": "LLMStudio core capabilities for routing llm calls for any vendor. No proxy server required. For that use llmstudio[proxy]",
"version": "1.0.3",
"project_urls": null,
"split_keywords": [],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "30c1d0b409694bdc8b74acbe02fdcd6c56994c267319fcb45c6b3db25ab78021",
"md5": "bc974838db8bf105aff72e65c8edfc83",
"sha256": "8e8ee3d09badcbaff85e1dfcc0c4e077f53e9a1e4dc7eae84861b02d61816583"
},
"downloads": -1,
"filename": "llmstudio_core-1.0.3-py3-none-any.whl",
"has_sig": false,
"md5_digest": "bc974838db8bf105aff72e65c8edfc83",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": "<4.0,>=3.9",
"size": 25444,
"upload_time": "2025-02-12T10:14:50",
"upload_time_iso_8601": "2025-02-12T10:14:50.746813Z",
"url": "https://files.pythonhosted.org/packages/30/c1/d0b409694bdc8b74acbe02fdcd6c56994c267319fcb45c6b3db25ab78021/llmstudio_core-1.0.3-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "0eed1503f288f94914850662af34eae9ca8643c48e958b08205cb5a2f1fe92c5",
"md5": "a063b06c14e73cd722ad238bc5ded148",
"sha256": "49e71363f883273f9a7f9a90c38f70931d369025909e5b5aa5f95b9ab79b6149"
},
"downloads": -1,
"filename": "llmstudio_core-1.0.3.tar.gz",
"has_sig": false,
"md5_digest": "a063b06c14e73cd722ad238bc5ded148",
"packagetype": "sdist",
"python_version": "source",
"requires_python": "<4.0,>=3.9",
"size": 21439,
"upload_time": "2025-02-12T10:14:51",
"upload_time_iso_8601": "2025-02-12T10:14:51.908559Z",
"url": "https://files.pythonhosted.org/packages/0e/ed/1503f288f94914850662af34eae9ca8643c48e958b08205cb5a2f1fe92c5/llmstudio_core-1.0.3.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2025-02-12 10:14:51",
"github": false,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"lcname": "llmstudio-core"
}