![Banner](https://raw.githubusercontent.com/JonahWhaler/llm-agent-toolkit/main/images/repo-banner.jpeg)
# LLM Agent Toolkit: Modular Components for AI Workflows
LLM Agent Toolkit provides minimal, modular interfaces for core components in LLM-based applications. Simplify workflows with stateless interaction, embedding encoders, memory management, tool integration, and data loaders, designed for compatibility and scalability. It prioritizes simplicity and modularity by proposing minimal wrappers designed to work across common tools, discouraging direct access to underlying technologies. Specific implementations and examples will be documented separately in a Cookbook (planned).
PyPI: ![PyPI Downloads](https://static.pepy.tech/badge/llm-agent-toolkit)
## Attention!!!
Using this toolkit simplifies integration by providing unified and modular interfaces across platforms. Many configurations are intentionally kept at their default settings to prioritize ease of use. However, most of these components are extensible through abstract classes, allowing developers to define their own desired configurations for greater flexibility. While this approach enhances consistency and reduces complexity, advanced customization may require extending the provided abstractions.
For developers requiring full-range customization or access to the latest features, it is recommended to consider using native libraries like `ollama` and `openai` directly.
# Table of Contents
- [LLM Agent Toolkit: Modular Components for AI Workflows](#llm-agent-toolkit-modular-components-for-ai-workflows)
- [Attention!!!](#attention)
- [Table of Contents](#table-of-contents)
- [Installation](#installation)
- [Fundamental Components](#fundamental-components)
- [Core:](#core)
- [Encoder:](#encoder)
- [Memory:](#memory)
- [Tool:](#tool)
- [Loader:](#loader)
- [Chunkers:](#chunkers)
- [Planned Feature](#planned-feature)
- [License](#license)
# Installation
`
pip install llm-agent-toolkit
`
# Fundamental Components
## Core:
A stateless chat completion interface to interact with the LLM.
**Purpose**: Serves as the central execution layer that abstracts interaction with the underlying LLM model.
**Features**:
* Supports Text-to-Text and Image-to-Text.
* Enables iterative executions for multi-step workflows.
* Facilitates tool invocation as part of the workflow.
* Support models from OpenAI and Ollama
## Encoder:
A standardized wrapper for embedding models.
**Purpose**: Provides a minimal API to transform text into embeddings, usable with any common embedding model.
**Features**:
* Abstracts away model-specific details (e.g., dimensionality, framework differences).
* Allows flexible integration with downstream components like Memory or retrieval mechanisms.
* Support OpenAI, Ollama and Transformers.
* Support asynchronous operation.
## Memory:
Offers essential context retention capabilities.
**Purpose**: Allows efficient context management without hardcoding database or memory solutions.
**Types**:
1. *Short-term Memory*:
* Maintains recent interactions for session-based context.
2. *Vector Memory*:
* Combines embedding and storage for retrieval-augmented workflows.
* Includes optional metadata management for filtering results.
* Support Faiss and Chroma
3. *Async Vector Memory*:
* Same as Vector Memory with async support.
## Tool:
A unified interface for augmenting the LLM's functionality.
**Purpose**: Provides a lightweight abstraction for tool integration, accommodating both simple and complex tools.
**Features**:
* *Simple Tools*: Lazy wrappers for functions or basic utilities.
* *Complex Tools*: Abstract class for external APIs or multi-step operations.
## Loader:
Responsible for converting raw data into text.
**Purpose**: Handles preprocessing and content extraction from diverse formats.
**Features**:
* Covering limited type of documents, images, and audio files.
## Chunkers:
Utility to split long text into chunks.
**Features**:
* **Basic**:
* *FixedCharacterChunker*: Split text into fixed-size character chunks with optional overlapping.
* *FixedGroupChunker*: Splits text into K chunks. Supporting two levels, `word` and `character`, default is `character`.
* **Semantic**:
* *SemanticChunker*: Split text into semantically coherent chunks.
* *SimulatedAnnealingSemanticChunker*: Enhanced with Simulated Annealing optimization technique.
# Planned Feature
- A Cookbook with detailed implementation examples.
# License
This project is licensed under the GNU General Public License v3.0 License. See the [LICENSE](LICENSE) file for details.
Raw data
{
"_id": null,
"home_page": null,
"name": "llm-agent-toolkit",
"maintainer": null,
"docs_url": null,
"requires_python": null,
"maintainer_email": null,
"keywords": "llm, agent, toolkit, large language model, memory management, tool integration, multi-modality interaction, multi-step workflow",
"author": "jonah_whaler_2348",
"author_email": "jk_saga@proton.me",
"download_url": "https://files.pythonhosted.org/packages/78/5e/69971dd1a8d547f4e58040905affe6e4ffc21771fdfa904157a4f846854c/llm_agent_toolkit-0.0.19.tar.gz",
"platform": null,
"description": "![Banner](https://raw.githubusercontent.com/JonahWhaler/llm-agent-toolkit/main/images/repo-banner.jpeg)\n\n# LLM Agent Toolkit: Modular Components for AI Workflows\nLLM Agent Toolkit provides minimal, modular interfaces for core components in LLM-based applications. Simplify workflows with stateless interaction, embedding encoders, memory management, tool integration, and data loaders, designed for compatibility and scalability. It prioritizes simplicity and modularity by proposing minimal wrappers designed to work across common tools, discouraging direct access to underlying technologies. Specific implementations and examples will be documented separately in a Cookbook (planned).\n\nPyPI: ![PyPI Downloads](https://static.pepy.tech/badge/llm-agent-toolkit)\n\n## Attention!!!\nUsing this toolkit simplifies integration by providing unified and modular interfaces across platforms. Many configurations are intentionally kept at their default settings to prioritize ease of use. However, most of these components are extensible through abstract classes, allowing developers to define their own desired configurations for greater flexibility. While this approach enhances consistency and reduces complexity, advanced customization may require extending the provided abstractions. \n\nFor developers requiring full-range customization or access to the latest features, it is recommended to consider using native libraries like `ollama` and `openai` directly.\n\n# Table of Contents\n- [LLM Agent Toolkit: Modular Components for AI Workflows](#llm-agent-toolkit-modular-components-for-ai-workflows)\n - [Attention!!!](#attention)\n- [Table of Contents](#table-of-contents)\n- [Installation](#installation)\n- [Fundamental Components](#fundamental-components)\n - [Core:](#core)\n - [Encoder:](#encoder)\n - [Memory:](#memory)\n - [Tool:](#tool)\n - [Loader:](#loader)\n - [Chunkers:](#chunkers)\n- [Planned Feature](#planned-feature)\n- [License](#license)\n\n# Installation\n `\n pip install llm-agent-toolkit\n `\n\n# Fundamental Components\n## Core: \n\nA stateless chat completion interface to interact with the LLM.\n\n**Purpose**: Serves as the central execution layer that abstracts interaction with the underlying LLM model.\n\n**Features**:\n* Supports Text-to-Text and Image-to-Text.\n* Enables iterative executions for multi-step workflows.\n* Facilitates tool invocation as part of the workflow.\n* Support models from OpenAI and Ollama\n\n## Encoder:\nA standardized wrapper for embedding models.\n\n**Purpose**: Provides a minimal API to transform text into embeddings, usable with any common embedding model.\n\n**Features**:\n* Abstracts away model-specific details (e.g., dimensionality, framework differences).\n* Allows flexible integration with downstream components like Memory or retrieval mechanisms.\n* Support OpenAI, Ollama and Transformers.\n* Support asynchronous operation.\n\n## Memory: \nOffers essential context retention capabilities.\n\n**Purpose**: Allows efficient context management without hardcoding database or memory solutions.\n\n**Types**:\n1. *Short-term Memory*:\n * Maintains recent interactions for session-based context.\n2. *Vector Memory*:\n * Combines embedding and storage for retrieval-augmented workflows.\n * Includes optional metadata management for filtering results.\n * Support Faiss and Chroma\n3. *Async Vector Memory*:\n * Same as Vector Memory with async support.\n\n## Tool:\nA unified interface for augmenting the LLM's functionality.\n\n**Purpose**: Provides a lightweight abstraction for tool integration, accommodating both simple and complex tools.\n\n**Features**:\n* *Simple Tools*: Lazy wrappers for functions or basic utilities.\n* *Complex Tools*: Abstract class for external APIs or multi-step operations.\n\n## Loader:\nResponsible for converting raw data into text.\n\n**Purpose**: Handles preprocessing and content extraction from diverse formats.\n\n**Features**:\n* Covering limited type of documents, images, and audio files.\n\n## Chunkers:\nUtility to split long text into chunks.\n\n**Features**:\n* **Basic**: \n * *FixedCharacterChunker*: Split text into fixed-size character chunks with optional overlapping.\n * *FixedGroupChunker*: Splits text into K chunks. Supporting two levels, `word` and `character`, default is `character`.\n* **Semantic**:\n * *SemanticChunker*: Split text into semantically coherent chunks.\n * *SimulatedAnnealingSemanticChunker*: Enhanced with Simulated Annealing optimization technique.\n\n# Planned Feature\n- A Cookbook with detailed implementation examples.\n\n# License\nThis project is licensed under the GNU General Public License v3.0 License. See the [LICENSE](LICENSE) file for details.\n\n\n",
"bugtrack_url": null,
"license": "GPLv3",
"summary": "LLM Agent Toolkit provides minimal, modular interfaces for core components in LLM-based applications.",
"version": "0.0.19",
"project_urls": null,
"split_keywords": [
"llm",
" agent",
" toolkit",
" large language model",
" memory management",
" tool integration",
" multi-modality interaction",
" multi-step workflow"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "4bc4515b3e72d8058ccd42f772039024bea597a166c7fc5ac9b896d348b6f593",
"md5": "10aa15a381929e0fa88706a5452aa67a",
"sha256": "11b7c9079cac7d2c4a1bfca0e14b4c24f9940212dd6e4d3b0b4c631fc4cf6b9a"
},
"downloads": -1,
"filename": "llm_agent_toolkit-0.0.19-py3-none-any.whl",
"has_sig": false,
"md5_digest": "10aa15a381929e0fa88706a5452aa67a",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": null,
"size": 90421,
"upload_time": "2024-12-20T08:30:49",
"upload_time_iso_8601": "2024-12-20T08:30:49.115031Z",
"url": "https://files.pythonhosted.org/packages/4b/c4/515b3e72d8058ccd42f772039024bea597a166c7fc5ac9b896d348b6f593/llm_agent_toolkit-0.0.19-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "785e69971dd1a8d547f4e58040905affe6e4ffc21771fdfa904157a4f846854c",
"md5": "4ec57a845a2c899f862d9d23b6818eec",
"sha256": "ac23a9c537fbd5f8e77065196fbcfe0efa722f3d50e44c64741c5b87156f24c0"
},
"downloads": -1,
"filename": "llm_agent_toolkit-0.0.19.tar.gz",
"has_sig": false,
"md5_digest": "4ec57a845a2c899f862d9d23b6818eec",
"packagetype": "sdist",
"python_version": "source",
"requires_python": null,
"size": 67712,
"upload_time": "2024-12-20T08:30:52",
"upload_time_iso_8601": "2024-12-20T08:30:52.015337Z",
"url": "https://files.pythonhosted.org/packages/78/5e/69971dd1a8d547f4e58040905affe6e4ffc21771fdfa904157a4f846854c/llm_agent_toolkit-0.0.19.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2024-12-20 08:30:52",
"github": false,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"lcname": "llm-agent-toolkit"
}