ggufloader


Nameggufloader JSON
Version 1.0.3 PyPI version JSON
download
home_pageNone
SummaryA local LLM runner for loading and chatting with GGUF models
upload_time2025-07-11 19:31:43
maintainerNone
docs_urlNone
authorNone
requires_python>=3.8
licenseMIT
keywords llm gguf llmloader llmrunner ggufloader inference
VCS
bugtrack_url
requirements llama-cpp-python PySide6
Travis-CI No Travis.
coveralls test coverage No coveralls.
            # GGUF Loader
![License](https://img.shields.io/github/license/ggufloader/gguf-loader)
![Last Commit](https://img.shields.io/github/last-commit/ggufloader/gguf-loader)
![Repo Size](https://img.shields.io/github/repo-size/ggufloader/gguf-loader)
![Issues](https://img.shields.io/github/issues/ggufloader/gguf-loader)

Easiest way to run and manage GGUF-based LLMs locally β€” with drag-and-drop GUI, plugin system, and zero terminal setup.

## πŸš€ Install in One Line

```bash
pip install ggufloader
```

```bash
ggufloader
```

Works on Windows, Linux, and macOS.

## 🌟 Why GGUF Loader?

- βœ… No terminal needed β€” fully GUI-based  
- πŸ”Œ Blender-style plugin system for custom tools  
- 🧠 Built for local LLMs: Mistral, LLaMA, DeepSeek, etc.  
- ⚑ Runs even on Intel i5 + 16GB RAM  
- πŸ“ Drag and drop `.gguf` models and run  

> Ideal for beginners, researchers, or anyone avoiding cloud APIs.

## πŸ–ΌοΈ Preview
![Image 1](https://raw.githubusercontent.com/GGUFloader/gguf-loader/main/1.png)  
![Image 2](https://raw.githubusercontent.com/GGUFloader/gguf-loader/main/2.png)

## πŸ“¦ Features

| Feature                 | Description                                     |
|------------------------|-------------------------------------------------|
| GUI for GGUF LLMs      | Run models without terminal or coding          |
| Addon System           | Extend with plugins like summarizers, exporters |
| Cross-platform Support | Windows / Linux / macOS                        |
| Multi-Model Friendly   | Mistral, LLaMA, DeepSeek, Yi, Gemma, OpenHermes |
| Memory-Efficient       | Works on low-spec laptops (16GB RAM)           |

## πŸ“˜ How It Works

1. Install via `pip install ggufloader`  
2. Launch it with `ggufloader`  
3. Drag and drop your `.gguf` models  
4. Start chatting with local LLMs instantly  
5. Add features via plugins (like PDF, email, spreadsheet)

## βš™οΈ Advanced Usage

- Run via CLI: `ggufloader --model mistral.gguf`  
- Use Addon: `ggufloader --addon summarize_pdf`  
- Import in Python:

```python
from ggufloader import chat
chat("Hello, local world!")
```

## 🧠 Use Cases

- Run LLaMA 3 or Mistral on your own machine  
- Build your own local ChatGPT  
- Summarize documents with AI locally  
- Run AI completely offline (no API needed)

> *Perfect for local-first AI, privacy-focused developers, and automation hackers.*

## πŸ’‘ Comparison with Similar Tools

| Tool           | GUI | Addons | Pip Install | Offline Use | Notes                          |
|----------------|-----|--------|-------------|-------------|-------------------------------|
| GGUF Loader βœ…  | βœ…  | βœ…     | βœ…          | βœ…          | Modular, privacy-first         |
| LM Studio      | βœ…  | ❌     | ❌          | βœ…          | More polished, less flexible   |
| Ollama         | ❌  | ❌     | ❌          | βœ…          | CLI-first, tightly scoped      |
| GPT4All        | βœ…  | ❌     | βœ…          | βœ…          | Limited extension support      |

## πŸ“Ž Links

- 🏠 Website: [ggufloader.github.io](https://ggufloader.github.io)  
- πŸ§ͺ PyPI: [https://pypi.org/project/ggufloader](https://pypi.org/project/ggufloader)
- πŸ“‚ GitHub: [github.com/ggufloader/ggufloader](https://github.com/ggufloader/ggufloader)

## ❓ FAQ

**Q: What is GGUF?**  
A new file format for optimized local LLM models (used in llama.cpp, Mistral, etc).

**Q: Can I use it offline?**  
Yes β€” GGUF Loader runs completely offline and doesn’t use OpenAI or API calls.

**Q: Does it support plugins?**  
Yes! Use addons for PDF reading, summarization, chatbot modes, spreadsheet processing, and more.


            

Raw data

            {
    "_id": null,
    "home_page": null,
    "name": "ggufloader",
    "maintainer": null,
    "docs_url": null,
    "requires_python": ">=3.8",
    "maintainer_email": null,
    "keywords": "llm, gguf, LLMLoader, LLMRunner, ggufloader, Inference",
    "author": null,
    "author_email": "Hussain <hossainnazary475@gmail.com>",
    "download_url": "https://files.pythonhosted.org/packages/9e/56/e7987d92d49928ba1f8e304df8c35c92dc3589e5acef9dfe13502101a375/ggufloader-1.0.3.tar.gz",
    "platform": null,
    "description": "# GGUF Loader\r\n![License](https://img.shields.io/github/license/ggufloader/gguf-loader)\r\n![Last Commit](https://img.shields.io/github/last-commit/ggufloader/gguf-loader)\r\n![Repo Size](https://img.shields.io/github/repo-size/ggufloader/gguf-loader)\r\n![Issues](https://img.shields.io/github/issues/ggufloader/gguf-loader)\r\n\r\nEasiest way to run and manage GGUF-based LLMs locally \u2014 with drag-and-drop GUI, plugin system, and zero terminal setup.\r\n\r\n## \ud83d\ude80 Install in One Line\r\n\r\n```bash\r\npip install ggufloader\r\n```\r\n\r\n```bash\r\nggufloader\r\n```\r\n\r\nWorks on Windows, Linux, and macOS.\r\n\r\n## \ud83c\udf1f Why GGUF Loader?\r\n\r\n- \u2705 No terminal needed \u2014 fully GUI-based  \r\n- \ud83d\udd0c Blender-style plugin system for custom tools  \r\n- \ud83e\udde0 Built for local LLMs: Mistral, LLaMA, DeepSeek, etc.  \r\n- \u26a1 Runs even on Intel i5 + 16GB RAM  \r\n- \ud83d\udcc1 Drag and drop `.gguf` models and run  \r\n\r\n> Ideal for beginners, researchers, or anyone avoiding cloud APIs.\r\n\r\n## \ud83d\uddbc\ufe0f Preview\r\n![Image 1](https://raw.githubusercontent.com/GGUFloader/gguf-loader/main/1.png)  \r\n![Image 2](https://raw.githubusercontent.com/GGUFloader/gguf-loader/main/2.png)\r\n\r\n## \ud83d\udce6 Features\r\n\r\n| Feature                 | Description                                     |\r\n|------------------------|-------------------------------------------------|\r\n| GUI for GGUF LLMs      | Run models without terminal or coding          |\r\n| Addon System           | Extend with plugins like summarizers, exporters |\r\n| Cross-platform Support | Windows / Linux / macOS                        |\r\n| Multi-Model Friendly   | Mistral, LLaMA, DeepSeek, Yi, Gemma, OpenHermes |\r\n| Memory-Efficient       | Works on low-spec laptops (16GB RAM)           |\r\n\r\n## \ud83d\udcd8 How It Works\r\n\r\n1. Install via `pip install ggufloader`  \r\n2. Launch it with `ggufloader`  \r\n3. Drag and drop your `.gguf` models  \r\n4. Start chatting with local LLMs instantly  \r\n5. Add features via plugins (like PDF, email, spreadsheet)\r\n\r\n## \u2699\ufe0f Advanced Usage\r\n\r\n- Run via CLI: `ggufloader --model mistral.gguf`  \r\n- Use Addon: `ggufloader --addon summarize_pdf`  \r\n- Import in Python:\r\n\r\n```python\r\nfrom ggufloader import chat\r\nchat(\"Hello, local world!\")\r\n```\r\n\r\n## \ud83e\udde0 Use Cases\r\n\r\n- Run LLaMA 3 or Mistral on your own machine  \r\n- Build your own local ChatGPT  \r\n- Summarize documents with AI locally  \r\n- Run AI completely offline (no API needed)\r\n\r\n> *Perfect for local-first AI, privacy-focused developers, and automation hackers.*\r\n\r\n## \ud83d\udca1 Comparison with Similar Tools\r\n\r\n| Tool           | GUI | Addons | Pip Install | Offline Use | Notes                          |\r\n|----------------|-----|--------|-------------|-------------|-------------------------------|\r\n| GGUF Loader \u2705  | \u2705  | \u2705     | \u2705          | \u2705          | Modular, privacy-first         |\r\n| LM Studio      | \u2705  | \u274c     | \u274c          | \u2705          | More polished, less flexible   |\r\n| Ollama         | \u274c  | \u274c     | \u274c          | \u2705          | CLI-first, tightly scoped      |\r\n| GPT4All        | \u2705  | \u274c     | \u2705          | \u2705          | Limited extension support      |\r\n\r\n## \ud83d\udcce Links\r\n\r\n- \ud83c\udfe0 Website: [ggufloader.github.io](https://ggufloader.github.io)  \r\n- \ud83e\uddea PyPI: [https://pypi.org/project/ggufloader](https://pypi.org/project/ggufloader)\r\n- \ud83d\udcc2 GitHub: [github.com/ggufloader/ggufloader](https://github.com/ggufloader/ggufloader)\r\n\r\n## \u2753 FAQ\r\n\r\n**Q: What is GGUF?**  \r\nA new file format for optimized local LLM models (used in llama.cpp, Mistral, etc).\r\n\r\n**Q: Can I use it offline?**  \r\nYes \u2014 GGUF Loader runs completely offline and doesn\u2019t use OpenAI or API calls.\r\n\r\n**Q: Does it support plugins?**  \r\nYes! Use addons for PDF reading, summarization, chatbot modes, spreadsheet processing, and more.\r\n\r\n",
    "bugtrack_url": null,
    "license": "MIT",
    "summary": "A local LLM runner for loading and chatting with GGUF models",
    "version": "1.0.3",
    "project_urls": {
        "Homepage": "https://github.com/GGUFloader/gguf-loader",
        "Issues": "https://github.com/GGUFloader/gguf-loader/issues",
        "Repository": "https://github.com/GGUFloader/gguf-loader"
    },
    "split_keywords": [
        "llm",
        " gguf",
        " llmloader",
        " llmrunner",
        " ggufloader",
        " inference"
    ],
    "urls": [
        {
            "comment_text": null,
            "digests": {
                "blake2b_256": "7328d45bdf16e82996fad005491ca98dce75e7f64a440b3b91498e7ad9b98ef9",
                "md5": "b1f2c21a1dcbf8ca16415c25174684e4",
                "sha256": "0cb21d07ba29f49db937d487c18037d9e7e4d213f8ca97571c63d50667e7a8f9"
            },
            "downloads": -1,
            "filename": "ggufloader-1.0.3-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "b1f2c21a1dcbf8ca16415c25174684e4",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.8",
            "size": 21597,
            "upload_time": "2025-07-11T19:30:49",
            "upload_time_iso_8601": "2025-07-11T19:30:49.726037Z",
            "url": "https://files.pythonhosted.org/packages/73/28/d45bdf16e82996fad005491ca98dce75e7f64a440b3b91498e7ad9b98ef9/ggufloader-1.0.3-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": null,
            "digests": {
                "blake2b_256": "9e56e7987d92d49928ba1f8e304df8c35c92dc3589e5acef9dfe13502101a375",
                "md5": "bc58029753d6aaa68e992c01cbcfbac3",
                "sha256": "6c0ff95fccb703bd8f238e6b699b95ceace1b218f6874b8577ca5d95b0c9d915"
            },
            "downloads": -1,
            "filename": "ggufloader-1.0.3.tar.gz",
            "has_sig": false,
            "md5_digest": "bc58029753d6aaa68e992c01cbcfbac3",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.8",
            "size": 20045,
            "upload_time": "2025-07-11T19:31:43",
            "upload_time_iso_8601": "2025-07-11T19:31:43.356999Z",
            "url": "https://files.pythonhosted.org/packages/9e/56/e7987d92d49928ba1f8e304df8c35c92dc3589e5acef9dfe13502101a375/ggufloader-1.0.3.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2025-07-11 19:31:43",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "GGUFloader",
    "github_project": "gguf-loader",
    "travis_ci": false,
    "coveralls": false,
    "github_actions": true,
    "requirements": [
        {
            "name": "llama-cpp-python",
            "specs": []
        },
        {
            "name": "PySide6",
            "specs": []
        }
    ],
    "lcname": "ggufloader"
}
        
Elapsed time: 1.30138s