llm-bench


Namellm-bench JSON
Version 0.4.27 PyPI version JSON
download
home_pagehttps://github.com/b-Snaas/ollama-benchmark.git
SummaryLLM Benchmarking tool for OLLAMA
upload_time2024-07-01 10:03:23
maintainerNone
docs_urlNone
authorSnaas
requires_python<4.0,>=3.8
licenseMIT
keywords benchmark llama ollama llms local
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            # llm-benchmark (ollama-benchmark)

LLM Benchmark for Throughput via Ollama (Local LLMs)

## Installation Steps

```bash
pip install llm-bench
```

## Usage for general users directly

```bash
llm_bench run
```

## ollama installation with the following models installed

7B model can be run on machines with 8GB of RAM

13B model can be run on machines with 16GB of RAM

## Usage explaination

On Windows, Linux, and macOS, it will detect memory RAM size to first download required LLM models.

When memory RAM size is greater than or equal to 4GB, but less than 7GB, it will check if gemma:2b exist. The program implicitly pull the model.

```bash
ollama pull gemma:2b
```

When memory RAM size is greater than 7GB, but less than 15GB, it will check if these models exist. The program implicitly pull these models

```bash
ollama pull gemma:2b
ollama pull gemma:7b
ollama pull mistral:7b
ollama pull llama2:7b
ollama pull llava:7b
```

When memory RAM siz is greater than 15GB, it will check if these models exist. The program implicitly pull these models

```bash
ollama pull gemma:2b
ollama pull gemma:7b
ollama pull mistral:7b
ollama pull llama2:7b
ollama pull llama2:13b
ollama pull llava:7b
ollama pull llava:13b
```

## Python Poetry manually(advanced) installation

<https://python-poetry.org/docs/#installing-manually>

## For developers to develop new features on Windows Powershell or on Ubuntu Linux or macOS

```bash
python3 -m venv .venv
. ./.venv/bin/activate
pip install -U pip setuptools
pip install poetry
```

## Usage in Python virtual environment

```bash
poetry shell
poetry install
llm_benchmark hello jason
```

### Example #1 send systeminfo and benchmark results to a remote server

```bash
llm_bench run
```

### Example #2 Do not send systeminfo and benchmark results to a remote server

```bash
llm_bench run --no-sendinfo
```

### Example #3 Benchmark run on explicitly given the path to the ollama executable (When you built your own developer version of ollama)

```bash
llm_bench run --ollamabin=~/code/ollama/ollama
```

## Reference

[Ollama](https://ollama.com)


            

Raw data

            {
    "_id": null,
    "home_page": "https://github.com/b-Snaas/ollama-benchmark.git",
    "name": "llm-bench",
    "maintainer": null,
    "docs_url": null,
    "requires_python": "<4.0,>=3.8",
    "maintainer_email": null,
    "keywords": "benchmark, llama, ollama, llms, local",
    "author": "Snaas",
    "author_email": null,
    "download_url": "https://files.pythonhosted.org/packages/f5/d3/793520855ea79dba46a520dae4aa2b9b649211c54abb9e9973d831335883/llm_bench-0.4.27.tar.gz",
    "platform": null,
    "description": "# llm-benchmark (ollama-benchmark)\n\nLLM Benchmark for Throughput via Ollama (Local LLMs)\n\n## Installation Steps\n\n```bash\npip install llm-bench\n```\n\n## Usage for general users directly\n\n```bash\nllm_bench run\n```\n\n## ollama installation with the following models installed\n\n7B model can be run on machines with 8GB of RAM\n\n13B model can be run on machines with 16GB of RAM\n\n## Usage explaination\n\nOn Windows, Linux, and macOS, it will detect memory RAM size to first download required LLM models.\n\nWhen memory RAM size is greater than or equal to 4GB, but less than 7GB, it will check if gemma:2b exist. The program implicitly pull the model.\n\n```bash\nollama pull gemma:2b\n```\n\nWhen memory RAM size is greater than 7GB, but less than 15GB, it will check if these models exist. The program implicitly pull these models\n\n```bash\nollama pull gemma:2b\nollama pull gemma:7b\nollama pull mistral:7b\nollama pull llama2:7b\nollama pull llava:7b\n```\n\nWhen memory RAM siz is greater than 15GB, it will check if these models exist. The program implicitly pull these models\n\n```bash\nollama pull gemma:2b\nollama pull gemma:7b\nollama pull mistral:7b\nollama pull llama2:7b\nollama pull llama2:13b\nollama pull llava:7b\nollama pull llava:13b\n```\n\n## Python Poetry manually(advanced) installation\n\n<https://python-poetry.org/docs/#installing-manually>\n\n## For developers to develop new features on Windows Powershell or on Ubuntu Linux or macOS\n\n```bash\npython3 -m venv .venv\n. ./.venv/bin/activate\npip install -U pip setuptools\npip install poetry\n```\n\n## Usage in Python virtual environment\n\n```bash\npoetry shell\npoetry install\nllm_benchmark hello jason\n```\n\n### Example #1 send systeminfo and benchmark results to a remote server\n\n```bash\nllm_bench run\n```\n\n### Example #2 Do not send systeminfo and benchmark results to a remote server\n\n```bash\nllm_bench run --no-sendinfo\n```\n\n### Example #3 Benchmark run on explicitly given the path to the ollama executable (When you built your own developer version of ollama)\n\n```bash\nllm_bench run --ollamabin=~/code/ollama/ollama\n```\n\n## Reference\n\n[Ollama](https://ollama.com)\n\n",
    "bugtrack_url": null,
    "license": "MIT",
    "summary": "LLM Benchmarking tool for OLLAMA",
    "version": "0.4.27",
    "project_urls": {
        "Homepage": "https://github.com/b-Snaas/ollama-benchmark.git"
    },
    "split_keywords": [
        "benchmark",
        " llama",
        " ollama",
        " llms",
        " local"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "414f0317a91095d1538918963cc915aedd0986972e2390cddb785f64c99fa1b3",
                "md5": "e1504f1d52aab847d4b0de54a6228792",
                "sha256": "887f7dc8a72b5f49f7fb61a24a7a6b55fe3c9a96180a1a859fe239bd0fe3e2e0"
            },
            "downloads": -1,
            "filename": "llm_bench-0.4.27-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "e1504f1d52aab847d4b0de54a6228792",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": "<4.0,>=3.8",
            "size": 18667,
            "upload_time": "2024-07-01T10:03:21",
            "upload_time_iso_8601": "2024-07-01T10:03:21.439372Z",
            "url": "https://files.pythonhosted.org/packages/41/4f/0317a91095d1538918963cc915aedd0986972e2390cddb785f64c99fa1b3/llm_bench-0.4.27-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "f5d3793520855ea79dba46a520dae4aa2b9b649211c54abb9e9973d831335883",
                "md5": "71fa1e3b5ed68e743d5e0c80529d86c4",
                "sha256": "e73325f6cb8aff0879f5d3f523b4533537a58ae7e5a9352e591d60ba7bdef78d"
            },
            "downloads": -1,
            "filename": "llm_bench-0.4.27.tar.gz",
            "has_sig": false,
            "md5_digest": "71fa1e3b5ed68e743d5e0c80529d86c4",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": "<4.0,>=3.8",
            "size": 12538,
            "upload_time": "2024-07-01T10:03:23",
            "upload_time_iso_8601": "2024-07-01T10:03:23.545084Z",
            "url": "https://files.pythonhosted.org/packages/f5/d3/793520855ea79dba46a520dae4aa2b9b649211c54abb9e9973d831335883/llm_bench-0.4.27.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-07-01 10:03:23",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "b-Snaas",
    "github_project": "ollama-benchmark",
    "travis_ci": false,
    "coveralls": false,
    "github_actions": true,
    "requirements": [],
    "lcname": "llm-bench"
}
        
Elapsed time: 0.28808s