Name | Version | Summary | date |
caption-flow |
0.1.0 |
Self-contained distributed community captioning system |
2025-08-12 03:39:03 |
llmq |
0.0.4 |
High-Performance vLLM Job Queue Package |
2025-08-11 08:21:32 |
langchain-llm-config |
0.2.0 |
A comprehensive LLM configuration package supporting multiple providers (OpenAI, VLLM, Gemini, Infinity) for chat assistants and embeddings |
2025-08-08 12:17:48 |
speculators |
0.1.0 |
A unified library for creating, representing, and storing speculative decoding algorithms for LLM serving such as in vLLM. |
2025-08-08 01:22:17 |
vllm-judge |
0.1.8 |
LLM-as-a-Judge evaluations for vLLM hosted models |
2025-08-07 21:45:49 |
ovllm |
0.3.0 |
One-line vLLM wrapper with gorgeous DSPy integration |
2025-08-04 22:42:36 |
kvcached |
0.0.1 |
A KV cache management system that supports on-demand KV cache allocation for LLMs with GPU virtual memory |
2025-07-25 20:40:58 |
guidellm |
0.2.1 |
Guidance platform for deploying and managing large language models. |
2025-04-29 17:49:39 |
vllmocr |
0.3.8 |
OCR project using LLMs |
2025-03-13 10:43:39 |
faster-translate |
1.0.2 |
A high-performance translation library using CTTranslate2 and vLLM. |
2025-03-03 23:14:12 |
bocr |
0.2.0 |
A Python package for OCR using Vision LLMs |
2025-02-22 13:33:12 |
openllm |
0.6.19 |
OpenLLM: Self-hosting LLMs Made Easy. |
2025-02-15 07:48:10 |
sparrow-parse |
0.5.0 |
Sparrow Parse is a Python package (part of Sparrow) for parsing and extracting information from documents. |
2025-01-09 12:28:45 |
happy-vllm |
1.1.12 |
happy_vllm is a REST API for vLLM, production ready |
2025-01-06 11:07:33 |
pictoken |
0.1.0 |
Calculate image tokens for (Azure) OpenAI models. Offers multiple utilities to resize images to reduce token usage. |
2024-10-07 21:19:08 |
optimum-benchmark |
0.4.0 |
Optimum-Benchmark is a unified multi-backend utility for benchmarking Transformers, Timm, Diffusers and Sentence-Transformers with full support of Optimum's hardware optimizations & quantization schemes. |
2024-07-31 08:35:53 |