Name | Version | Summary | date |
autonomize-model-sdk |
1.1.49 |
SDK for creating and managing machine learning pipelines. |
2025-07-15 10:41:56 |
yaicli |
0.8.1 |
A simple CLI tool to interact with LLM |
2025-07-15 08:44:03 |
probabilistic-quantum-reasoner |
0.1.0 |
A quantum-classical hybrid reasoning engine for uncertainty-aware AI inference |
2025-07-14 12:34:17 |
ggufloader |
1.0.3 |
A local LLM runner for loading and chatting with GGUF models |
2025-07-11 19:31:43 |
xgrammar |
0.1.21 |
Efficient, Flexible and Portable Structured Generation |
2025-07-10 19:34:14 |
optimum-rbln |
0.8.1 |
Optimum RBLN is the interface between the HuggingFace Transformers and Diffusers libraries and RBLN accelerators. It provides a set of tools enabling easy model loading and inference on single and multiple rbln device settings for different downstream tasks. |
2025-07-04 02:37:11 |
guidellm |
0.2.1 |
Guidance platform for deploying and managing large language models. |
2025-04-29 17:49:39 |
anemoi-inference |
0.5.1 |
A package to run inference from data-driven forecasts weather models. |
2025-04-09 10:51:44 |
tritonclient |
2.55.0 |
Python client library and utilities for communicating with Triton Inference Server |
2025-02-26 20:07:24 |
figaro |
1.7.8 |
FIGARO: Fast Inference for GW Astronomy, Research & Observations |
2025-02-19 15:38:29 |
pixano-inference |
0.4.0 |
Deploy inference models served for Artificial Intelligence solutions including but not limited to Pixano. |
2025-02-18 15:35:05 |
causaltensor |
0.1.10 |
Package for causal inference in panels |
2025-02-07 16:58:55 |
optimum-neuron |
0.0.28 |
Optimum Neuron is the interface between the Hugging Face Transformers and Diffusers libraries and AWS Trainium and Inferentia accelerators. It provides a set of tools enabling easy model loading, training and inference on single and multiple neuron core settings for different downstream tasks. |
2025-02-07 10:24:33 |
vidur |
0.0.9 |
A LLM inference cluster simulator |
2025-02-02 07:31:31 |
friendli-client |
1.5.8 |
Client of Friendli Suite. |
2025-01-24 05:33:38 |
hot-fair-utilities |
2.0.10 |
Utilities for AI - Assisted Mapping fAIr |
2025-01-12 21:41:38 |
finsim |
1.0.2 |
Financial simulation and inference |
2025-01-10 18:26:26 |
tritony |
0.0.20 |
Tiny configuration for Triton Inference Server |
2025-01-10 10:25:04 |
mcp-server-replicate |
0.1.9 |
FastMCP server implementation for the Replicate API, providing resource-based access to AI model inference |
2025-01-06 23:10:23 |
pysmatch |
1.0 |
Propensity Score Matching(PSM) on python |
2025-01-06 14:12:15 |