Name | Version | Summary | date |
torch-model-archiver |
0.10.0 |
Torch Model Archiver is used for creating archives of trained neural net models that can be consumed by TorchServe inference |
2024-03-14 21:47:38 |
torchserve |
0.10.0 |
TorchServe is a tool for serving neural net models for inference |
2024-03-14 21:47:35 |
new-ai-benchmark |
2.7.0 |
AI Benchmark is an open source python library for evaluating AI performance of various hardware platforms, including CPUs, GPUs and TPUs. |
2024-03-10 18:53:48 |
sparsify-nightly |
1.7.0.20240304 |
Easy-to-use UI for automatically sparsifying neural networks and creating sparsification recipes for better inference performance and a smaller footprint |
2024-03-05 13:38:57 |
superlaser |
0.0.6 |
An MLOps library for LLM deployment w/ the vLLM engine on RunPod's infra. |
2024-03-02 20:54:35 |
cynet |
2.0.4 |
Learning Point Processes Using Deep Granger Nets |
2024-02-26 02:06:19 |
ANU-inversion-course |
0.1.1 |
ANU Inversion Course Package |
2024-02-16 04:10:46 |
torch-tensorrt |
2.2.0 |
Torch-TensorRT is a package which allows users to automatically compile PyTorch and TorchScript modules to TensorRT while remaining in PyTorch |
2024-02-14 01:49:39 |
infer-camembert |
0.2.0 |
Python implementation for text classification inference with CamemBERT fine-tuned models |
2024-02-08 09:52:19 |
samplex |
0.0.2 |
Samplers in MLX |
2024-02-02 21:44:41 |
psimpy |
0.2.0 |
Predictive and probabilistic simulation tools. |
2024-01-26 15:44:37 |
glai |
0.1.3 |
Easy deployment of quantized llama models on cpu |
2024-01-13 19:04:27 |
llama-memory |
0.0.1a1 |
Easy deployment of quantized llama models on cpu |
2024-01-09 01:59:55 |
tsdate |
0.1.6 |
Infer node ages from a tree sequence topology. |
2024-01-06 23:12:00 |
sparsify |
1.6.1 |
Easy-to-use UI for automatically sparsifying neural networks and creating sparsification recipes for better inference performance and a smaller footprint |
2023-12-20 14:28:37 |
sparseml |
1.6.1 |
Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models |
2023-12-20 14:24:12 |
sparsezoo |
1.6.1 |
Neural network model repository for highly sparse and sparse-quantized models with matching sparsification recipes |
2023-12-20 14:23:57 |
tritony |
0.0.16 |
Tiny configuration for Triton Inference Server |
2023-12-15 04:15:36 |
deepsparse-ent |
1.6.0 |
An inference runtime offering GPU-class performance on CPUs and APIs to integrate ML into your application |
2023-12-04 16:24:07 |
sbi |
0.22.0 |
Simulation-based inference. |
2023-12-04 11:12:10 |