Name | Version | Summary | date |
intel-extension-for-transformers |
1.4.1 |
Repository of Intel® Intel Extension for Transformers |
2024-04-21 08:29:37 |
bitsandbytes |
0.43.1 |
k-bit optimizers and matrix multiplication routines. |
2024-04-11 18:23:49 |
ctranslate2 |
4.2.0 |
Fast inference engine for Transformer models |
2024-04-10 17:23:45 |
neural-compressor-3x-tf |
2.5.1 |
Repository of Intel® Neural Compressor |
2024-04-03 14:12:01 |
neural-compressor-3x-pt |
2.5.1 |
Repository of Intel® Neural Compressor |
2024-04-03 14:11:47 |
neural-compressor-3x-ort |
2.5.1 |
Repository of Intel® Neural Compressor |
2024-04-03 14:11:31 |
neural-compressor |
2.5.1 |
Repository of Intel® Neural Compressor |
2024-04-03 14:11:06 |
bitlinear-pytorch |
0.4.0 |
Implementation of the BitLinear layer from: The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits |
2024-03-19 00:07:54 |
quanto |
0.1.0 |
A quantization toolkit for pytorch. |
2024-03-13 08:42:22 |
owlite |
0.0.5 |
A fake package to warn the user they are not installing the correct package. |
2024-03-11 10:26:06 |
auto-round |
0.1 |
Repository of AutoRound: Advanced Weight-Only Quantization Algorithm for LLMs |
2024-03-08 07:52:40 |
nncf |
2.9.0 |
Neural Networks Compression Framework |
2024-03-06 11:39:35 |
sparsify-nightly |
1.7.0.20240304 |
Easy-to-use UI for automatically sparsifying neural networks and creating sparsification recipes for better inference performance and a smaller footprint |
2024-03-05 13:38:57 |
faster-whisper |
1.0.1 |
Faster Whisper transcription with CTranslate2 |
2024-03-01 10:44:19 |
autoawq-kernels |
0.0.6 |
AutoAWQ Kernels implements the AWQ kernels. |
2024-02-24 22:42:37 |
optimum-intel |
1.15.2 |
Optimum Library is an extension of the Hugging Face Transformers library, providing a framework to integrate third-party libraries from Hardware Partners and interface with their specific functionality. |
2024-02-22 17:19:09 |
nendo-plugin-quantize-core |
0.2.6 |
Nendo Plugin for audio quantization with grid detection and time-stretching |
2024-02-21 09:42:22 |
auto-gptq |
0.7.0 |
An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm. |
2024-02-16 12:52:41 |
clika-inference |
0.0.2 |
A fake package to warn the user they are not installing the correct package. |
2024-01-31 06:43:35 |
clika-compression |
0.0.2 |
A fake package to warn the user they are not installing the correct package. |
2024-01-31 06:43:33 |