Name | Version | Summary | date |
turkish-tokenizer |
0.2.25 |
Turkish tokenizer for Turkish language processing |
2025-08-27 13:54:11 |
hf-vram-calc |
1.0.2 |
GPU memory calculator for Hugging Face models with different data types and parallelization strategies |
2025-08-25 15:50:02 |
xlstm |
2.0.5 |
A novel LSTM variant with promising performance compared to Transformers or State Space Models. |
2025-08-24 14:38:49 |
megatron-fsdp |
0.1.0rc0 |
**Megatron-FSDP** is an NVIDIA-developed PyTorch extension that provides a high-performance implementation of Fully Sharded Data Parallelism (FSDP) |
2025-08-22 17:13:31 |
multihead-latent-attention |
0.1.0 |
Multi-head Latent Attention (MLA) - PyTorch |
2025-08-22 00:59:26 |
rotary-spatial-embeddings |
2025.8.21.2030 |
PyTorch implementation of Rotary Spatial Embeddings |
2025-08-21 20:33:18 |
spatial-grouping-attention |
2025.8.21.1739 |
PyTorch Implementation of Spatial Grouping Attention Layer |
2025-08-21 17:52:29 |
ptflops |
0.7.5 |
Flops counter for neural networks in pytorch framework |
2025-08-20 17:17:47 |
mantis-tsfm |
0.2.0 |
Mantis: Lightweight Calibrated Foundation Model for User-Friendly Time Series Classification |
2025-08-20 15:32:16 |
haystack-ai |
2.17.1 |
LLM framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. |
2025-08-20 09:18:52 |
polymesh-ai |
0.3.4 |
Transformer Library for 3D Mesh Processing |
2025-08-16 03:13:16 |
pymesh3d-nielsen |
0.2.0 |
Transformer Library for 3D Mesh Processing |
2025-08-16 00:22:51 |
model-vram-calc |
1.0.0 |
GPU memory calculator for Hugging Face models with different data types and parallelization strategies |
2025-08-15 13:27:18 |
trading-models |
0.1.3 |
MLP, CNN, Transformer models for time-series trading predictions. |
2025-08-15 09:44:16 |
timecopilot-uni2ts |
0.1.2 |
Unified Training of Universal Time Series Forecasting Transformers |
2025-08-14 03:24:56 |
megatron-core |
0.13.1 |
Megatron Core - a library for efficient and scalable training of transformer based models |
2025-08-12 18:34:07 |
megania |
0.0.1 |
IA Megan con tensores y tokenizador básico sin dependencias |
2025-08-11 01:28:45 |
microgpt |
0.0.2 |
Lightweight GPT implementation designed for resource-constrained environments |
2025-08-10 07:17:04 |
aviary-models |
1.2.1 |
A collection of machine learning models for materials discovery |
2025-08-10 00:36:56 |
convai-innovations |
1.1.2 |
Interactive LLM Training Academy - Learn to build language models from scratch |
2025-08-05 17:49:25 |