PyDigger - unearthing stuff about Python


NameVersionSummarydate
ragtime 0.0.43 Ragtime 🎹 is an LLMOps framework to automatically evaluate Retrieval Augmented Generation (RAG) systems and compare different RAGs / LLMs 2024-06-10 15:20:30
indic-eval 0.1.0 A package to make LLM evaluation easier 2024-06-01 07:07:20
nutcracker-py 0.0.2a1 streamline LLM evaluation 2024-05-29 19:50:18
sed-scores-eval 0.0.4 (Threshold-Independent) Evaluation of Sound Event Detection Scores 2024-05-23 19:52:56
chainforge 0.3.2.0 A Visual Programming Environment for Prompt Engineering 2024-05-18 06:19:36
evaluate 0.4.2 HuggingFace community-driven open-source library of evaluation 2024-04-30 09:44:19
ntqr 0.3.2 Tools for the logic of evaluation using unlabeled data 2024-04-18 12:30:10
llama-index-packs-llama-dataset-metadata 0.1.4 llama-index packs llama_dataset_metadata integration 2024-04-08 19:39:22
enoslib 9.2.0 None 2024-04-03 11:34:17
lighteval 0.3.0 A lightweight and configurable evaluation package 2024-03-29 16:52:04
rag-eval 0.1.3 A RAG evaluation framework 2024-03-19 17:16:27
synthesized-datasets 1.7 Publically available datasets for benchmarking and evaluation. 2024-03-13 14:54:22
unbabel-comet 2.2.2 High-quality Machine Translation Evaluation 2024-03-13 11:27:34
easy-evaluator 0.0.0 A library for easy evaluation of language models 2024-03-03 15:30:15
reseval 0.1.6 Reproducible Subjective Evaluation 2024-03-03 05:32:16
tno.sdg.tabular.eval.utility-metrics 0.3.0 Utility metrics for tabular data 2024-02-28 13:23:02
llama-index-packs-rag-evaluator 0.1.3 llama-index packs rag_evaluator integration 2024-02-22 01:29:47
mt-thresholds 0.0.4 Tool to check how metric deltas for machine translation reflect on system-level human accuracies. 2024-02-12 20:40:24
lighthouz 0.0.5 Lighthouz AI Python SDK 2024-02-12 07:27:52
v-stream 0.1.2 STREAM: Spatio-TempoRal Evaluation and Analysis Metric for Video Generative Models 2024-01-25 08:02:46
hourdayweektotal
5415709397219316
Elapsed time: 0.68930s