Evaluate Digitalization Data
Mir_eval package ported to pytorch
rank_eval: A Blazing Fast Python Library for Ranking Evaluation and Comparison
This package is written for the evaluation of audio generation model.
An easy to use collection of Speech enhancement measures
Python SDK to configure and run evaluations for your LLM-based application
ADC Evaluation Library
Time Series analysis and evaluation tools
TTS Evaluation
This package provides measurement tools for Generative Adversarial Networks (GANs), including Inception Score (IS), Fréchet Inception Distance (FID), Kernel Inception Distance (KID), and Precision and Recall (PR). These metrics are used to evaluate the quality and diversity of generated images in GANs. The package streamlines the use of these metrics, making it easier to apply them to your work.
A user-friendly feature evaluation and selection package.
Evaluation method for the DRAGON benchmark
A plug & play evaluator for self-supervised image classification.
Library for evaluating SafeGraph data
A PEP 523 compatible frame evaluator
A RAG evaluation framework
Embodied agent interface evaluation for VirtualHome
The evaluation of the epic cardio biosensor integrated into napari
Interpretable Evaluation for Natural Language Processing
Vital Agent Eval Env
Used to print entries from jsonl files, developed with LLM evals in mind.
Tool for safe (or less safe) evaluation of strings as math expressions
Evaluate arbitrary JavaScript from Python using a NodeJS sidecar
LLM Application Debug/Eval UI on top of AIConfig
A visualization package for model score evaluation.
Agiflow (EVAL) for Python
TestScript FHIR resource evaluator
AI Maintainer Agent Harness for our benchmarking and Marketplace API and platform
The evaluation component of the sci-annot framework
Topic Modeling Evaluation
Supplementary code and materials for paper "On Model Evaluation under Non-constant Class Imbalance" by Brabec et al.
A library for evaluating image generation models
Safe execute expr code.
Jinja2 Extension for getting eval() result.
An Extendable Evaluation Pipeline for Named Entity Drill-Down Analysis
A placeholder package to reserve the name llms.
An asr evaluation tool targetting for testing Portuguese(Brazil) sentence.
Quickly evaluate multi-label classifiers in various metrics
calculations with physical quantities
Library for evaluating RAG using Nuclia's models
A Comprehensive Platform for Automated Testing and Analysis For Supervised Machine Learning Tasks.
aigc_evals
Dashboard for Quality-driven NER.
A package with utility functions for evaluating conformal predictors
This repository provides a small Python wrapper for the Matlab tool SNR Eval provided by Labrosa: https://labrosa.ee.columbia.edu/projects/snreval
A Python library for Multi-Agent Reinforcement Learning evaluation.
Eval