an AutoML library that builds, optimizes, and evaluates machine learning pipelines using domain-specific objective functions
Testing framework for sequence labeling
A framework for evaluating autoregressive language models
MS-COCO Caption Evaluation for Python 3
Safe, minimalistic evaluator of python expression using ast module
Send Sir Perceval on a quest to fetch and gather data from software repositories.
Python Mathematical Expression Evaluator
Provides Python bindings for popular Information Retrieval measures implemented within trec_eval.
Evaluation tools for the SIGSEP MUS database
Discover and retrieve water data from U.S. federal hydrologic web services.
Safely evaluate AST nodes without side effects
evalutils helps users create extensions for grand-challenge.org
"EvalPlus for rigourous evaluation of LLM-synthesized code"
scikit-learn model evaluation made easy: plots, tables andmarkdown reports.
Well tested evaluation framework for Text summarization
EM algorithms for integrated spatial and spectral models.
A library for providing a simple interface to create new metrics and an easy-to-use toolkit for metric computations and checkpointing.
Contains the integration code of AzureML Evaluate with Mlflow.
Easily computing clip embeddings and building a clip retrieval system with them
Provides Python bindings for popular Information Retrieval measures implemented within trec_eval.
Package for fast computation of BSS Eval metrics for source separation
Model evaluation for Machine Learning pipelines.
Toolkit for summarization evaluation
Use EvalAI through command line interface
Show value of an importable object
Validation and secure evaluation of untrusted python expressions
HydroEval: An Evaluator for Streamflow Time Series In Python
clusteval is a python package for unsupervised cluster validation.
A package to evaluate how close a synthetic data set is to real data.
Performs async evaluations of strings
Command-line evaluation tools for named entity linking and (cross-document) coreference resolution
Debugging middlewares for zope.publisher-based web applications
Evaluating ASR (automatic speech recognition) hypotheses, i.e. computing word error rate.
PMML evaluator library for Python
An information retrieval evaluation script based on the C/W/L framework that is TREC Compatible and provides a replacement for INST_EVAL, RBP_EVAL, TBG_EVAL, UMeasure and TREC_EVAL scripts. All measurements are reported in the same units making all metrics directly comparable.
A package providing text segmentation evaluation metrics and utilities
A library for providing a simple interface to create new metrics and an easy-to-use toolkit for metric computations and checkpointing.
ranx: A Blazing-Fast Python Library for Ranking Evaluation, Comparison, and Fusion
PH Evaluator - an efficient Poker Hand Evaluator based on a Perfect Hash algorithm