Show value of an importable object
HuggingFace community-driven open-source library of evaluation
Safely evaluate AST nodes without side effects
Like `typing._eval_type`, but lets older Python versions use newer typing features.
A simple, safe single expression evaluator library.
Testing framework for sequence labeling
Safe, minimalistic evaluator of python expression using ast module
A framework for evaluating language models
Validation and secure evaluation of untrusted python expressions
MS-COCO Caption Evaluation for Python 3
an AutoML library that builds, optimizes, and evaluates machine learning pipelines using domain-specific objective functions
A getattr and setattr that works on nested objects, lists, dicts, and any combination thereof without resorting to eval
Evaluation tools for the SIGSEP MUS database
Use EvalAI through command line interface
Python Mathematical Expression Evaluator
Provides Python bindings for popular Information Retrieval measures implemented within trec_eval.
Limited evaluator
A library for providing a simple interface to create new metrics and an easy-to-use toolkit for metric computations and checkpointing.
The open-source evaluation framework for LLMs.
"EvalPlus for rigourous evaluation of LLM-synthesized code"
Universal library for evaluating AI models
Package for fast computation of BSS Eval metrics for source separation
Faster interpretation of the original COCOEval
Easily computing clip embeddings and building a clip retrieval system with them
evalutils helps users create extensions for grand-challenge.org
A poker hand evaluation and equity calculation library
Library to systematically track and evaluate LLM based applications.
A library for providing a simple interface to create new metrics and an easy-to-use toolkit for metric computations and checkpointing.
Relative clustering validation to select best number of clusters
Provides Python bindings for popular Information Retrieval measures implemented within trec_eval.
Evaluating and scoring financial data
Contains the integration code of AzureML Evaluate with Mlflow.
Serialization based on ast.literal_eval
A custom Streamlit component to evaluate arbitrary Javascript expressions.
Performs async evaluations of strings
An information retrieval evaluation script based on the C/W/L framework that is TREC Compatible and provides a replacement for INST_EVAL, RBP_EVAL, TBG_EVAL, UMeasure and TREC_EVAL scripts. All measurements are reported in the same units making all metrics directly comparable.
LLM Evaluations
AlpacaEval : An Automatic Evaluator of Instruction-following Models
Send Sir Perceval on a quest to fetch and gather data from software repositories.
A flexible, generalized tree-based data structure.
EM algorithms for integrated spatial and spectral models.
Wrapper around ast.literal_eval with new {foo='bar', key=None} dict syntax.
A toolkit for survey evaluation
Bundle of Perceval backends for OPNFV ecosystem.
Bundle of Perceval backends for Mozilla ecosystem.