Show value of an importable object
Like `typing._eval_type`, but lets older Python versions use newer typing features.
HuggingFace community-driven open-source library of evaluation
Safely evaluate AST nodes without side effects
A simple, safe single expression evaluator library.
Evalica, your favourite evaluation toolkit.
Safe, minimalistic evaluator of python expression using ast module
Validation and secure evaluation of untrusted python expressions
an AutoML library that builds, optimizes, and evaluates machine learning pipelines using domain-specific objective functions
Testing framework for sequence labeling
A framework for evaluating language models
A getattr and setattr that works on nested objects, lists, dicts, and any combination thereof without resorting to eval
Use EvalAI through command line interface
The open-source evaluation framework for LLMs.
Evaluation tools for the SIGSEP MUS database
MS-COCO Caption Evaluation for Python 3
EvalScope: Lightweight LLMs Evaluation Framework
evalutils helps users create extensions for grand-challenge.org
A poker hand evaluation and equity calculation library
LLM Evaluations
Evaluating and scoring financial data
Python Mathematical Expression Evaluator
Limited evaluator
A library for providing a simple interface to create new metrics and an easy-to-use toolkit for metric computations and checkpointing.
Provides Python bindings for popular Information Retrieval measures implemented within trec_eval.
Universal library for evaluating AI models
A library for providing a simple interface to create new metrics and an easy-to-use toolkit for metric computations and checkpointing.
Faster interpretation of the original COCOEval
Contains the integration code of AzureML Evaluate with Mlflow.
AlpacaEval : An Automatic Evaluator of Instruction-following Models
Backwards-compatibility package for API of trulens_eval<1.0.0 using API of trulens-*>=1.0.0.
Python package for evaluating neuron segmentations in terms of the number of splits and merges
A custom Streamlit component to evaluate arbitrary Javascript expressions.
Provides Python bindings for popular Information Retrieval measures implemented within trec_eval.
An information retrieval evaluation script based on the C/W/L framework that is TREC Compatible and provides a replacement for INST_EVAL, RBP_EVAL, TBG_EVAL, UMeasure and TREC_EVAL scripts. All measurements are reported in the same units making all metrics directly comparable.
User-friendly evaluation framework: Eval Suite & Benchmarks: UHGEval, HaluEval, HalluQA, etc.
A flexible, generalized tree-based data structure.
Send Sir Perceval on a quest to fetch and gather data from software repositories.
Performs async evaluations of strings
Prompt flow evals
Easily computing clip embeddings and building a clip retrieval system with them
Bundle of Perceval backends for Mozilla ecosystem.
Interface to ndeval.c
Python code evaluation system and submissions server capable of unit tests, tracing, and AST inspection. Server can run on Python 2.7 but evaluation requires 3.7+.
Bundle of Perceval backends for OPNFV ecosystem.
Serialization based on ast.literal_eval
Package for fast computation of BSS Eval metrics for source separation