autoevals
Advanced tools
Comparing version 0.0.21 to 0.0.22
{ | ||
"name": "autoevals", | ||
"version": "0.0.21", | ||
"version": "0.0.22", | ||
"description": "Universal library for evaluating AI models", | ||
@@ -5,0 +5,0 @@ "main": "jsdist/bundle.js", |
@@ -151,3 +151,3 @@ # AutoEvals | ||
- Normalizing metrics between 0 and 1 is tough. For example, check out the calculation in [number.py][/py/autoevals/number.py] to see how it's done for numeric differences. | ||
- Normalizing metrics between 0 and 1 is tough. For example, check out the calculation in [number.py](/py/autoevals/number.py) to see how it's done for numeric differences. | ||
- Parsing the outputs on model-graded evaluations is also challenging. There are frameworks that do this, but it's hard to | ||
@@ -154,0 +154,0 @@ debug one output at a time, propagate errors, and tweak the prompts. AutoEvals makes these tasks easy. |
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package
License Policy Violation
LicenseThis package is not allowed per your license policy. Review the package's license to ensure compliance.
Found 1 instance in 1 package