Spaces:
Runtime error
Runtime error
metadata
title: Zeno Evals Hub
emoji: π
colorFrom: pink
colorTo: indigo
sdk: docker
pinned: false
license: mit
fullWidth: true
Zeno + OpenAI Evals
OpenAI's Evals library is a great resource providing evaluation sets for LLMS.
This repo provides a hub for exploring these results using the Zeno evaluation tool.
Add New Evals
To add new evals, add a new entry to evals/evals.yaml
with the following fields:
results-file
: The first.jsonl
result fromoaievals
link
: A link to the evals commit for this evaluationdescription
: A succint description of what the evaluation is testingsecond-results-file
: An optional second.jsonl
result fromoaievals
. Must be the same dataset as the first one.functions-file
: An optional Python file with Zeno functions for the evaluations.
Make sure you test your evals locally before submitting a PR!
Running
poetry install
python -m zeno-evals-hub evals/evals.yaml