Every Eval Ever is a shared schema and crowdsourced eval database. It defines a standardized metadata format for storing AI evaluation results — from leaderboard scrapes and research papers to local evaluation runs — so that results from different frameworks can be compared, reproduced, and reused.
Stars
50
Forks
24
Watchers
50
Open Issues
10
Overall repository health assessment
No package.json found
This might not be a Node.js project
85
commits
38
commits
24
commits
15
commits
14
commits
10
commits
6
commits
5
commits
4
commits
4
commits
refactor: remove unused save_to_file from inspect/utils (#100)
899a12eView on GitHubfix: preserve falsy generation args in HELM adapter (#102)
8e8833cView on GitHubMerge pull request #90 from evaleval/fix_different_uuids_for_json_and_jsonl
411f60bView on GitHubEliminate extraction uuid with regex and pass it via metadata_args from cli
c577a4fView on GitHubRemove code duplication for uuid extraction function
2725a2dView on GitHubUse regex for retrieving uuid from instance-level info file
6a46243View on GitHubFix different uuids for json and jsonl files for the same log in the CLI
016f851View on GitHubRead sample scores from sample reductions for inspect eval logs (#89)
332304eView on GitHubAdded option to provide additional eval details as json file (#66)
6dbbd54View on GitHubMerge pull request #81 from elronbandel/exgentic-adapter
5c5e43dView on GitHubMerge pull request #63 from StevenDillmann/submission/terminal-bench-2.0
5074410View on GitHub