Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Guide on how to evaluate models #180

Open
kisimoff opened this issue Apr 3, 2024 · 1 comment
Open

Guide on how to evaluate models #180

kisimoff opened this issue Apr 3, 2024 · 1 comment

Comments

@kisimoff
Copy link

kisimoff commented Apr 3, 2024

Im willing to test a few models and share the results.
I've looked at the readme, but couldn't wrap my head around how to benchmark a model. Any help would be appriciated!

@the-crypt-keeper
Copy link
Owner

The docs definitely need a rewrite my apologies here.

The general flow is:

  1. prepare.py
  2. interview*.py
  3. eval.py

In the dark days we had to deal with dozens of prompt formats, but these days prepare.py can be run with --chat hfmodel and it will sort it out.

Note that there are two interviews junior-v2 and senior, I usually only run senior on strong models that get >90% on junior.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants