Skip to content

somosnlp/la-leaderboard-backend

Folders and files

NameName
Last commit message
Last commit date

Latest commit

bc0f728 · Dec 7, 2024

History

27 Commits
Nov 6, 2024
Dec 5, 2024
Oct 30, 2024
Oct 22, 2024
Dec 7, 2024
Oct 22, 2024
Dec 6, 2024
Nov 6, 2024
Oct 23, 2024
Dec 4, 2024
Oct 30, 2024
Dec 7, 2024
Dec 7, 2024

Repository files navigation

Backend de "La Leaderboard"

  • To evaluate the models in the requests dataset, run python3 -m main_eval_basic_queue.py
  • To evaluate a custom combination of models and tasks, run python3 -m main_eval_internal_queue.py with an optional argument containing the path to the JSON file with the tasks to run which defaults to "internal_queue/tasks_todo.json"
  • To evaluate the models in the cluster's queue managed by slurm, run python3 -m main_eval_slurm_queue.py

Notes:

  • Check the version of the lm-evaluation-harness used in the requirements

About

No description, website, or topics provided.

Resources

Code of conduct

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published