Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
Quazim0t0 
posted an update 12 days ago
Post
1086
Thank you to the Open LLM Leaderboard's team for offering it to the community for as long as they did. I only recently joined HF, and it provided a lot of incentive and information to make better models.

Always will remember getting to #112 :D

Anyone have a solid way to test my models privately? Please let me know!

you can use lm-evaluation-harness from Eleuther AI, tho its a bit slow from my testing.
Alternatively, you can use hf evals to match the scores from the public leaderboard.

Side note: The hf llm leaderboard seems to be outdated a bit, so to use the new and better benchmarks, I suggest evaluating locally.

The links:
https://github.com/EleutherAI/lm-evaluation-harness
https://huggingface.co/docs/evaluate/

In this post