Spaces:
Running
Running
| # ML.ENERGY Leaderboard | |
| How much energy do LLMs consume? | |
| ## Devs | |
| Current setup in `ampere02`: | |
| 1. Find model weights in `/data/leaderboard/weights/`, e.g. subdirectory `llama` and `vicuna`. | |
| 2. Let's share the Huggingface Transformer cache: | |
| ```bash | |
| export TRANSFORMERS_CACHE=/data/leaderboard/hfcache | |
| ``` | |
| Run benchmarks like this: | |
| ```console | |
| $ docker build -t leaderboard:latest . | |
| $ docker run -it --name jw-leaderboard --gpus all --cap-add SYS_ADMIN -v /data/leaderboard:/data/leaderboard -v $HOME/workspace/leaderboard:/workspace/leaderboard leaderboard:latest bash | |
| # cd leaderboard | |
| # python scripts/benchmark.py --model-path /data/leaderboard/weights/lmsys/vicuna-7B --input-file /data/leaderboard/sharegpt/sg_90k_part1_html_cleaned_lang_first_sampled.json | |
| # python scripts/benchmark.py --model-path databricks/dolly-v2-12b --input-file /data/leaderboard/sharegpt/sg_90k_part1_html_cleaned_lang_first_sampled.json | |
| ``` | |