Hardware & VRAM Requirement for Llama 4 Scout in BF16

#66
by YongchengYAO - opened

I have access to Nvidia 80G A100 and H100 GPUs.

  • Is it possible to run inference for the meta-llama/Llama-4-Scout-17B-16E-Instruct model in BF16?
  • What is the VRAM requirement?

Found the instruction: https://github.com/meta-llama/llama-cookbook/blob/main/getting-started/build_with_llama_4.ipynb

"You'll need at least 4 GPUs with >= 80GB each"

YongchengYAO changed discussion status to closed

Sign up or log in to comment