Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

RedHatAI
/
Llama-4-Scout-17B-16E-Instruct-FP8-dynamic

Image-Text-to-Text
Safetensors
PyTorch
vllm
llama4
facebook
meta
llama
neuralmagic
redhat
llmcompressor
quantized
FP8
conversational
compressed-tensors
Model card Files Files and versions
xet
Community
5
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

Failing to quantize using your method

#4 opened 24 days ago by
redd2dead

VLLM launch parametrs

๐Ÿ‘ 3
#3 opened about 2 months ago by
Clutchkin

Why not FP8 with static and per-tensor quantization?

๐Ÿ‘ 1
1
#2 opened about 2 months ago by
wanzhenchn

Thank you uploading this.

โค๏ธ 6
#1 opened about 2 months ago by
getfit
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs