How much GPU memory does the 7B parameter model use?

#1
by SOT1k - opened

Hi team,

Thanks for releasing F Lite 7B, it looks like a very promising lightweight model!

I’d like to get a better idea of its hardware requirements. Could you share details on:

The approximate GPU memory usage when running inference (e.g. batch size 1)

Memory usage during training or fine-tuning, if applicable

Any optimizations (e.g., quantization, reduced precision) that affect memory footprint

This information would really help in planning deployments and evaluating compatibility with various hardware setups.

Thanks in advance!

Sign up or log in to comment