torch==2.6.0 llama-cpp-python gradio>=3.0.0 huggingface_hub transformers==4.49.0 accelerate==1.3.0 #flash_attn==2.7.4.post1