bump flash-attn version
Browse files- requirements.txt +5 -4
requirements.txt
CHANGED
|
@@ -1,6 +1,7 @@
|
|
| 1 |
# Core ML dependencies for DSE-Qwen2-2B-MRL-V1
|
| 2 |
-
torch==2.
|
| 3 |
-
torchvision
|
|
|
|
| 4 |
transformers==4.55.2
|
| 5 |
tokenizers>=0.15.0
|
| 6 |
accelerate>=0.26.0
|
|
@@ -17,8 +18,8 @@ numpy<2.0.0
|
|
| 17 |
# HuggingFace Hub integration
|
| 18 |
huggingface-hub>=0.19.0
|
| 19 |
|
| 20 |
-
# Flash Attention (optional,
|
| 21 |
-
flash-attn==2.
|
| 22 |
|
| 23 |
# Additional utilities
|
| 24 |
tqdm==4.66.1
|
|
|
|
| 1 |
# Core ML dependencies for DSE-Qwen2-2B-MRL-V1
|
| 2 |
+
torch==2.6.0
|
| 3 |
+
torchvision==0.21.0
|
| 4 |
+
torchaudio==2.6.0
|
| 5 |
transformers==4.55.2
|
| 6 |
tokenizers>=0.15.0
|
| 7 |
accelerate>=0.26.0
|
|
|
|
| 18 |
# HuggingFace Hub integration
|
| 19 |
huggingface-hub>=0.19.0
|
| 20 |
|
| 21 |
+
# Flash Attention (optional, commented out due to slow build times on HF Inference Endpoints)
|
| 22 |
+
flash-attn==2.7.4.post1
|
| 23 |
|
| 24 |
# Additional utilities
|
| 25 |
tqdm==4.66.1
|