hanglics commited on
Commit
4ada353
·
1 Parent(s): e2da048

bump flash-attn version

Browse files
Files changed (1) hide show
  1. requirements.txt +5 -4
requirements.txt CHANGED
@@ -1,6 +1,7 @@
1
  # Core ML dependencies for DSE-Qwen2-2B-MRL-V1
2
- torch==2.2.0
3
- torchvision
 
4
  transformers==4.55.2
5
  tokenizers>=0.15.0
6
  accelerate>=0.26.0
@@ -17,8 +18,8 @@ numpy<2.0.0
17
  # HuggingFace Hub integration
18
  huggingface-hub>=0.19.0
19
 
20
- # Flash Attention (optional, may not install on all systems)
21
- flash-attn==2.3.0
22
 
23
  # Additional utilities
24
  tqdm==4.66.1
 
1
  # Core ML dependencies for DSE-Qwen2-2B-MRL-V1
2
+ torch==2.6.0
3
+ torchvision==0.21.0
4
+ torchaudio==2.6.0
5
  transformers==4.55.2
6
  tokenizers>=0.15.0
7
  accelerate>=0.26.0
 
18
  # HuggingFace Hub integration
19
  huggingface-hub>=0.19.0
20
 
21
+ # Flash Attention (optional, commented out due to slow build times on HF Inference Endpoints)
22
+ flash-attn==2.7.4.post1
23
 
24
  # Additional utilities
25
  tqdm==4.66.1