Finetuning model performance decreases when using memory_efficient_attenttion
#11 opened 2 months ago
by
hrushikesh1
Any Plan to release tensorflow based model?
1
#10 opened 5 months ago
by
undefined-x
adding to transformers officially?
2
#9 opened 5 months ago
by
pszemraj

Is flash-attention-2 suppported
1
#8 opened 7 months ago
by
Jack7777777
Xformer support for Qwen1.5B
3
#6 opened 8 months ago
by
le723z
backbone模型不开源吗?
10
#4 opened 11 months ago
by
JaheimLee
Disable trust_remote_code
4
14
#2 opened 12 months ago
by
veeravignesh