Quantization
#16 opened 6 months ago
by
jntjdbhvebjynfbjdv
Add exported onnx model 'model_qint8_arm64.onnx'
#15 opened 6 months ago
by
tonatiuh-synopsys
Matryoshka embeddings
1
#14 opened 9 months ago
by
ghidav

model doesnt seem to support device_map="auto" for multi GPU .
2
#11 opened about 1 year ago
by
pulkitmehtametacube
How to change the batch size? or any tips to make the model faster?
1
#7 opened about 1 year ago
by
sdvfsfwfgfwf
Help with error message "Could not locate the configuration.py inside Alibaba-NLP/new-impl."
1
#6 opened about 1 year ago
by
azhang01