Vikhr-Nemo fine-tuned with contrastive Russian literature.
- Base model: https://huggingface.co/Vikhrmodels/Vikhr-Nemo-12B-Instruct-R-21-09-24
- Dataset: https://huggingface.co/datasets/40umov/dostoevsky
- Method: ORPO
- Training config: https://github.com/IlyaGusev/saiga/blob/main/configs/models/doestoevsky_nemo_12b_orpo_m1.json
- WandB: https://wandb.ai/ilyagusev/rulm_self_instruct/runs/4v4pcgej
- Downloads last month
- 14
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for IlyaGusev/vikhr_nemo_orpo_dostoevsky_12b
Base model
mistralai/Mistral-Nemo-Base-2407
Finetuned
mistralai/Mistral-Nemo-Instruct-2407