Yosef Worku Alemneh
rasyosef
AI & ML interests
Pretraining, Supervised Fine Tuning, Direct Preference Optimization, Retrieval Augmented Generation (RAG), Function Calling
Recent Activity
updated
a collection
6 days ago
Phi 1.5 Chat Models
new activity
19 days ago
rasyosef/flores_english_amharic_mt:Update README.md to include ISO code for English
Organizations
rasyosef's activity
Update README.md to include ISO code for English
#2 opened 4 months ago
by
weezygeezer
Update model metadata to set pipeline tag to the new `text-ranking` and tags to `sentence-transformers`
#1 opened about 2 months ago
by
tomaarsen

Using hard negatives VS query, pos pair to train embedding models
4
#2 opened 3 months ago
by
rasyosef
Adding Evaluation Results
#1 opened 8 months ago
by
leaderboard-pr-bot

Adding Evaluation Results
#3 opened 9 months ago
by
leaderboard-pr-bot

Phi-2-Instruct-APO: aligned with Anchored Preference Optimization
16
#3 opened 8 months ago
by
rasyosef
[Query-ISSUE] tokenizer.vocab_size is 128000, however len(tokenizer) is 128256, which prevents me from using those other tokens.
1
#34 opened 7 months ago
by
HV-Khurdula

What are the start and stop tokens of this model?
1
#40 opened 7 months ago
by
aryaash
Is the BOS token id of 128000 hardcoded into the llama 3.2 tokenizer?
2
#17 opened 8 months ago
by
rasyosef
Mistral-NeMo-Minitron-8B-Chat
🚀
1
5
#5 opened 9 months ago
by
rasyosef
APO Trainer in TRL?
1
#2 opened 9 months ago
by
rasyosef
ChatML template does not work properly
10
#2 opened 9 months ago
by
WasamiKirua

Collaboration
1
#1 opened 9 months ago
by
deleted
Error when trying to run
1
#1 opened 9 months ago
by
ctranslate2-4you
What changed for people using this model in english?
👍
1
3
#3 opened 9 months ago
by
migueltalka
What should a finetuned model's license be if the model is MIT but the datasets are Apache 2.0 and cc-by-4.0
5
#866 opened 10 months ago
by
rasyosef
Update README.md
1
#2 opened 10 months ago
by
seyyaw

Duplicate?
❤️
1
1
#2 opened about 1 year ago
by
israel
