Post
3350
I think we have released the best Arabic model under 25B at least based on
inceptionai/AraGen-Leaderboard
Yehia = ALLaM-AI/ALLaM-7B-Instruct-preview + GRPO
and its ranked number one model under the 25B parameter size mark.
Now, i said "i think" not "i am sure" because this model used the same metric of evaluation the AraGen developers use (the 3C3H) as a reward model to improve its responses and this sparks the question. Is this something good for users or is it another type of overfitting that we don't want ?
I don't know if this is a good thing or a bad thing but what i know is that you can try it from here:
Navid-AI/Yehia-7B-preview
or Download it for your personal experiments from here:
Navid-AI/Yehia-7B-preview
Ramadan Kareem 🌙
Yehia = ALLaM-AI/ALLaM-7B-Instruct-preview + GRPO
and its ranked number one model under the 25B parameter size mark.
Now, i said "i think" not "i am sure" because this model used the same metric of evaluation the AraGen developers use (the 3C3H) as a reward model to improve its responses and this sparks the question. Is this something good for users or is it another type of overfitting that we don't want ?
I don't know if this is a good thing or a bad thing but what i know is that you can try it from here:
Navid-AI/Yehia-7B-preview
or Download it for your personal experiments from here:
Navid-AI/Yehia-7B-preview
Ramadan Kareem 🌙