Add note about repetition penalty to readme

#4
by adamo1139 - opened

During a discussion over at r/localllama, it became quite apparent that using repetition penalty settings different from 1.0 will very likely negatively affect performance of this model, since coding very often has repeating patterns.

Conversation in this thread.
https://old.reddit.com/r/LocalLLaMA/comments/17yda6k/having_a_hard_time_setting_deepseek_coder/

Would it be possible for you to add information that setting repetition penalty to 1.0 may improve coding performance of DeepSeek models to readme pages of your deepseek quants? That's seemingly how DeepSeek is running their model on their demo page, so I think this is how they intend it to be used. It is actually probably applicable to all coding models, but we noticed it for DeepSeek now. I think this could help out people who get bad first experience with this model and they are not sure why this hyped up thing runs poorly for them.

Sign up or log in to comment