Optimal settings for running Small-24b using Ollama?
#14
by
AaronFeng753
- opened
This model seems to be quite sensitive to settings. I've seen many posts and comments on Reddit about this model having weird bugs due to inference settings. Could you share the optimal settings for running this model?
For example, what should the repeat penalty, top p & k values be? And is the optimal temperature indeed 0.15 as stated in the model card?