Optimal settings for running Small-24b using Ollama?

#14
by AaronFeng753 - opened

This model seems to be quite sensitive to settings. I've seen many posts and comments on Reddit about this model having weird bugs due to inference settings. Could you share the optimal settings for running this model?

For example, what should the repeat penalty, top p & k values be? And is the optimal temperature indeed 0.15 as stated in the model card?

Sign up or log in to comment