LLM Settings

#1
by webboty - opened

Hi, thanks for the work.

What temperatur, top-p and top-k are recommended for this model?

I did not try anything other than defaults.

As recently reported, this quant might have issues with tool calls, so use it for chat only and should be fine--it was optimized for chat interaction. I will post in a bit a new quant of the VCoder with just 8-hi, that seems to be cleaner, and the size does not change by much

Temp=1 seems to be most efficient, from what I heard

Sign up or log in to comment