Replies: 2 comments 1 reply
-
Seconded. Actually, on a per response basis - add to that, the ability to generate 2 or more responses to the same query, in paralleled, the way Gemini does it, so that you can choose among them. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Newcomer question: how do you even set these settings in general? The docs state the fact we can teak them, but... how? I cannot find any GUI section that let me tweak temperature, sampling, etc.. am I blind? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi, I was thinking, since the settings we use (like temperature, top_k, top_p, etc.) really depend on what we're trying to do, it would be helpful to be able to tweak them on a per-conversation basis. In a similar fashion as how Lobe-chat lets you adjust the temperature for each new chat. For example, when I'm coding, I need the temperature to be pretty low so I don't get many mistakes, but that's not ideal for other use cases where I want the AI to be more creative.
Implementation of:
Beta Was this translation helpful? Give feedback.
All reactions