Is there somewhere more detailed insight on arguments and their influence on the output? #620
Replies: 2 comments
-
found this |
Beta Was this translation helpful? Give feedback.
-
I did some digging on the Notably, this entropy is not derived from the softmaxed token probabilities -- that's what (the Shannon entropy), where The behavior is that if the last 32 tokens fall below the given To get a sense of the scale, I used ChatGPT to calculate the entropy you get if the last 32 tokens are a sentence of ![]() As you can see from the plot, the highest entropy you can get with all-distinct tokens is 3.47. To suppress repetitions more aggressively, you can increase the threshold from the default of 2.40, but you probably don't want to get too close to the maximum of 3.47, or you'll get a lot of retries and unnecessarily increase the temperature. In some very brief testing I found that 3.0 seems to do a good job getting rid of repetitions, and looking at the transcript diff, I can't make out much of a decrease in quality. |
Beta Was this translation helpful? Give feedback.
-
Is there any more in depth info on how to adjust the arguments for best results? More specifically i mean these arguments, for us, who are not at home with ML, please?
I was playing with some random adjustments but see no effect in the output.
Beta Was this translation helpful? Give feedback.
All reactions