AlexeyKruglov t1_j04rrg5 wrote on December 14, 2022 at 1:40 AM Reply to [D] Why are ChatGPT's initial responses so unrepresentative of the distribution of possibilities that its training data surely offers? by Osemwaro Probably because temperature parameter is not 1.0 when the model samples next tokens. Having it above 1 leads to the bias towards the more probable tokens. Permalink 2
AlexeyKruglov t1_j04rrg5 wrote
Reply to [D] Why are ChatGPT's initial responses so unrepresentative of the distribution of possibilities that its training data surely offers? by Osemwaro
Probably because temperature parameter is not 1.0 when the model samples next tokens. Having it above 1 leads to the bias towards the more probable tokens.