Submitted by imgonnarelph t3_11wqmga in MachineLearning
hosjiu t1_jd1a6az wrote
Reply to comment by Civil_Collection7267 in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
"They also have the tendency to hallucinate frequently unless parameters are made more restrictive."
I am not really understand this point in term of technical
royalemate357 t1_jd1stda wrote
Not op, but I imagine they're referring to the sampling hyperparameters that control the text generation process. For example there is a temperature setting, a lower temperature makes it sample more from the most likely choices. So it would potentially be more precise/accurate but also less diverse and creative in it's outputs
Viewing a single comment thread. View all comments