Submitted by mettle t3_10oyllu in MachineLearning
currentscurrents t1_j6jbokk wrote
I think hallucination occurs because of the next-word-prediction task on which these models were trained. No matter how good a model is, it can never predict the irreducible entropy of the sentence - the 1.5 bits per word or whatever that contains the actual information content. The best it can do is guess.
This is exactly what hallucination looks like; all the sentence structure is right, but the information is wrong. Unfortunately, this is also the most important part of the sentence.
mettle OP t1_j6jgkz8 wrote
Sure, but the question is how often does it happen to get the right answer vs. the wrong answer and how would be measure that.
Viewing a single comment thread. View all comments