Submitted by buggaby t3_11qgasm in MachineLearning
I just posted this on r/ChatGPT but thought there might be some great thoughts here, too.
ChatGPT generates believable output but, as many have noted, not trustworthy output. A lot of the use cases I see for future generative AI models seem to crucially depend on making believable AND truthful responses. But given that it's probably easier to make believable but non-truth responses (since more of them exist), I imagine that this is a very hard prospect. Is it even possible with current methods?
From my read, modern generative AI models can only increase correctness of output in 2 ways. Using more correct data, and using human labellers for fine-tuning. Having more correct data either requires much smaller datasets (even academic journals can't be considered correct since science evolves over time) or human expertise in correcting the data. So it seems like human expertise remains vital.
Now I know that human labellers were necessary to reduce the toxicity of GPT-3 responses. I read that something like dozens were used over a period of months, though I don't know if this is publicly shared by OpenAI. But how important is human training in driving up "truthfulness" of these models?
I briefly reviewed this paper and it talks about InstructGPT being better than GPT-3 at truthfulness, even with 1/100th of the parameters (1.3B parameters vs 175B of GPT-3). But I also understand that larger models tend to lie more, so that could be part of it. And even though it is "more truthful", the metric used to compare seems suspect to me, especially since "InstructGPT still makes simple mistakes", including making up facts.
It seems here like little improvement in truthfulness.
Without a clear path to increasing this vital metric, I struggle to see how modern generative AI models can be used for any important tasks that are sensitive to correctness. That's still a lot of cool things, but we seem far from even a good search engine, from assisting researchers, or even from coding support. (I have used ChatGPT for this latter purpose, and sometimes it helps me more quickly, but sometimes it makes it slower because it's flat-out false. Stackoverflow is generally much more trustworthy and useful for me so far.) And certainly we are really far from anything remotely "AGI".
abriec t1_jc34zx3 wrote
Given the constant evolution of information through time combining LLMs with retrieval and reasoning modules is the way forward imo