OpenAI admits AI hallucinations can't be fixed

Draugoth

Gold Member
bafkreiaian6pxyvxafqbcry2wbf26boddcqw4k426dfa6sqj6ji63riocm@jpeg


Source


In a landmark study, OpenAI researchers reveal that large language models will always produce plausible but false outputs, even with perfect data, due to fundamental statistical and computational limits.

"Large language models sometimes guess when uncertain, producing plausible yet incorrect statements instead of admitting uncertainty. Such 'hallucinations' persist even in state-of-the-art systems."

"The study established that "the generative error rate is at least twice the IIV misclassification rate," where IIV referred to "Is-It-Valid" and demonstrated mathematical lower bounds that prove AI systems will always make a certain percentage of mistakes, no matter how much the technology improves."

"The OpenAI research also revealed that industry evaluation methods actively encouraged the problem. Analysis of popular benchmarks, including GPQA, MMLU-Pro, and SWE-bench, found nine out of 10 major evaluations used binary grading that penalized "I don't know" responses while rewarding incorrect but confident answers."
 
Last edited:
I love what AI is doing in the medical fields, they can check for blood cancers via eye tests now but you always need a person checking it to make sure it's not producing utter bollocks.

Sadly those pushing it hard seem to think it's magic and not easily derailed.
 
Where one company has flaws, another will solve. Only ever a matter of time, so it doesn't even matter. It's just something we should know when using it for now.
 
Top Bottom