OpenAI announced on May 31st, its efforts to enhance ChatGPT's mathematical problem-solving capabilities, aiming to reduce instances of artificial intelligence (AI) hallucinations. OpenAI emphasized mitigating hallucinations as a crucial step towards developing aligned AGI.
In March, the introduction of the latest version of ChatGPT, GPT-4, further propelled artificial intelligence into the mainstream. However, generative AI chatbots have long grappled with factual accuracy, occasionally generating false information, commonly referred to as "hallucinations." The efforts to reduce these AI hallucinations were announced through a post on their website.
AI hallucinations refer to instances where artificial intelligence systems generate outputs that are factually incorrect, misleading or unsupported by real-world data. These hallucinations can manifest in various forms, such as generating false information, making up nonexistent events or people or providing inaccurate details about certain topics.
OpenAI conducted research to examine the effectiveness of two types of feedback– "outcome supervision" and "process supervision." Outcome supervision involves feedback based on the final result, while process supervision provides input for each step in a chain of thought. OpenAI evaluated these models using math problems, generating multiple solutions and selecting the highest-ranked solution according to each feedback model.
After thorough analysis, the research team found that process supervision yielded a superior performance as it encouraged the model to adhere to a human-approved process. In contrast, outcome supervision proved more challenging to scrutinize consistently.
OpenAI recognized that the implications of process
Read more on cointelegraph.com