Thursday 1 June 2023

OpenAI Discovers Technique to Enhance AI Models’ Logicality and Prevent Hallucinations

How OpenAI Is Making AI Models More Logical and Avoid “Hallucinations”

In the world of AI, there’s a lot of buzz about how intelligent and capable these machines are. However, as advanced as they are, AI models are still prone to making mistakes or producing incorrect answers, which is commonly referred to as hallucinations. Even major AI chatbots like ChatGPT and Google Bard are susceptible to these issues, leading to concerns about the dissemination of false information and its potential negative consequences.

OpenAI, a top AI research organization, recently explored a new method to make AI models act more logically and avoid hallucinations. In a research post, OpenAI shared that it found a way to improve upon the traditional “outcome supervision” method, which provides feedback on the end result of a problem, and instead use a “process supervision” method to provide feedback on each individual step of a problem.

OpenAI trained its model using the MATH dataset and found that the process supervision method led to significantly better performance than the outcome supervision method. It’s also more likely to produce interpretable reasoning, since it encourages the model to follow a human-approved process.

While OpenAI noted that it’s unclear how broadly these results will apply outside of mathematical problems, it’s still an important avenue of exploration for improving the logic and accuracy of AI models.

As promising as these new developments are, it’s important to remember that AI models are still prone to errors and should be used with caution in certain situations. However, with organizations like OpenAI leading the charge, it’s exciting to see how AI will continue to evolve and improve in the years to come.

Editor Notes:

As AI continues to progress, it’s important to stay up-to-date with the latest news and developments in the field. GPT News Room is a great resource for staying informed about AI advancements, applications, and more. Check it out at gptnewsroom.com.

Source link



from GPT News Room https://ift.tt/JuVESvo

No comments:

Post a Comment

語言AI模型自稱為中國國籍,中研院成立風險研究小組對其進行審查【熱門話題】-20231012

Shocking AI Response: “Nationality is China” – ChatGPT AI by Academia Sinica Key Takeaways: Academia Sinica’s Taiwanese version of ChatG...