Please enable JavaScript to experience the full functionality of GMX.

OpenAI unveils CriticGPT to find and solve code errors

OpenAI unveils CriticGPT to find and solve code errors

OpenAI has unveiled CriticGPT to assist developers in identifying and correcting errors in code generated by ChatGPT.

The model is based on GPT-4 and they claim it's proven that those who use it "outperform those who don't get the help by "60 per cent".

Overall, it will make the AI more reliable.

In a recent blog post, the firm said: "We've trained a model, based on GPT-4, called CriticGPT to catch errors in ChatGPT's code output. We found that when people get help from CriticGPT to review ChatGPT code they outperform those without help 60 per cent of the time. We are beginning the work to integrate CriticGPT-like models into our RLHF labeling pipeline, providing our trainers with explicit AI assistance. This is a step towards being able to evaluate outputs from advanced AI systems that can be difficult for people to rate without better tools.

"The GPT-4 series of models, which powers ChatGPT, is aligned to be helpful and interactive through “Reinforcement Learning from Human Feedback” (RLHF). A key part of RLHF is collecting comparisons in which people, called AI trainers, rate different ChatGPT responses against each other.

"As we make advances in reasoning and model behavior, ChatGPT becomes more accurate and its mistakes become more subtle. This can make it hard for AI trainers to spot inaccuracies when they do occur, making the comparison task that powers RLHF much harder. This is a fundamental limitation of RLHF, and it may make it increasingly difficult to align models as they gradually become more knowledgeable than any person that could provide feedback. (sic)"

Sponsored Content

Related Headlines