


To solve the biggest obstacle to the implementation of AI, has OpenAI found a way?
OpenAI seems to have found a solution to the "serious nonsense" of generative artificial intelligence.
On May 31, OpenAI announced on its official website that it had trained a model that can help eliminate common "illusions" and other common problems in generative AI.
OpenAI stated that reward models can be trained to detect hallucinations, and reward models are divided into outcome supervision (providing feedback based on the final result) or process supervision (providing feedback for each step in the thinking chain) model.
That is, process supervision rewards each correct step in reasoning, while outcome supervision simply rewards correct answers.
OpenAI said that in contrast, process supervision has an important advantage-It directly trains the model to produce human-approved thought chains:
Process supervision has several consistency advantages over outcome supervision. Each step is precisely supervised, so it rewards behavior that follows a consistent chain of thought model.
Process supervision is also more likely to produce explainable reasoning because it encourages the model to follow a human-approved process
Outcome monitoring can reward an inconsistent process and is often more difficult to review.
OpenAI tested both models on a mathematical dataset and found that the process supervision approach resulted in "significantly better performance."
It is important to note, however, that so far the process supervision approach has only been tested in the mathematical domain, and more work is needed to see how it performs more generally.
In addition, OpenAI did not indicate how long it will take for this research to be applied to ChatGPT, which is still in the research stage.
While the initial results are good, OpenAI does mention that the safer approach incurs reduced performance, called an alignment tax.
Current results show that process supervision does not generate alignment taxes when dealing with mathematical problems, but the situation in general information is not known.
The “illusion” of generative AI
Since the advent of generative AI, accusations of fabricating false information and “generating hallucinations” have never disappeared. This is also one of the biggest problems with current generative AI models.
In February of this year, Google hastily launched the chatbot Bard in response to ChatGPT funded by Microsoft. However, it was found that common sense errors were made in the demonstration, causing Google's stock price to plummet.
There are many reasons that cause AI hallucinations. One of them is inputting data to trick the AI program into misclassifying.
For example, developers use data (such as images, text, or other types) to train artificial intelligence systems. If the data is changed or distorted, the application will interpret the input differently and produce incorrect results.
Illusions can occur in large language-based models like ChatGPT due to incorrect transformer decoding, causing the language model to potentially produce a story or narrative that is not illogical or ambiguous.
The above is the detailed content of To solve the biggest obstacle to the implementation of AI, has OpenAI found a way?. For more information, please follow other related articles on the PHP Chinese website!

Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

Video Face Swap
Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Article

Hot Tools

Notepad++7.3.1
Easy-to-use and free code editor

SublimeText3 Chinese version
Chinese version, very easy to use

Zend Studio 13.0.1
Powerful PHP integrated development environment

Dreamweaver CS6
Visual web development tools

SublimeText3 Mac version
God-level code editing software (SublimeText3)

Hot Topics

The article reviews top AI art generators, discussing their features, suitability for creative projects, and value. It highlights Midjourney as the best value for professionals and recommends DALL-E 2 for high-quality, customizable art.

Meta's Llama 3.2: A Leap Forward in Multimodal and Mobile AI Meta recently unveiled Llama 3.2, a significant advancement in AI featuring powerful vision capabilities and lightweight text models optimized for mobile devices. Building on the success o

The article compares top AI chatbots like ChatGPT, Gemini, and Claude, focusing on their unique features, customization options, and performance in natural language processing and reliability.

ChatGPT 4 is currently available and widely used, demonstrating significant improvements in understanding context and generating coherent responses compared to its predecessors like ChatGPT 3.5. Future developments may include more personalized interactions and real-time data processing capabilities, further enhancing its potential for various applications.

The article discusses top AI writing assistants like Grammarly, Jasper, Copy.ai, Writesonic, and Rytr, focusing on their unique features for content creation. It argues that Jasper excels in SEO optimization, while AI tools help maintain tone consist

The article reviews top AI voice generators like Google Cloud, Amazon Polly, Microsoft Azure, IBM Watson, and Descript, focusing on their features, voice quality, and suitability for different needs.

2024 witnessed a shift from simply using LLMs for content generation to understanding their inner workings. This exploration led to the discovery of AI Agents – autonomous systems handling tasks and decisions with minimal human intervention. Buildin

This week's AI landscape: A whirlwind of advancements, ethical considerations, and regulatory debates. Major players like OpenAI, Google, Meta, and Microsoft have unleashed a torrent of updates, from groundbreaking new models to crucial shifts in le
