search
HomeTechnology peripheralsAIGoogle proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training

The effect is more stable and the implementation is simpler.

The success of large language models (LLM) is inseparable from "reinforcement learning based on human feedback (RLHF)". RLHF can be roughly divided into two stages. First, given a pair of preferred and unpreferred behaviors, a reward model is trained to assign a higher score to the former by classifying the target. This reward function is then optimized through some kind of reinforcement learning algorithm. However, key elements of the reward model may have some undesirable effects.

Researchers from Carnegie Mellon University (CMU) and Google Research jointly proposed a simple, theoretically rigorous, and experimentally effective new RLHF Method - Self-Play Preference Optimization (SPO). This approach eliminates reward models and does not require adversarial training.

Google proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training

Paper: A Minimaximalist Approach to Reinforcement Learning from Human Feedback
Paper address: https://arxiv.org/abs/2401.04056

Method Introduction

The SPO method mainly includes two aspects. First, this research truly eliminates the reward model by constructing RLHF as a zero-sum game, making it more capable of handling noisy, non-Markovian preferences that often appear in practice. Second, by exploiting the symmetry of the game, this study demonstrates that a single agent can simply be trained in a self-game manner, thereby eliminating the need for unstable adversarial training.

In practice, this is equivalent to sampling multiple trajectories from the agent, asking the evaluator or preference model to compare each pair of trajectories, and setting the reward to the winning rate of the trajectory .
Google proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training
#SPO avoids reward modeling, compound errors, and adversarial training. By establishing the concept of minmax winner from social choice theory, this study constructs RLHF as a two-person zero-sum game and exploits the symmetry of the game's payoff matrix to demonstrate that a single agent can be simply trained to against itself.

Google proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training

Google proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training

The study also analyzed the convergence characteristics of SPO and proved that when the potential reward function does exist, SPO can be compared with The fast rate of convergence to the optimal policy is comparable to standard methods.

Experiment

This study performed a series of continuous control tasks with realistic preference functions On the above, it is proved that SPO performs better than methods based on reward models. SPO is able to learn samples more efficiently than reward model-based methods in various preference settings, as shown in Figure 2 below.

Google proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training

Google proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training

Google proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training

This study combines SPO with the iterative reward modeling (RM) method from multiple dimensions A comparison is made to answer 4 questions:

  1. Can SPO calculate MW when facing intransitive preferences?
  2. Can SPO match or exceed RM sample efficiency on problems with unique Copeland Winners/optimal strategies?
  3. How robust is SPO to random preferences?
  4. Can SPO handle non-Markovian preferences?

Google proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training

In terms of maximum reward preference, noise preference, and non-Markov preference, the experimental results of this study are shown in Figures 6, 7, and 8 below respectively. Show:

Google proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training

Google proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training

Google proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training

Interested readers can read the original text of the paper to learn more about the research content.

The above is the detailed content of Google proposes a new RLHF method: eliminating reward models and eliminating the need for adversarial training. For more information, please follow other related articles on the PHP Chinese website!

Statement
This article is reproduced at:机器之心. If there is any infringement, please contact admin@php.cn delete
Sam's Club Bets On AI To Eliminate Receipt Checks And Enhance RetailSam's Club Bets On AI To Eliminate Receipt Checks And Enhance RetailApr 22, 2025 am 11:29 AM

Revolutionizing the Checkout Experience Sam's Club's innovative "Just Go" system builds on its existing AI-powered "Scan & Go" technology, allowing members to scan purchases via the Sam's Club app during their shopping trip.

Nvidia's AI Omniverse Expands At GTC 2025Nvidia's AI Omniverse Expands At GTC 2025Apr 22, 2025 am 11:28 AM

Nvidia's Enhanced Predictability and New Product Lineup at GTC 2025 Nvidia, a key player in AI infrastructure, is focusing on increased predictability for its clients. This involves consistent product delivery, meeting performance expectations, and

Exploring the Capabilities of Google's Gemma 2 ModelsExploring the Capabilities of Google's Gemma 2 ModelsApr 22, 2025 am 11:26 AM

Google's Gemma 2: A Powerful, Efficient Language Model Google's Gemma family of language models, celebrated for efficiency and performance, has expanded with the arrival of Gemma 2. This latest release comprises two models: a 27-billion parameter ver

The Next Wave of GenAI: Perspectives with Dr. Kirk Borne - Analytics VidhyaThe Next Wave of GenAI: Perspectives with Dr. Kirk Borne - Analytics VidhyaApr 22, 2025 am 11:21 AM

This Leading with Data episode features Dr. Kirk Borne, a leading data scientist, astrophysicist, and TEDx speaker. A renowned expert in big data, AI, and machine learning, Dr. Borne offers invaluable insights into the current state and future traje

AI For Runners And Athletes: We're Making Excellent ProgressAI For Runners And Athletes: We're Making Excellent ProgressApr 22, 2025 am 11:12 AM

There were some very insightful perspectives in this speech—background information about engineering that showed us why artificial intelligence is so good at supporting people’s physical exercise. I will outline a core idea from each contributor’s perspective to demonstrate three design aspects that are an important part of our exploration of the application of artificial intelligence in sports. Edge devices and raw personal data This idea about artificial intelligence actually contains two components—one related to where we place large language models and the other is related to the differences between our human language and the language that our vital signs “express” when measured in real time. Alexander Amini knows a lot about running and tennis, but he still

Jamie Engstrom On Technology, Talent And Transformation At CaterpillarJamie Engstrom On Technology, Talent And Transformation At CaterpillarApr 22, 2025 am 11:10 AM

Caterpillar's Chief Information Officer and Senior Vice President of IT, Jamie Engstrom, leads a global team of over 2,200 IT professionals across 28 countries. With 26 years at Caterpillar, including four and a half years in her current role, Engst

New Google Photos Update Makes Any Photo Pop With Ultra HDR QualityNew Google Photos Update Makes Any Photo Pop With Ultra HDR QualityApr 22, 2025 am 11:09 AM

Google Photos' New Ultra HDR Tool: A Quick Guide Enhance your photos with Google Photos' new Ultra HDR tool, transforming standard images into vibrant, high-dynamic-range masterpieces. Ideal for social media, this tool boosts the impact of any photo,

What are the TCL Commands in SQL? - Analytics VidhyaWhat are the TCL Commands in SQL? - Analytics VidhyaApr 22, 2025 am 11:07 AM

Introduction Transaction Control Language (TCL) commands are essential in SQL for managing changes made by Data Manipulation Language (DML) statements. These commands allow database administrators and users to control transaction processes, thereby

See all articles

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

Video Face Swap

Video Face Swap

Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Tools

SublimeText3 English version

SublimeText3 English version

Recommended: Win version, supports code prompts!

mPDF

mPDF

mPDF is a PHP library that can generate PDF files from UTF-8 encoded HTML. The original author, Ian Back, wrote mPDF to output PDF files "on the fly" from his website and handle different languages. It is slower than original scripts like HTML2FPDF and produces larger files when using Unicode fonts, but supports CSS styles etc. and has a lot of enhancements. Supports almost all languages, including RTL (Arabic and Hebrew) and CJK (Chinese, Japanese and Korean). Supports nested block-level elements (such as P, DIV),

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)

MinGW - Minimalist GNU for Windows

MinGW - Minimalist GNU for Windows

This project is in the process of being migrated to osdn.net/projects/mingw, you can continue to follow us there. MinGW: A native Windows port of the GNU Compiler Collection (GCC), freely distributable import libraries and header files for building native Windows applications; includes extensions to the MSVC runtime to support C99 functionality. All MinGW software can run on 64-bit Windows platforms.

Atom editor mac version download

Atom editor mac version download

The most popular open source editor