< Explain other AI papers

Sotopia-RL: Reward Design for Social Intelligence

Haofei Yu, Zhengyang Qi, Yining Zhao, Kolby Nottingham, Keyang Xuan, Bodhisattwa Prasad Majumder, Hao Zhu, Paul Pu Liang, Jiaxuan You

2025-08-07

Sotopia-RL: Reward Design for Social Intelligence

Summary

This paper talks about Sotopia-RL, a new reinforcement learning method that helps large language models become better at social tasks by giving detailed feedback on every part of a conversation, not just the whole interaction.

What's the problem?

The problem is that teaching AI models to be socially smart is hard because social interactions are complex and unfold over many steps. Usually, the feedback AI models get is too general and simple, which makes it difficult for them to learn the important details of social behavior like building relationships or sharing knowledge.

What's the solution?

The solution was to create Sotopia-RL, which breaks down feedback into rewards for each individual sentence a model says during a conversation. It also uses multiple types of rewards that focus on different social goals, such as keeping good relationships, seeking knowledge, and completing tasks. This helps the model learn social skills more effectively by understanding the impact of each part of the conversation.

Why it matters?

This matters because better social intelligence in AI can improve how machines interact with humans in everyday situations like negotiating, collaborating, or persuading. Sotopia-RL makes AI models more skilled in understanding and responding to social cues, which can lead to more natural and helpful interactions with technology.

Abstract

Sotopia-RL, a novel reinforcement learning framework, enhances social intelligence in large language models by refining feedback into utterance-level, multi-dimensional rewards, improving performance in social tasks.