CoSER: Coordinating LLM-Based Persona Simulation of Established Roles
Abstract
Role-playing language agents (RPLAs) have emerged as promising applications of large language models (LLMs). However, simulating established characters presents a challenging task for RPLAs, due to the lack of authentic character datasets and nuanced evaluation methods using such data. In this paper, we present CoSER, a collection of a high-quality dataset, open models, and an evaluation protocol towards effective RPLAs of established characters. The CoSER dataset covers 17,966 characters from 771 renowned books. It provides authentic dialogues with real-world intricacies, as well as diverse data types such as conversation setups, character experiences and internal thoughts. Drawing from acting methodology, we introduce given-circumstance acting for training and evaluating role-playing LLMs, where LLMs sequentially portray multiple characters in book scenes. Using our dataset, we develop CoSER 8B and CoSER 70B, i.e., advanced open role-playing LLMs built on LLaMA-3.1 models. Extensive experiments demonstrate the value of the CoSER dataset for RPLA training, evaluation and retrieval. Moreover, CoSER 70B exhibits state-of-the-art performance surpassing or matching GPT-4o on our evaluation and three existing benchmarks, i.e., achieving 75.80% and 93.47% accuracy on the InCharacter and LifeChoice benchmarks respectively.
Community
๐ข Introducing CoSER: Advancing AI Character Role-Playing with High-Quality Data from Best-Ever Books
We're excited to present CoSER (Coordinating LLM-Based Persona Simulation of Established Roles), a collection of a high-quality dataset, open models, and novel evaluation protocol for more authentic AI character role-playing!
๐ Key Features:
๐ 17,966 characters, 29,798 authentic conversations, from 771 renowned books (top-rated on the Best-Books-Ever List)
๐ Comprehensive data of other types, such as conversation setups, character experiences and internal thoughts.
๐ State-of-the-art role-playing models: CoSER-8B and CoSER-70B (built upon LLaMA 3.1)
๐ญ Novel Methodology for training and evaluating role-playing LLMs: Given-Circumstance Acting.
๐ Paper: https://arxiv.org/pdf/2502.09082
๐ป Source code: https://github.com/Neph0s/COSER
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- OpenCharacter: Training Customizable Role-Playing LLMs with Large-Scale Synthetic Personas (2025)
- RPGBENCH: Evaluating Large Language Models as Role-Playing Game Engines (2025)
- Leveraging Graph Structures and Large Language Models for End-to-End Synthetic Task-Oriented Dialogues (2025)
- PokerBench: Training Large Language Models to become Professional Poker Players (2025)
- MIRAGE: Exploring How Large Language Models Perform in Complex Social Interactive Environments (2025)
- ACEBench: Who Wins the Match Point in Tool Usage? (2025)
- PsyPlay: Personality-Infused Role-Playing Conversational Agents (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper