AI Safety & Alignment

TriPlay-RL: Tri-Role Self-Play Reinforcement Learning for LLM Safety Alignment

ZZhewen TanWWenhan YuJJianfeng SiTTongxin LiuKKaiqi GuanHHuiyan JinJJiawen TaoXXiaokun YuanDDuohe MaXXiangzheng ZhangTTong YangLLin Sun
Published
January 26, 2026
Authors
12

Abstract

In recent years, safety risks associated with large language models have become increasingly prominent, highlighting the urgent need to mitigate the generation of toxic and harmful content. The mainstream paradigm for LLM safety alignment typically adopts a collaborative framework involving three roles: an attacker for adversarial prompt generation, a defender for safety defense, and an evaluator for response assessment. In this paper, we propose a closed-loop reinforcement learning framework called TriPlay-RL that enables iterative and co-improving collaboration among three roles with near-zero manual annotation. Experimental results show that the attacker preserves high output diversity while achieving a 20%-50% improvement in adversarial effectiveness; the defender attains 10%-30% gains in safety performance without degrading general reasoning capability; and the evaluator continuously refines its fine-grained judgment ability through iterations, accurately distinguishing unsafe responses, simple refusals, and useful guidance. Overall, our framework establishes an efficient and scalable paradigm for LLM safety alignment, enabling continuous co-evolution within a unified learning loop.

Keywords

reinforcement learninglarge language modelssafety alignmentadversarial prompt generationsafety defenseresponse assessmentclosed-loop frameworkco-improving collaborationiterative learning

More in AI Safety & Alignment

View all
TriPlay-RL: Tri-Role Self-Play Reinforcement Learning for LLM Safety Alignment | Paperchime