Skip to yearly menu bar Skip to main content


Poster

Quality Diversity through Human Feedback: an Open-Ended Backend for Diversity-Driven Optimization

Li Ding · Jenny Zhang · Jeff Clune · Lee Spector · Joel Lehman


Abstract:

Reinforcement Learning from Human Feedback (RLHF) has shown potential in qualitative tasks where easily-defined performance measures are lacking. However, there are drawbacks when it is used to optimize for average human preferences (as is common practice in RLHF), especially in generative tasks that demand diverse model responses. Meanwhile, Quality Diversity (QD) algorithms excel at identifying diverse and high-quality solutions but often rely on manually-crafted diversity metrics. This paper introduces Quality Diversity through Human Feedback (QDHF), a novel approach integrating human feedback into the QD framework. QDHF progressively infers diversity metrics from human judgments of similarity among solutions, thereby enhancing the applicability and effectiveness of QD algorithms. Our empirical studies show that QDHF significantly outperforms state-of-the-art methods in automatic diversity discovery and matches the efficacy of manually-crafted metrics for QD on standard benchmarks in robotics and reinforcement learning. Notably, in a latent space illumination task, QDHF substantially enhances the diversity of images generated by a diffusion model and was more favorably received in user studies. We conclude by analyzing QDHF's scalability and the quality of its derived diversity metrics, emphasizing its potential to improve exploration and diversity in complex, open-ended optimization tasks.

Live content is unavailable. Log in and register to view live content