Training Natively Interactive Models
Anne Wu, Cornell University
Large language models are interactive artifacts, but interaction is baked into them almost offhandedly, both for deployment and learning, and this has far-reaching consequences: rather than supporting the naturalness of human interactions, interactions with LLMs are stilted turn-taking experiences, a far cry from the dynamic full-duplex (i.e., concurrent input/output) nature of human interaction. While reinforcement learning methods utilize interactions, they remain mostly applicable to narrow domains due to their dependence on simplified rewards or the risks of reward hacking.
This talk covers two projects: I will spend most of the talk describing full-duplex speech LLMs and the methods I designed to post-train them, showing consistent improvements in conversational engagement, safety, and factuality; and in the second part, I will discuss an RL benchmark for visuo-linguistic reasoning I created, focused on designing robust reward signals that are not susceptible to reward hacking or require significant task simplifications.
