Aligning every agent individually does not scale. We train cooperative behavior into a single seed agent — and watch it spread to untrained agents through interaction alone.
Multi-agent systems require robust alignment, but aligning every agent individually does not scale to open environments with many interacting models. We propose Alignment Propagation, where cooperative behavior is instilled in a single fine-tuned seed agent and spreads to untrained agents through interaction.
To study this effect, we introduce the Alignment Propagation Playground with two complementary settings: Red-Black Game, a discrete social dilemma with broadcast deliberation, and Sugarscape, a continuous resource-competition world with pairwise negotiation. We use a frontier model to generate cooperative trajectories and fine-tune a seed agent, then deploy seeds into otherwise untrained collectives.
A single seed more than doubles cooperation on held-out Red-Black scenarios (26% → 62%), scaling to 96% with five seeds. Without retraining, seeds transfer zero-shot to Sugarscape (91.5% trade success vs. 21.6% for an untrained baseline) and outperform prompt-based Gemini 3 Pro. Finally, topology governs propagation efficiency: broadcast deliberation requires 20% seeds to shift the group, whereas pairwise negotiation requires ~50%.
Two environments, each designed to pressure agents to defect for local gain — and to test whether cooperation, instilled in a few, can propagate through interaction.
@inproceedings{hsing2026yoao,
title = {You Only Align Once: Propagating Cooperative Behaviors
in Multi-Agent Systems},
author = {Hsing, Nicole and Zheng, Asuka Yuxi and Zhao, Yi
and Tu, Haoqin and Huang, Jen-tse},
booktitle = {Workshop on Lifelong Agents: Learning, Aligning, Evolving,
ICLR},
year = {2026}
}