People form ad hoc conventions, by establishing linguistic & gestural abstractions, and shift information across speech and gesture to communicate more efficiently over time.
In our upcoming #CHI2026 paper, we study how these multimodal communications evolve in repeated physical collaboration.
Led by Kiyosu Maeda in close collaboration with @jefan, @rdhawkins, and team: William McCarthy, Ching-Yi Tsai, Jeffrey Mu, and Haoliang Wang.
🧵👇 1/4
Using #AR, we carefully isolate speech and gestures, removing other cues (e.g., gaze, facial expressions). This allows us to analyze how partners coordinate on abstractions and how information shifts across these modalities over time.
We develop a computational model, extending the Rational Speech Act (RSA) framework to multimodal settings, and simulate the behaviors we observe.
2/4