Free Lunch for User Experience: Crowdsourcing Agents for Scalable User Studies
Siyang Liu, Sahand Sabour, Xiaoyang Wang, Rada Mihalcea
Run your next usability test with 1,000 agents before recruiting humans. Use it for rapid iteration on information architecture or onboarding flows. Validate edge cases (low-literacy users, non-native speakers) that are expensive to recruit traditionally.
User studies cost $50-150 per participant and take weeks to recruit. You're bottlenecked by budget and timeline, so you test with 12 people and call it validated.
Method: This team recruited simulated users from billion-scale profile datasets—not generic LLM personas, but agents instantiated from real demographic and behavioral data. They ran studies at 100x scale compared to traditional recruitment. The key mechanism: agents inherit attributes from actual user profiles (age, tech literacy, domain expertise) rather than relying on prompt engineering alone. This grounds the simulation in empirical distributions, not researcher assumptions about 'typical users.'
Caveats: Fidelity doubts remain—agents may miss embodied or emotional nuances. Best for cognitive tasks, not affective design decisions.
Reflections: Which UX tasks show the largest fidelity gap between agent and human responses? · Can agents simulate frustration or fatigue during long tasks? · How do you validate that agent diversity matches real population variance?