As researchers, you don't need hype. You need evidence.
We collaborate with independent researchers and academic institutions to evaluate AI-moderated interviews under real research conditions. Glaut provides full platform access, while our partners run the tests and conduct the analysis independently.
Abstract
A comparative study with the University of Mannheim found that AI-moderated interviews (AIMI) generated higher-quality open-ended responses than a static online survey using the same questionnaire. The responses were more linguistically rich, with increases of 39% in word count, 51% in unique words, and 12% in lexical diversity, without reducing readability or the proportion of content words. AIMI also covered a wider range of themes (+36% unique themes), eliminated gibberish answers (0% vs. 10%), and enhanced the participant experience by 6%.

We test before we tell.
Every claim about AIMIs is grounded in comparative experiments, evaluating performance, engagement, and data quality against established methods like surveys, IDIs, and CATI.
A full-service agency crafting digital experiences that inspire awe and wonder.
We design studies to learn, not to confirm.
From dynamic follow-ups to voice-based interviews, each experiment challenges how research is conducted and opens new possibilities for scalability, empathy, and precision.
Research should reach everyone.
We explore how AI-moderated interviews make participation accessible to children, elderly adults, and other groups often excluded by traditional methods.
We believe rigorous research is collective work.
Every study is run internally at Glaut and in partnership with independent researchers, MR firms, and academic teams to ensure transparency, credibility, and shared advancement.
Abstract
A comparative study with the University of Mannheim found that AI-moderated interviews (AIMI) generated higher-quality open-ended responses than a static online survey using the same questionnaire. The responses were more linguistically rich, with increases of 39% in word count, 51% in unique words, and 12% in lexical diversity, without reducing readability or the proportion of content words. AIMI also covered a wider range of themes (+36% unique themes), eliminated gibberish answers (0% vs. 10%), and enhanced the participant experience by 6%.
Abstract
When people speak instead of type, they share more, and differently. Across 252 AI-moderated interviews, voice responses were 236% longer, 138% more varied, and 28% richer in themes than text. Yet participants rated all formats equally high for ease, empathy, and openness, with 55% still preferring text for privacy and control. Voice brings depth, text offers comfort, and hybrid AIMIs balance both.
Together, we advance research.
Join us in testing and shaping a new hybrid methodology
Researchers have already partnered with us across Europe, Australia and the US.
Comparative studies in progress, exploring data quality, empathy, inclusion, and efficiency.
Collaborate with Glaut to test, compare, and evolve AI-Moderated Interview methods through open, evidence-based experimentation.