What is it about?

We created a web-based platform which makes it easy to test out chatbots, and rate them based on how well they perform. It allows you to have conversations with different bots, and judge which ones seem, for example, more human-like, accurate, and helpful in their responses.

Featured Image

Why is it important?

Given the increase of chatbots in our everyday life, we need a way to rate them based on human interaction. Our platform makes it easy for anyone to evaluate chatbots through actual conversations, rather than just technical tests. By making the platform easy to set up, anyone can test their bots according to their personal questions and needs.

Perspectives

What made working on this article unique, is that it is not just a research article, but it is software that has been built from an actual need from within our research group. We do not only provide the platform for others to use, but actually use it in our yearly course ourselves, which is not often the case in our research work.

Daan van der Weijden
Universitat Zurich

Read the Original

This page is a summary of: Alan's Speakeasy - An Ecosystem for the Evaluation of Conversational Agents, May 2025, ACM (Association for Computing Machinery),
DOI: 10.1145/3701716.3715165.
You can read the full text:

Read

Resources

Contributors

The following have contributed to this page