conferencePaper

Evaluating non-AI experts' interaction with AI: a case study in library context

Qingxiao Zheng, Minrui Chen, Hyanghee Park, Zhongwei Xu, Yun Huang

2025-00-00 2025

Publisher: Association for Computing Machinery

DOI: 10.1145/3706598.3714219

URL: https://doi.org/10.1145/3706598.3714219

ISBN: 979-8-4007-1394-1

SSH Fingerprint

SHA256:7oLxJHveRRFPrv15lHs1Q8v91v9QuW9lwmOMqN3aa5c

Zotero Key: 2RYTA6QS

Abstract

Public libraries in the U.S. are increasingly facing labor shortages, tight budgets, and overworked staff, creating a pressing need for conversational agents to assist patrons. The democratization of generative AI has empowered public service professionals to develop AI agents by leveraging large language models. To understand the needs of non-AI library professionals in creating their own conversational agents, we conducted semi-structured interviews with library professionals (n=11) across the U.S. Insights from these interviews informed the design of AgentBuilder, a prototype tool that enables non-AI experts to create conversational agents without coding skills. We then conducted think-aloud sessions and follow-up interviews to evaluate the prototype experience and identify the key evaluation criteria emphasized by library professionals (n=12) when developing conversational agents. Our findings highlight how these professionals perceive the prototype experience and reveal five essential evaluation criteria: interpreting user intent, faithful paraphrasing, proper alignment with authoritative sources, tailoring the tone of voice, and handling unknown answers effectively. These insights provide valuable guidance for designing AI-supported "end-user creation tools" in public service domains beyond libraries.

Tags

Schools of Thought