Evaluating LLM-Generated Topics from Survey Responses: Identifying Challenges in Recruiting Participants through Crowdsourcing

Reham Al Tamime*, Joni Salminen, Soon Gyo Jung, Bernard Jansen

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

The evolution of generative artificial intelligence (AI) technologies, particularly large language models (LLMs), has lead to consequences for the field of Human-Computer Interaction (HCI) in areas such as personalization, predictive analytics, automation, and data analysis. This research aims to evaluate LLM-generated topics derived from survey responses in comparison with topics suggested by humans, particularly participants recruited through a crowdsourcing experiment. We present an evaluation results to compare LLM-generated topics with human-generated topics in terms of Quality, Usefulness, Accuracy, Interestingness, and Completeness. This involves three stages: (1) Design and Generate Topics with an LLM (OpenAI's GPT-4); (2) Crowdsourcing Human-Generated Topics; and (3) Evaluation of Human-Generated Topics and LLM-Generated Topics. However, a feasibility study with 33 crowdworkers indicated challenges in using participants for LLM evaluation, particularly in inviting humans participants to suggest topics based on open-ended survey answers. We highlight several challenges in recruiting crowdsourcing participants for generating topics from survey responses. We recommend using well-trained human experts rather than crowdsourcing to generate human baselines for LLM evaluation.

Original languageEnglish
Title of host publicationProceedings - 2024 IEEE Symposium on Visual Languages and Human-Centric Computing, VL/HCC 2024
PublisherIEEE Computer Society
Pages412-416
Number of pages5
ISBN (Electronic)9798350366136
DOIs
Publication statusPublished - 2024
Event2024 IEEE Symposium on Visual Languages and Human-Centric Computing, VL/HCC 2024 - Liverpool, United Kingdom
Duration: 2 Sept 20246 Sept 2024

Publication series

NameProceedings of IEEE Symposium on Visual Languages and Human-Centric Computing, VL/HCC
ISSN (Print)1943-6092
ISSN (Electronic)1943-6106

Conference

Conference2024 IEEE Symposium on Visual Languages and Human-Centric Computing, VL/HCC 2024
Country/TerritoryUnited Kingdom
CityLiverpool
Period2/09/246/09/24

Keywords

  • Challenges in Recruitment
  • Crowdsourcing for Human-centric Computing
  • Feasibility Study
  • LLM Evaluation

Fingerprint

Dive into the research topics of 'Evaluating LLM-Generated Topics from Survey Responses: Identifying Challenges in Recruiting Participants through Crowdsourcing'. Together they form a unique fingerprint.

Cite this