Skip to main navigation Skip to search Skip to main content

Prompt refinement or fine-tuning? best practices for using LLMs in computational social science tasks

  • Pioneer Center for AI

Research output: Conference Article in Proceeding or Book/Report chapterArticle in proceedingsResearchpeer-review

Abstract

Large Language Models are expressive tools that enable complex tasks of text understanding within Computational Social Science. Their versatility, while beneficial, poses a barrier for establishing standardized best practices within the field. To bring clarity on the values of different strategies, we present an overview of the performance of modern LLM-based classification methods on a benchmark of 23 social knowledge tasks. Our results point to three best practices: prioritize models with larger vocabulary and pre-training corpora; avoid simple zero-shot in favor of AI-enhanced prompting; fine-tune on task-specific data, and consider more complex forms instruction-tuning on multiple datasets only when only training data is more abundant.
Original languageEnglish
Title of host publicationProceedings of the Third Workshop on Social Influence in Conversations (SICon 2025)
PublisherAssociation for Computational Linguistics
Publication dateJul 2025
ISBN (Electronic)979-8-89176-266-4
DOIs
Publication statusPublished - Jul 2025
EventProceedings of the Third Workshop on Social Influence in Conversations - Vienna, Austria
Duration: 31 Jul 202531 Jul 2025

Conference

ConferenceProceedings of the Third Workshop on Social Influence in Conversations
Country/TerritoryAustria
CityVienna
Period31/07/202531/07/2025

Keywords

  • Large Language Models
  • Computational Social Science
  • AI-enhanced prompting
  • fine-tuning
  • instruction-tuning

Fingerprint

Dive into the research topics of 'Prompt refinement or fine-tuning? best practices for using LLMs in computational social science tasks'. Together they form a unique fingerprint.

Cite this