Prompt refinement or fine-tuning? best practices for using LLMs in computational social science tasks

Publikation: Konference artikel i Proceeding eller bog/rapport kapitelKonferencebidrag i proceedingsForskningpeer review

Abstract

Large Language Models are expressive tools that enable complex tasks of text understanding within Computational Social Science. Their versatility, while beneficial, poses a barrier for establishing standardized best practices within the field. To bring clarity on the values of different strategies, we present an overview of the performance of modern LLM-based classification methods on a benchmark of 23 social knowledge tasks. Our results point to three best practices: prioritize models with larger vocabulary and pre-training corpora; avoid simple zero-shot in favor of AI-enhanced prompting; fine-tune on task-specific data, and consider more complex forms instruction-tuning on multiple datasets only when only training data is more abundant.
OriginalsprogEngelsk
TitelProceedings of the Third Workshop on Social Influence in Conversations (SICon 2025)
ForlagAssociation for Computational Linguistics
Publikationsdatojul. 2025
ISBN (Elektronisk)979-8-89176-266-4
DOI
StatusUdgivet - jul. 2025
BegivenhedProceedings of the Third Workshop on Social Influence in Conversations - Vienna, Østrig
Varighed: 31 jul. 202531 jul. 2025

Konference

KonferenceProceedings of the Third Workshop on Social Influence in Conversations
Land/OmrådeØstrig
ByVienna
Periode31/07/202531/07/2025

Fingeraftryk

Dyk ned i forskningsemnerne om 'Prompt refinement or fine-tuning? best practices for using LLMs in computational social science tasks'. Sammen danner de et unikt fingeraftryk.

Citationsformater