Abstract
Large Language Models are expressive tools that enable complex tasks of text understanding within Computational Social Science. Their versatility, while beneficial, poses a barrier for establishing standardized best practices within the field. To bring clarity on the values of different strategies, we present an overview of the performance of modern LLM-based classification methods on a benchmark of 23 social knowledge tasks. Our results point to three best practices: prioritize models with larger vocabulary and pre-training corpora; avoid simple zero-shot in favor of AI-enhanced prompting; fine-tune on task-specific data, and consider more complex forms instruction-tuning on multiple datasets only when only training data is more abundant.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of the Third Workshop on Social Influence in Conversations (SICon 2025) |
| Publisher | Association for Computational Linguistics |
| Publication date | Jul 2025 |
| ISBN (Electronic) | 979-8-89176-266-4 |
| DOIs | |
| Publication status | Published - Jul 2025 |
| Event | Proceedings of the Third Workshop on Social Influence in Conversations - Vienna, Austria Duration: 31 Jul 2025 → 31 Jul 2025 |
Conference
| Conference | Proceedings of the Third Workshop on Social Influence in Conversations |
|---|---|
| Country/Territory | Austria |
| City | Vienna |
| Period | 31/07/2025 → 31/07/2025 |
Keywords
- Large Language Models
- Computational Social Science
- AI-enhanced prompting
- fine-tuning
- instruction-tuning
Fingerprint
Dive into the research topics of 'Prompt refinement or fine-tuning? best practices for using LLMs in computational social science tasks'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver