Skip to main navigation Skip to search Skip to main content

The Unseen Hand: How User Background Shapes AI Text-to-Image Generation

  • Royal Danish Academy

Research output: Conference Article in Proceeding or Book/Report chapterArticle in proceedingsResearchpeer-review

Abstract

Text-To-Image (TTI) generators are becoming widely used and are often promoted as "democratizing" the making of images independently of the skill set of the user. But relatively little is known about whether people with different educations, skill sets and literacies use these tools differently, and how their backgrounds influence the quality of the results. In this article we investigate the impact of Visual Literacy (VL), AI Literacy (AIL) and Prompt Engineering Literacy (PEL) on prompt use. More precisely we are examining how they each
impact prompt usage patterns, linguistic and semantic composition, and the structure and morphology of prompts. Additionally, we employed Process Analysis to investigate how participants with different literacy
approach the creative design process with TTI. Our results show that individuals scoring high on Visual Literacy (VL) tend to employ a larger vocabulary and more nuanced language, greater prompt variety and make more references to art genres, styles, and movements. In contrast, participants scoring high AI Literacy (AIL) demonstrated a deeper understanding of AI interaction, influencing their prompting strategies, and tended to adhere closely to the exact wording of the brief, treating it as precise specifications. The study's findings have the potential of significantly impacting educations, both by actively teaching students to experiment with prompt variations and deepen expressive visual vocabulary as well as designing curricula that incorporate structured exercises on iterative prompt refinement, explicitly teaching how to adjust prompts based on AI
output to achieve desired visual outcomes.
Original languageEnglish
Title of host publicationProceedings of the International Association of Societies of Design Research Conference (IASDR2025)
Number of pages17
Place of PublicationTaipei, Taiwan
Publication date1 Dec 2025
Publication statusPublished - 1 Dec 2025
EventInternational Association of Societies of Design Research Conference - Taipei, Taiwan, Province of China
Duration: 2 Dec 20255 Dec 2025

Conference

ConferenceInternational Association of Societies of Design Research Conference
Country/TerritoryTaiwan, Province of China
CityTaipei
Period02/12/202505/12/2025

Keywords

  • Text-to-Image Synthesis
  • Visual Literacy
  • AI Literacy
  • Prompt Engineering Literacy
  • Educational Implications of Generative AI

Fingerprint

Dive into the research topics of 'The Unseen Hand: How User Background Shapes AI Text-to-Image Generation'. Together they form a unique fingerprint.

Cite this