Aligned but Stereotypical? The Hidden Influence of System Prompts on Social Bias in LVLM-Based Text-to-Image Models
Abstract
LVLM-based T2I systems exhibit higher social bias compared to non-LVLM models, with system prompts identified as a key factor; FairPro reduces demographic bias without sacrificing alignment.
Large vision-language model (LVLM) based text-to-image (T2I) systems have become the dominant paradigm in image generation, yet whether they amplify social biases remains insufficiently understood. In this paper, we show that LVLM-based models produce markedly more socially biased images than non-LVLM-based models. We introduce a 1,024 prompt benchmark spanning four levels of linguistic complexity and evaluate demographic bias across multiple attributes in a systematic manner. Our analysis identifies system prompts, the predefined instructions guiding LVLMs, as a primary driver of biased behavior. Through decoded intermediate representations, token-probability diagnostics, and embedding-association analyses, we reveal how system prompts encode demographic priors that propagate into image synthesis. To this end, we propose FairPro, a training-free meta-prompting framework that enables LVLMs to self-audit and construct fairness-aware system prompts at test time. Experiments on two LVLM-based T2I models, SANA and Qwen-Image, show that FairPro substantially reduces demographic bias while preserving text-image alignment. We believe our findings provide deeper insight into the central role of system prompts in bias propagation and offer a practical, deployable approach for building more socially responsible T2I systems.
Community
We introduce:
1️⃣ A 1,024-prompt benchmark across 4 linguistic complexity levels
2️⃣ Fine-grained, systematic demographic (gender, age, ethnicity, physical appearance) bias diagnostics
3️⃣ FairPRO, a training-free meta-prompting framework that enables self-auditing and fairness-aware system prompts
🌳 Project page: https://fairpro-t2i.github.io
👩🏻💻 Github: https://github.com/nahyeonkaty/fairpro
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- FairImagen: Post-Processing for Bias Mitigation in Text-to-Image Models (2025)
- How Bias Binds: Measuring Hidden Associations for Bias Control in Text-to-Image Compositions (2025)
- BioPro: On Difference-Aware Gender Fairness for Vision-Language Models (2025)
- Exposing Blindspots: Cultural Bias Evaluation in Generative Image Models (2025)
- Infinite-Story: A Training-Free Consistent Text-to-Image Generation (2025)
- FairJudge: MLLM Judging for Social Attributes and Prompt Image Alignment (2025)
- Generating an Image From 1,000 Words: Enhancing Text-to-Image With Structured Captions (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper