Repository for the Description-based Bias Benchmark dataset.
The Description-Based Bias Benchmark (DBB) is a large-scale dataset designed to systematically evaluate social bias in large language models (LLMs) at the semantic, contextual, and descriptive levels.
- Instances: ~103,649 pairs of questions/options
- Categories: Age, Gender, Race/Ethnicity, Socioeconomic Status (SES), Religions
- Purpose: Benchmark and analyze social bias in LLMs at description level.
- Users: NLP researchers, LLM developers, fairness auditors.
Each item contains:
- A scenario with demographic identity (explicitly or implicitly) (context)
- Two answer options reflecting opposing concepts
- Concept pairs
- Traditional stereotype explanation
- Category label (e.g., gender, SES, etc.)
- Biased Target (e.g. male, young, etc.)
- Bias concepts adapted from SOFA, BBQ, StereoSets, and Crows-Pairs.
- Contexts and options generated using GPT-4o, then refined.
- Manual Review: Every sample included has been individually reviewed and confirmed to meet success criteria for fluency, coherence, and semantic alignment.
We retrieve stereotypical concepts by using GPT-4o from SOFA, BBQ, StereoSets, and Crows-Pairs.
And pairing with anti-stereotypical concepts correspondingly.
Concepts are in 📂 concept_lists/📄modified_all_concepts074.csv
Our DBB dataset is in 📂 data/📄 Bias-Dataset.csv
Bias-Dataset-More-Samples.zip has more samples for the dataset.
Below is the instructions you can generate a dataset to explore bias via the description-based method. Codes are in 📂 src/
python concept_analysis.py --model_name=gpt-4o --dataset=bbq --all
Can use any datasets you want. NOT Only limited to the datasets mentioned before.
python q_generate.py --model_name=gpt-4o --all_q
Use questions_final.ipynb to replace [[X]] to finish up question generation.
GPT-4o-results.zip contains results of each question for GPT-4o in DBB.
If you use DBB in your work, please cite:
@inproceedings{pan-etal-2025-whats,
title = "What{'}s Not Said Still Hurts: A Description-Based Evaluation Framework for Measuring Social Bias in {LLM}s",
author = "Pan, Jinhao and
Raj, Chahat and
Yao, Ziyu and
Zhu, Ziwei",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2025.findings-emnlp.76/",
pages = "1438--1459",
ISBN = "979-8-89176-335-7",
}