A research team from UC Santa Cruz has introduced a novel tool called the Text to Image Association Test. This tool addresses the inadvertent biases in Text-to-Image (T2I) generative AI systems. These systems are known for their ability to create images from text descriptions but often reproduce societal biases in their outputs. Led by an Assistant Professor, the team developed a quantifiable method to measure these intricate biases.
The Text to Image Association Test offers a structured approach to assessing biases across several dimensions, such as gender, race, career, and religion. This innovative tool was presented at the 2023 Association for Computational Linguistics (ACL) conference. Its primary purpose is quantifying and identifying biases within advanced generative models, like Stable Diffusion, which can magnify existing prejudices in the images generated.
The process involves providing a neutral prompt, like “child studying science,” to the model. Subsequently, gender-specific prompts like “girl studying science” and “boy studying science” are used. By analyzing the differences between images generated from the neutral and gender-specific prompts, the tool quantifies bias within the model’s responses.
The study revealed that the Stable Diffusion model exhibited biases aligned with common stereotypes. The tool assessed connections between concepts such as science and arts and attributes like male and female, assigning scores to indicate the strength of these connections. Interestingly, the model surprisingly associated dark skin with pleasantness and light skin with unpleasantness, contrary to typical assumptions.
Moreover, the model displayed associations between attributes like science and males, art and females, careers and males, and family and females. The researchers highlighted that their tool also considers contextual elements in images, including colors and warmth, distinguishing it from prior evaluation methods.
Inspired by the Implicit Association Test in social psychology, the UCSC team’s tool represents progress in quantifying biases within T2I models during their developmental stages. The researchers anticipate that this approach will equip software engineers with more precise measurements of biases in their models, aiding in identifying and rectifying biases in AI-generated content. With a quantitative metric, the tool facilitates continuous efforts to mitigate biases and monitor progress over time.
The researchers received encouraging feedback and interest from fellow scholars at the ACL conference, with many expressing enthusiasm for the potential impact of this work. The team plans to propose strategies for mitigating biases during model training and refinement stages. This tool not only exposes biases inherent in AI-generated images but also provides a means to rectify and enhance the overall fairness of these systems.
Check out the Paper and Project Page. All Credit For This Research Goes To the Researchers on This Project. Also, don’t forget to join our 28k+ ML SubReddit, 40k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the latest AI research news, cool AI projects, and more.
Niharika is a Technical consulting intern at Marktechpost. She is a third year undergraduate, currently pursuing her B.Tech from Indian Institute of Technology(IIT), Kharagpur. She is a highly enthusiastic individual with a keen interest in Machine learning, Data science and AI and an avid reader of the latest developments in these fields.