BiasAware is a comprehensive research platform that exposes and quantifies social biases in modern AI text-to-image generation models. Through the analysis of over 3,300 generated images across 11 bias categories, this project reveals how AI systems perpetuate harmful demographic stereotypes and societal inequalities.
Built as an undergraduate engineering capstone project at Toronto Metropolitan University (2024), BiasAware combines machine learning research, statistical analysis, and interactive web development to make complex bias patterns accessible and understandable.
- 3,300+ Images Analyzed across 11 distinct bias categories
- Quantified Demographic Bias in race, gender, and age representation
- Custom AI Model Training on Toronto neighborhood data
- Interactive Web Platform with statistical visualizations
- Comprehensive Research Findings published with academic rigor
Based on our comprehensive analysis documented in the research findings, we examined bias across the following categories:
| Category | Focus Area | Sample Size | Key Bias Patterns |
|---|---|---|---|
| Activities | Sports, recreation, hobbies | 150+ images | Male-dominated (65%), White majority (70%) |
| Addictions | Substance use representations | 150+ images | Gender and racial stereotyping |
| Crime | Law enforcement, criminal justice | 150+ images | Severe racial bias (60% racialized) |
| Emotions | Emotional expression patterns | 150+ images | White bias in positive emotions (72%) |
| Engineering | Technical field diversity | 150+ images | Strong male bias (82%), varies by field |
| Healthcare | Medical profession representations | 150+ images | Female majority (62%), moderate diversity |
| Neighborhoods | Socioeconomic geography (Toronto focus) | 150+ images | Clear wealth-based architectural patterns |
| Professions | Career representations across sectors | 150+ images | Varies significantly by profession type |
| Quality of Life | Lifestyle and socioeconomic indicators | 150+ images | Racial disparities in affluence depiction |
- React.js - Modern single-page application
- Material-UI - Consistent component library
- Custom CSS - Responsive design with advanced animations
- Node.js + Express.js - RESTful API architecture
- MongoDB Atlas - Cloud database with GridFS image storage
- Statistical APIs - Real-time bias calculation endpoints
- Stable Diffusion - Primary image generation model
- NightCafe - Secondary generation platform
- Custom LoRA Training - Fine-tuned models for neighborhood bias
- HuggingFace - Model training and deployment infrastructure
Based on our detailed analysis in the research findings documentation, we discovered significant bias patterns:
| Bias Type | Most Biased Category | Representation Gap |
|---|---|---|
| Gender | Engineering (82% male) | 64% over-representation |
| Race | Crime (60% racialized) | 45% misrepresentation |
| Age | All categories (90% adult average) | 75% age concentration |
All findings are statistically validated with 95% confidence intervals on all measurements. Chi-square tests confirm bias significance (p < 0.05), and we developed a custom Bias Severity Index (0-100 scale) to quantify bias levels across categories. Complete statistical analysis is available in our research findings.
Our research methodology followed a systematic approach to identify and quantify bias patterns:
-
Initial Research
- Conducted a comprehensive literature review on AI biases
- Explored and evaluated multiple image generation platforms
-
Prompt Development
- Created diverse prompts targeting bias categories including age, gender, and race
- Generated multiple image variations for each prompt to ensure statistical validity
-
Image Analysis
- Manually annotated and quantified biases using demographic distribution metrics
- Applied statistical methods to validate findings across all categories
-
Web Development
- Built an interactive web application using the MERN stack
- Integrated dynamic UI components including heatmaps and image galleries
-
Model Fine-Tuning
- Trained custom models with Toronto neighborhood datasets to explore geographical bias
- Compared pre-training and post-training bias patterns
-
Evaluation
- Visualized comprehensive results through the web application
- Documented prominent stereotypes and demographic disparities across all categories
-
Clone the repository:
git clone https://github.com/mpat247/biasaware.git cd biasaware -
Install dependencies:
# Backend cd server && npm install # Frontend cd ../client && npm install
-
Start development servers:
# Backend (Terminal 1) cd server && npm start # Frontend (Terminal 2) cd client && npm start
Access the application at http://localhost:3000
Comprehensive technical documentation is available in the /docs folder:
- Project Overview - Complete technical summary and architecture details
- API Documentation - Backend endpoints and usage guidelines
- Deployment Guide - Production setup instructions
- Research Findings - Detailed bias analysis results and statistical validation