A Project on Mitigating Racial Bias in Face Aging Models
Facial Age Progression (FAP)—the ability to predict what someone will look like as they age—is a fascinating frontier in computer vision. But many state-of-the-art models have a critical flaw: they are trained on datasets that don't represent the diversity of the real world. Datasets like FFHQ, while high-quality, are heavily skewed towards caucasian faces.
This imbalance means that models trained on this data often fail to accurately capture the unique aging patterns of non-white individuals. An age progression model that changes a person's perceived race is not just inaccurate; it's a failure of fairness. This project tackles that problem head-on.
Our starting point was a powerful model called SAM (Style-based Age Manipulation). SAM is an image-to-image translation model built on StyleGAN, a renowned generative network. It works by encoding a person's face into a latent space—a compressed representation of facial features—and then cleverly manipulating that representation to change their age while preserving their identity.
The core of our idea was simple: what if we could penalize the model every time it generated a face whose race was different from the input? To do this, we integrated a pre-trained race classifier, DeepFace, directly into the training loop.
During training, after SAM generated an aged face, DeepFace would analyze both the original and the new face. If it detected a change in racial characteristics, it would send a "race loss" signal back to the main model, nudging its parameters to correct the mistake. In essence, we taught the model to preserve racial identity as a core objective, alongside preserving personal identity and accurately portraying age.
While quantitative metrics are still a work in progress, the qualitative results, even from early training, are striking. The most powerful demonstration is a direct video comparison.
The effect is also clear in still images. We compared generations from our model (trained from scratch with a race loss lambda of 15) against the original vanilla SAM.
This project, while challenging due to computational constraints, shows significant promise. Our key takeaway is that incorporating auxiliary information—like a race classifier—directly into the loss function can be a powerful strategy for mitigating dataset bias without needing to curate a perfectly balanced dataset.
Ultimately, this work is a step towards building AI systems that are not only technologically advanced but also socially conscious and equitable for everyone.
You can experiment with our fairness-enhanced model yourself using these Google Colab notebooks. No setup is required.