A Prompt Array Keeps the Bias Away: Debiasing Vision-Language Models with Adversarial Learning

03/22/2022
by   Hugo Elias Berg, et al.
2

Vision-language models can encode societal biases and stereotypes, but there are challenges to measuring and mitigating these harms. Prior proposed bias measurements lack robustness and feature degradation occurs when mitigating bias without access to pretraining data. We address both of these challenges in this paper: First, we evaluate different bias measures and propose the use of retrieval metrics to image-text representations via a bias measuring framework. Second, we investigate debiasing methods and show that optimizing for adversarial loss via learnable token embeddings minimizes various bias measures without substantially degrading feature representations.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset