Large Language Models (LLMs), now used daily by millions of users, can e...
Without proper safeguards, large language models will readily follow
mal...
Public figures receive a disproportionate amount of abuse on social medi...
We introduce VisoGender, a novel dataset for benchmarking gender bias in...
Vision-language models are growing in popularity and public visibility t...
The generative AI revolution in recent years has been spurred by an expa...
Large language models (LLMs) are used to generate content for a wide ran...
Online sexism is a widespread and harmful phenomenon. Automated tools ca...
The emergence of large language models (LLMs) represents a major advance...
Annotating abusive language is expensive, logistically complex and creat...
The growing capability and availability of generative language models ha...
Textual data can pose a risk of serious harm. These harms can be categor...
Vision-language models can encode societal biases and stereotypes, but t...
Detecting online hate is a complex task, and low-performing models have
...
Hateful memes pose a unique challenge for current machine learning syste...