Enabling Classifiers to Make Judgements Explicitly Aligned with Human Values

10/14/2022
by   Yejin Bang, et al.
2

Many NLP classification tasks, such as sexism/racism detection or toxicity detection, are based on human values. Yet, human values can vary under diverse cultural conditions. Therefore, we introduce a framework for value-aligned classification that performs prediction based on explicitly written human values in the command. Along with the task, we propose a practical approach that distills value-aligned knowledge from large-scale language models (LLMs) to construct value-aligned classifiers in two steps. First, we generate value-aligned training data from LLMs by prompt-based few-shot learning. Next, we fine-tune smaller classification models with the generated data for the task. Empirical results show that our VA-Models surpass multiple baselines by at least 15.56 existing text augmentation methods. We suggest that using classifiers with explicit human value input improves both inclusivity explainability in AI.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset