A bilingual(English&Chinese) benchmark for implicit bias evaluation in Large Language Models using psychometric attack methods
- Install the required packages by running
pip install -r requirements.txt
. cd scripts/<MODEL>
to cd into some directory.- Fill in the
config.py
file in the directory, including API key and url. - Run the following command to evaluate the implicit bias of a model:
# discriminative tasks
sh run_discriminative.sh
# generative tasks
sh run_generative.sh