Skip to content

A bilingual (English&Chinese) benchmark for implicit bias evaluation in Large Language Models using psychometric attack methods, forked from @yuchenwen1

Notifications You must be signed in to change notification settings

CAS-AISafetyBasicResearchGroup/BUMBLE

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

BUMBLE Benchmark

A bilingual(English&Chinese) benchmark for implicit bias evaluation in Large Language Models using psychometric attack methods

Run evaluations

  1. Install the required packages by running pip install -r requirements.txt.
  2. cd scripts/<MODEL> to cd into some directory.
  3. Fill in the config.py file in the directory, including API key and url.
  4. Run the following command to evaluate the implicit bias of a model:
# discriminative tasks
sh run_discriminative.sh

# generative tasks
sh run_generative.sh

About

A bilingual (English&Chinese) benchmark for implicit bias evaluation in Large Language Models using psychometric attack methods, forked from @yuchenwen1

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 89.4%
  • Shell 10.6%