AgentPsyAssessment is a portable, comprehensive psychological assessment framework that combines various psychometric models (Big Five, MBTI, cognitive functions, and more) with AI-powered analysis. The framework provides unified assessment skills supporting 6 professional evaluation types with intelligent type detection and multi-language support.
- β¨ Unified Assessment Skills System - Configuration-driven architecture supporting 6 professional evaluation types
- π€ Intelligent Type Detection - Automatic assessment type identification without manual configuration
- π Interactive HTML Reports - Chart.js-powered visualization with responsive design
- π Multi-language Support - Chinese, English, German, French, Russian, Japanese, Spanish
- π 16 MBTI Personalities - Comprehensive personality profiling with Big Five mapping
- π Optimized Performance Engine - Async processing, intelligent caching, and enhanced metrics
- β‘ Smart Configuration Management - Auto-validation, parameter adjustment, and fault tolerance
- π Core Features
- ποΈ System Architecture
- β‘ Installation
- π Quick Start (5-Minute Experience)
- π§ Assessment Types
- π Usage Examples
- βοΈ Configuration
- π€ Models
- π Multi-language Support
- π Documentation
- π€ Contributing
- π License
- π Contact
- π§ 6 Professional Evaluation Types: Big Five Personality, Citizenship Knowledge, Financial Professional, Legal Knowledge, Motivation Psychology, Political Literacy
- π€ Intelligent Type Detection: Automatic assessment type identification from questionnaire content
- π Interactive HTML Reports: Chart.js visualization with responsive design and professional styling
- π 16 MBTI Personalities: Comprehensive personality profiling with detailed trait analysis
- π Multi-language Support: Full interface localization (Chinese, English, German, French, Russian, Japanese, Spanish)
- βοΈ Configuration-Driven: JSON-based configuration system for flexible assessment types
- Multi-Psychometric Assessment: Combines Big Five personality model, MBTI, cognitive functions, and specialized domains
- AI-Powered Analysis: Uses LLMs to analyze personality traits from question responses
- Flexible Model Support: Compatible with various LLMs (OpenAI, Claude, Ollama, local models)
- Comprehensive Output: Generates detailed personality profiles with confidence scoring
- Batch Processing: Supports bulk assessment processing with error recovery
- Consensus-Based Verification: Multi-model evaluation for reliable results
- Specialized Analysis: Domain-specific assessments for professional contexts
.claude/skills/unified-assessment-system/
βββ π config_validator.py # Configuration validation
βββ π assessment_detector.py # Intelligent assessment type detection
βββ ποΈ skill_base.py # Base skill architecture
βββ π unified_questionnaire_responder.py # Unified questionnaire responder
βββ π unified_psychological_analyzer.py # Unified psychological analyzer
βββ π unified_report_generator.py # Unified report generator
βββ π configs/ # Assessment type configurations
βββ big_five_personality.json # Big Five personality evaluation
βββ citizenship_knowledge.json # Citizenship knowledge evaluation
βββ financial_professional.json # Financial professional evaluation
βββ legal_knowledge.json # Legal knowledge evaluation
βββ motivation_psychology.json # Motivation psychology evaluation
βββ political_literacy.json # Political literacy evaluation
- Uses LLMs to generate responses to psychological questionnaires
- Configurable parameters for temperature, role-playing, context, stress levels
- Supports 16 MBTI personality profiles with Big Five tendencies mapping
- Provides standardized questionnaire administration for 6 assessment types
- Multi-language support with cultural adaptation
- Evaluates generated responses using scientific psychometric methods
- Performs Big Five personality scoring with MBTI type inference
- Conducts domain-specific analysis for professional contexts
- Creates interactive HTML reports with Chart.js visualization
- Generates confidence scores and validation reports
- Session-based evaluation with adaptive consensus algorithms
The assessment process follows a cyclical approach: test, evaluate, targeted test, evaluate, analyze:
- Initial Test: Administer standardized questionnaires to the LLM
- Initial Evaluation: Analyze responses for initial personality profile
- Targeted Test: Administer specific follow-up questions based on initial results
- Secondary Evaluation: Refine personality profile based on targeted questions
- Comprehensive Analysis: Generate detailed psychological report with recommendations
git clone https://github.com/ptreezh/AgentPsyAssessment.git
cd AgentPsyAssessment# Create virtual environment (recommended)
python -m venv psyagent-env
# Windows
psyagent-env\Scripts\activate
# Linux/macOS
source psyagent-env/bin/activate
# Install dependencies
pip install -r requirements.txt # if available
pip install ollama requests numpy pandas# Set provider (local or cloud)
export PROVIDER="local" # or "cloud"
# Windows (PowerShell)
$env:OPENAI_API_KEY = "your-openai-key"
$env:ANTHROPIC_API_KEY = "your-anthropic-key"
# macOS/Linux
export OPENAI_API_KEY="your-openai-key"
export ANTHROPIC_API_KEY="your-anthropic-key"# Run unified assessment system tests
cd .claude/skills/unified-assessment-system
python test_runner.py
# Expected output: π ALL TESTS PASSED!If you plan to use local models:
Download from https://ollama.ai and install according to your system.
# Start Ollama service
ollama serve
# Download recommended models
ollama pull qwen3:8b
ollama pull deepseek-r1:8b
ollama pull mistral-nemo:latest
ollama pull llama3:latest# Test unified assessment system (v1.0)
cd .claude/skills/unified-assessment-system
python test_runner.py
# Expected output: π ALL TESTS PASSED!# 1. Experience questionnaire generation
python llm_assessment/run_assessment_unified.py \
--model_name def \
--test_file llm_assessment/test_files/single_test_question_10.json \
--role_name def \
--tmpr 0.7
# 2. Experience batch analysis
python production_pipelines/local_batch_production/cli.py \
assess --model gpt-4o --role def
# 3. View results
ls results/# Start Ollama (if using local models)
ollama serve
# Download model
ollama pull llama3.1
# Run local evaluation
python llm_assessment/run_assessment_unified.py \
--model llama3.1 \
--role a1 \
--provider local# Run skills demo
python skills_demo_chinese_questionnaire.py
# View generated HTML reports
ls html/- π English Guide
- π©πͺ German Guide
- π«π· French Guide
- π·πΊ Russian Guide
- π―π΅ Japanese Guide
- πͺπΈ Spanish Guide
- π¨π³ Chinese Guide
- File Pattern:
*big_five*,*personality*,*ocean* - Dimensions: Openness, Conscientiousness, Extraversion, Agreeableness, Neuroticism
- Output: MBTI type inference + Big Five scoring + 16 personality profiles
- Use Case: General personality assessment and team compatibility analysis
- File Pattern:
*citizenship*,*ε ¬ζ°* - Focus: Civic rights & obligations, political system awareness
- Output: Civic literacy score + political knowledge assessment
- Use Case: Citizenship education and civic engagement evaluation
- File Pattern:
*financial*,*ιθ*,*bank* - Focus: Financial expertise, risk identification capabilities
- Output: Financial competency score + risk assessment profile
- Use Case: Financial industry recruitment and skill assessment
- File Pattern:
*legal*,*law*,*ζ³εΎ* - Focus: Legal foundations, practical operational capabilities
- Output: Legal knowledge score + practical judgment assessment
- Use Case: Legal profession evaluation and compliance assessment
- File Pattern:
*motivation*,*ε¨ζΊ* - Focus: Achievement motivation, power motivation, affiliation motivation
- Output: Motivation profile + drive pattern analysis
- Use Case: Career counseling and team motivation analysis
- File Pattern:
*political*,*ζΏζ²»* - Focus: Political system awareness, critical thinking
- Output: Political literacy score + analytical capability assessment
- Use Case: Political education and civic competency evaluation
# Test unified assessment system
cd .claude/skills/unified-assessment-system
python test_runner.py
# Expected output:
# β
PASS Configuration System (6/6 configs loaded)
# β
PASS Assessment Detection (2/2 detections successful)
# β
PASS Questionnaire Response (Generated 2 responses)
# β
PASS Psychological Analysis (Big Five + MBTI analysis)
# β
PASS Report Generation (HTML report generated)
# π ALL TESTS PASSED!# Big Five personality assessment
python llm_assessment/run_assessment_unified.py \
--test_file llm_assessment/test_files/agent-big-five-50-complete2.json \
--model_name gpt-4o --role_name enfj
# Citizenship knowledge assessment
python llm_assessment/run_assessment_unified.py \
--test_file llm_assessment/test_files/agent-citizenship-test.json \
--model_name claude-3-5-sonnet --role_name def
# Financial professional assessment
python llm_assessment/run_assessment_unified.py \
--test_file llm_assessment/test_files/agent-fund-management-test.json \
--model_name llama3.1 --role_name a1 --provider local# Multiple roles batch processing
python production_pipelines/local_batch_production/run_batch_suite.py \
--model llama3.1 --roles a1,a2,b1
# Enhanced batch analysis
python production_pipelines/local_batch_production/cli.py \
analyze --input results/latest_batch.json
# Generate HTML reports
python generate_all_html_reports.pyThe assessment component supports multiple configuration options:
--model: LLM model to use (e.g., gpt-4o, claude-3-5-sonnet, llama3.1)--role: Personality role to apply during assessment (a1-a10, b1-b10, def)--questions: Path to custom questions JSON file--temperature: Model temperature (default: 0)--top_p: Model top_p parameter (default: 0.9)--ollama: Use Ollama models instead of cloud APIs--host: Ollama host (default: http://localhost:11434)--context: Additional context to influence responses
# Basic assessment
python llm_assessment/run_assessment_unified.py --model gpt-4o --role def
# Assessment with specific parameters
python llm_assessment/run_assessment_unified.py --model claude-3-5-sonnet --role a1 --temperature 0.2The analysis component evaluates responses and generates comprehensive reports:
--analysis-type: Type of analysis (bigfive, mbti, belbin, comprehensive)--input: Path to assessment results to analyze--confidence-threshold: Confidence threshold for recommendations (default: 0.7)
# Analyze assessment results
python analyze_results.py --input results/assessment_result.json --analysis-type comprehensive
# Generate Big Five analysis
python analyze_big5_results.py --input results/assessment_result.json
# Generate MBTI analysis
python analyze_mbti_results.py --input results/assessment_result.jsonThe project includes both components accessible through the CLI:
# Run complete assessment and analysis
python cli.py assess --model gpt-4o --role def
python cli.py analyze --input results/latest_assessment.json
# Run assessment with integrated analysis
python cli.py assess --model gpt-4o --role def --analyze-
Initial Test: Administer standardized questionnaires to the LLM
python cli.py assess --model gpt-4o --role def
-
Initial Evaluation: Analyze initial responses
python cli.py analyze --input results/initial_assessment.json
-
Targeted Test: Run follow-up questions based on initial analysis
python cli.py assess --model gpt-4o --role targeted --context "focus on neuroticism and agreeableness" -
Secondary Evaluation: Refine personality profile
python cli.py analyze --input results/targeted_assessment.json
-
Comprehensive Analysis: Generate detailed report
python cli.py analyze --input results/combined_assessments.json --analysis-type comprehensive
Create a .env file with:
OPENAI_API_KEY=your_openai_api_key
ANTHROPIC_API_KEY=your_anthropic_api_key
OLLAMA_HOST=http://localhost:11434
Model-specific settings can be configured in config/ directory. Different models may require different parameters for optimal performance.
The system supports various LLMs with different capabilities:
- OpenAI: gpt-4, gpt-4o, gpt-4-turbo
- Anthropic: claude-3-5-sonnet, claude-3-opus, claude-3-sonnet
- Google: gemini-pro, gemini-1.5-pro
- Mistral: mistral-large, mistral-small
- Llama family: llama3.1, llama3.2, llama3
- Mixtral: mixtral-8x7b
- Others: phi3, command-r, deepseek-coder
- Questionnaire input (currently using Big Five 50-item model)
- Personality role application (optional)
- Context preparation for LLM
- Multi-step analysis for complex traits
- Evidence-based reasoning
- Confidence scoring
- Trait extraction and scoring
- Consistency checks
- Final report generation
- Cross-model verification (optional)
- Consistency checking
- Result validation
Assessment results are saved in JSON format with:
- Personality trait scores (Big Five dimensions)
- MBTI type determination
- Cognitive function preferences
- Confidence scores for each assessment
- Detailed analysis notes
The project includes tools for:
- Batch Analysis: Compare results across multiple assessments
- Reliability Analysis: Check consistency of results
- Segmentation Analysis: Analyze results by question categories
- Big Five Analysis: Detailed trait analysis
- Motivation Analysis: Motivation and drive assessment
We welcome contributions! Here's how you can help:
- Fork the repository
- Create a feature branch:
git checkout -b feature/amazing-feature - Commit your changes:
git commit -m 'Add amazing feature' - Push to the branch:
git push origin feature/amazing-feature - Open a Pull Request
- Follow the existing code style
- Write clear documentation
- Add tests for new features
- Use descriptive commit messages
- Ensure code passes linting
For development:
# Install in development mode
pip install -e .
# Run tests
python -m pytest tests/
# Format code
black .This project is licensed under the MIT License - see the LICENSE file for details.
AgentPsyAssessment provides comprehensive multi-language support:
- π¨π³ Chinese (Simplified) - Primary documentation and interface
- πΊπΈ English - Full documentation and interface support
- π©πͺ German - Complete documentation translation
- π«π· French - Comprehensive user guides
- π·πΊ Russian - Full localized documentation
- π―π΅ Japanese - Complete user manual and guides
- πͺπΈ Spanish - Comprehensive documentation
- Localized Interface: All UI elements translated
- Cultural Adaptation: Assessment questions adapted for cultural contexts
- Regional Content: Domain-specific examples localized by region
- Language Detection: Automatic language detection from questionnaire content
- Unicode Support: Full UTF-8 support for international characters
# Access language-specific documentation
docs/en/QUICK_START_GUIDE.md # English
docs/de/QUICK_START_GUIDE.md # German
docs/fr/QUICK_START_GUIDE.md # French
docs/ru/QUICK_START_GUIDE.md # Russian
docs/ja/QUICK_START_GUIDE.md # Japanese
docs/es/QUICK_START_GUIDE.md # Spanish
QUICK_START_GUIDE.md # Chinese (primary)- π User Manual - Comprehensive system documentation
- π Quick Start Guide - 5-minute setup and experience
- ποΈ System Architecture - Technical architecture and development guide
- π§ Configuration Guide - Detailed configuration options
- π§ Assessment Types Guide - All 6 evaluation types explained
- π Report Generation Guide - HTML report customization
- π Multi-language Guide - Internationalization features
- π Plugin Development Guide - Extending the system
- π Batch Processing Tutorial - Large-scale assessment
- βοΈ Cloud Deployment Guide - Production deployment
- π§ͺ Testing Guide - System testing and validation
We welcome contributions! Here's how you can help:
- Fork the repository
- Create a feature branch:
git checkout -b feature/amazing-feature - Commit your changes:
git commit -m 'Add amazing feature' - Push to the branch:
git push origin feature/amazing-feature - Open a Pull Request
- Follow the existing code style and architectural patterns
- Write clear documentation for new features
- Add comprehensive tests for new functionality
- Use descriptive commit messages with proper formatting
- Ensure code passes linting and all tests
- Update relevant documentation when adding features
# Install in development mode
pip install -e .
# Run tests
python -m pytest tests/
# Format code
black .
# Validate unified system
cd .claude/skills/unified-assessment-system
python test_runner.pyThis project is licensed under the MIT License - see the LICENSE file for details.
- π Project Homepage: https://agentpsy.com
- π¦ GitHub Repository: https://github.com/ptreezh/AgentPsyAssessment
- π Issue Tracking: GitHub Issues
- π¬ Discussions: GitHub Discussions
- π§ Email: contact@agentpsy.com
- π¬ WeChat: 3061176
For support, questions, or collaboration opportunities:
- π Check the User Manual first
- π Search existing GitHub Issues
- π¬ Join our GitHub Discussions
- π§ Contact us directly for urgent matters
- π§ Psychological Research Community - For foundational psychometric models and research
- π€ Open Source AI Community - For making powerful LLMs accessible
- π International Contributors - For multi-language support and cultural adaptation
- π Documentation Contributors - For comprehensive guides and tutorials
- π§ͺ Testing Community - For ensuring system reliability and validation
π Thank you for using AgentPsyAssessment v1.0!
Version: v1.0.0 | Updated: 2025-01-08 | Team: AgentPsyAssessment Development Team