Skip to content

cyberark/simple-llm-eval

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Simple LLM Evaluation

version Build Status Code Coverage PyPI - Python Version license OpenSSF Scorecard

Simpleval Banner

Welcome to the simple LLM evaluation frameworkโ€”simpleval, for short.

simpleval is a Python package designed to make evaluating Large Language Models (LLMs) easier, using the "LLM as a Judge" technique.

It supports a variety of LLM providers, including OpenAI, Google (Gemini API, Vertex), AWS Bedrock, Anthropic, Azure, and more (via LiteLLM).

simpleval also includes several reports to help you analyze, compare, and summarize your evaluation results. See the available reports for more details.

Getting Started

See the ๐Ÿ“š Quickstart Guide ๐Ÿ“š

Documentation

See ๐Ÿ“š Project Documentation ๐Ÿ“š

Contributing

We appreciate your help in making this project better! โœจ

If you would like to contribute to this project, please follow the guidelines outlined in the CONTRIBUTING.md file.

License

simpleval is released under the Apache License. See the LICENSE file for more details.

Contact

If you have any questions or suggestions, feel free to join our GitHub discussions forum ๐Ÿ’ฌ

If you want to report a bug or request a feature, please open an issue in the GitHub issues tracker ๐Ÿ›


About

Simple LLM Evaluation Using LLM As a Judge ๐Ÿ‘ฉโ€โš–๏ธ

Resources

License

Code of conduct

Security policy

Stars

Watchers

Forks

Packages

No packages published

Contributors 5