Releases: KRLabsOrg/LettuceDetect
Releases · KRLabsOrg/LettuceDetect
TinyLettuce + Integrations
v0.1.7: LettuceDetect multilingual + LLM baselines
What's Changed
- Preprocess RagBenchData & translate files by @nadiaverdha in #7
- Feature/translation ragbench by @adaamko in #9
- Add Web API and Python Client Library by @ceeli in #12
- Fix/Copying tensor by @lebe1 in #18
- Chat GPT and RAGAS Baseline by @nadiaverdha in #14
- Added seed to training, fixed encoder SEP token by @adaamko in #19
- Feature/llm baseline opt by @adaamko in #21
- Feature/cn llm eval by @adaamko in #22
New Contributors
- @nadiaverdha made their first contribution in #7
- @ceeli made their first contribution in #12
- @lebe1 made their first contribution in #18
Full Changelog: 0.1.6...0.1.7
0.1.6
Released 0.1.5, inference api changed, README changes
Release 0.1.5
0.1.4
0.1.3
Full Changelog: 0.1.2...0.1.3
0.1.2
Full Changelog: 0.1.1...0.1.2
0.1.1
Full Changelog: 0.1.0...0.1.1
0.1.0
First version of the model, results:
-- Token-Level Evaluation ----
Detailed Classification Report:
precision recall f1-score support
Supported 0.9799 0.9859 0.9829 422046
Hallucinated 0.6096 0.5222 0.5625 17844
accuracy 0.9671 439890
macro avg 0.7947 0.7540 0.7727 439890
weighted avg 0.9649 0.9671 0.9658 439890
Evaluation Results:
Hallucination Detection (Class 1):
Precision: 0.6096
Recall: 0.5222
F1: 0.5625
Supported Content (Class 0):
Precision: 0.9799
Recall: 0.9859
F1: 0.9829
---- Example-Level Evaluation ----
Detailed Example-Level Classification Report:
precision recall f1-score support
Supported 0.8696 0.8765 0.8730 1757
Hallucinated 0.7664 0.7550 0.7607 943
accuracy 0.8341 2700
macro avg 0.8180 0.8158 0.8168 2700
weighted avg 0.8335 0.8341 0.8338 2700
Example-Level Evaluation Results:
Hallucination Detection (Example Level) - Class 1:
Precision: 0.7664
Recall: 0.7550
F1: 0.7607
Supported Content (Example Level) - Class 0:
Precision: 0.8696
Recall: 0.8765
F1: 0.8730
Full Changelog: https://github.com/KRLabsOrg/LettuceDetect/commits/0.1.0