Skip to content

Commit

Permalink
fix typos and supplement README
Browse files Browse the repository at this point in the history
  • Loading branch information
QianHaosheng committed Apr 13, 2024
1 parent c8fbf95 commit eb9a672
Show file tree
Hide file tree
Showing 2 changed files with 4 additions and 1 deletion.
3 changes: 3 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,9 @@ The generate task is to answer the question based on the contexts provided by re
* [Answer chrF Score](./rageval/metrics/_answer_chrf.py): also known as *character n-gram F-score* as used in [the paper (Popovic et al.)](https://aclanthology.org/W15-3049).
* [Answer Disambig-F1](./rageval/metrics/_answer_disambig_f1.py): also known as *Disambig-F1* as used in [the paper (Ivan Stelmakh et al.)](https://arxiv.org/abs/2204.06092) and [the paper (Zhengbao Jiang et al.)](https://arxiv.org/abs/2305.06983).
* [Answer Rouge Correctness](./rageval/metrics/_answer_rouge_correctness.py): also known as *Rouge* as used in [the paper (Chin-Yew Lin)](https://aclanthology.org/W04-1013.pdf).
* [Answer Accuracy](./rageval/metrics/_answer_accuracy.py): also known as *Accuracy* as used in [the paper (Dan Hendrycks et al.)](https://arxiv.org/abs/2009.03300).
* [Answer LCS Ratio](./rageval/metrics/_answer_lcs_ratio.py): also know as *LCS(%)* as used in [the paper (Nashid et al.)](https://ieeexplore.ieee.org/abstract/document/10172590).
* [Answer Edit Distance](./rageval/metrics/_answer_edit_distance.py): also know as *Edit distance* as used in [the paper (Nashid et al.)](https://ieeexplore.ieee.org/abstract/document/10172590).

(2) **Answer Groundedness**: this category of metrics is to evaluate the groundedness (also known as factual consistency) by comparing the generated answer with the provided contexts. Here are some commonly used metrics:

Expand Down
2 changes: 1 addition & 1 deletion tests/units/test_answer_accuracy.py
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ def testset(sample):


@pytest.mark.slow
def test_case_on_answer_lcs_ratio(testset):
def test_case_on_answer_accuracy(testset):
metric = AnswerAccuracy()
assert metric.name == "answer_accuracy"
assert metric.homepage == ""
Expand Down

0 comments on commit eb9a672

Please sign in to comment.