Skip to content

Questions for mode == "single_inference" #21

@KZF-kzf

Description

@KZF-kzf

Thank you very much for your work! Here are some of my questions, which I hope you can answer:

I want to know if this mode is suitable for evaluating the performance of editing models on a fixed test set. The scores output by this EditReward are usually between -3 and 2; I don’t know if this is normal (the comparison tables in the paper seem to be all percentages). The code shows that editreward will be evaluated from two dimensions, so why is the final value taken as reward[0][0].item? Does reward[0][1] have any practical meaning?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions