Skip to content

Scoring and Data for ScholarQA-CS2 Benchmark #128

@pangxianghe

Description

@pangxianghe

Hello,

I am currently working on scoring the ScholarQA-CS2 benchmark and have a few questions regarding the process:

Scoring Query-Answer Pairs: Could you please clarify the correct function to use when scoring existing query-answer pairs? Which format should I organize my answers?

Discrepancy Between Paper and Code Rubrics: I noticed a discrepancy between the rubric example in the paper (Section H.3.2) and the code found at this GitHub link. The rubric in the paper mentions the query "how the AI hallucination is linked to the AI bias," but this query is not present in the data.json file. Additionally, the rubric in the paper includes a key called 'ingredients,' which is also absent in the data.json. It appears that the data.json is based on the original ScholarQA-Bench, and I am wondering where the data for ScholarQA-CS2 can be found.

Thank you in advance for your assistance!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions