-
Notifications
You must be signed in to change notification settings - Fork 15
Description
Hello,
I am currently working on scoring the ScholarQA-CS2 benchmark and have a few questions regarding the process:
Scoring Query-Answer Pairs: Could you please clarify the correct function to use when scoring existing query-answer pairs? Which format should I organize my answers?
Discrepancy Between Paper and Code Rubrics: I noticed a discrepancy between the rubric example in the paper (Section H.3.2) and the code found at this GitHub link. The rubric in the paper mentions the query "how the AI hallucination is linked to the AI bias," but this query is not present in the data.json file. Additionally, the rubric in the paper includes a key called 'ingredients,' which is also absent in the data.json. It appears that the data.json is based on the original ScholarQA-Bench, and I am wondering where the data for ScholarQA-CS2 can be found.
Thank you in advance for your assistance!