Skip to content

Question: Best practice for evaluating RAG models using the MiniRAG dataset with missing answers? #101

@Mihubaba

Description

@Mihubaba

I'm using the QA dataset from your open-source project (MiniRAG) to evaluate my own RAG question-answering framework. However, I noticed that some of the QA pairs in the dataset do not have reference answers.

I was wondering—when calculating the accuracy, should I exclude these questions without reference answers and only evaluate those that have standard answers? Or should I include all the questions and manually verify whether the model’s responses are correct?

I’d like to kindly ask for your advice on how to handle this situation. Thank you in advance!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions