Skip to content

Computation of IAA in inceptalytics #23

@stbachinger

Description

@stbachinger

Hello :)
I am currently in the process of calculating different values of the IAA for an annotation project in inception. We used inceptalytics and just to make sure, I manually calculated Cohens kappa and Krippendorffs alpha. But the values I get from inceptalytics are vastly different than the ones from my manual calculation. There, I use the IOB files and add for each annotator the annotation to an individual array ( arr1 for annotator 1, arr2 for annotator 2, etc.). So for each token, there is an annotation. And those arrays are used to get the following results:

Manuell calculation:
Pair-wise Cohen-Kappa-Scores for all annotations (including O)
Ann1: Ann2 (pairwise Cohen-Kappa-Score) 0.27756503652905506
Ann1: Ann3 (pairwise Cohen-Kappa-Score) 0.12717909466991562
Ann2: Ann3 (pairwise Cohen-Kappa-Score) 0.24369264052470352
Krippendorff: 0.22980529268056016

Inceptalytics' calculations
Ann1: Ann2 (pairwise Cohen-Kappa-Score) 0.8176
Ann1: Ann3 (pairwise Cohen-Kappa-Score) 0.8185
Ann2: Ann3 (pairwise Cohen-Kappa-Score) 0.9121
Krippendorff: 0.8258

I pulled the newest version of inceptalytics and am using the template from examples/dashboard.py without changing anything.
My guess is that inceptalytics is using another way of creating the arrays from the source files, which would also explain why n in the Cohen's kappa is relatively low. But so far, I wasn't able to find out why that is.
I would be very grateful for any help or insights into the generation of input values for the IAA scores and the reasons why they were implemented the way they were.
Thank you so much!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions