google / wmt-mqm-human-evaluation Goto Github PK
View Code? Open in Web Editor NEWLicense: Apache License 2.0
License: Apache License 2.0
Hi,
I found out that the wmt 2020 zhen source file had missing columns: seg_id
, rater
. Please check.
Sincerely,
Hi,
I'm now trying to figure out how to compute pairwise inter-rater agreement, but my results are lower (avg: 0.279) than the results from the paper (avg. 0.584) for English→German MQM.
To compute the agreement, I followed the steps:
news2020
, news2021
, and ted
; increment doc_id
in news2021
and ted
by 1000
and 2000
to avoid the doc_id
overlap.s
to the 7-point liker-type score by mapping s==0
to 0
, 0<s<=5
to 1
, ..., 24.99<s<=25
to 6
sklearn.metrics.cohen_kappa_score
May I have any advice on computing the agreement? Or a sample script?
I appreciate any help you can provide.
Thank you for creating this amazing resource!
The annotator guidelines (Table 12 of your paper) contain the following instruction:
To identify an error, highlight the relevant span of text [...] The span of text may be in the source segment if the error is a source error or an omission.
In the dataset, all the spans are on the target side, including omission errors:
Source: Setting the example? Income inequality in the US is at an all-time high
Target: <v>Die</v> Einkommensungleichheit in den USA ist auf einem Allzeithoch
Category: Accuracy/Omission
I'm wondering: Did the annotators deviate from the guidelines in this respect, or is it maybe a data processing mistake that could still be fixed?
Hi,
When I look at the English to German table for Newstest2020, DA scores are exactly those from the official findings report here: https://statmt.org/wmt20/pdf/2020.wmt-1.1.pdf
For WMT21 scores and rankings are completely off (page 19 Table 10 of this: https://statmt.org/wmt21/pdf/2021.wmt-1.1.pdf )
Is there a specific reason ?
I found out mqm_newstest2020_zhen.tsv has minor data missing issue for the following systems at Seg ID 181.
181 Online-B.1605
181 WeChat_AI.1525
181 Tencent_Translation.1249
181 THUNLP.1498
181 OPPO.1422
181 Huoshan_Translate.919
This caused a mismatch in between mqm_newstest2020_zhen.tsv and mqm_newstest2020_zhen.avg_seg_scores.tsv
mqm_newstest2020_zhen.tsv only has 19994 unique system+id.
Thanks again for creating this great resource. I noticed some minor inconsistencies in the newstest2020 dataset that could be relevant for other people working with it:
Traces of Post-Editing
The raters did not just highlight error spans (as indicated by the guidelines) but also seem to have made post-edits. The "target" column sometimes contains text that deviates from the original translation. This affects 8255 out of 79020 ratings for EN–DE and 32184 out of 124292 ratings for ZH–EN.
Example:
ID: Human-A.0 | independent.281139 | 4 | 4
Rater: rater2
Original translation: Um auf Titelseiten zu gelangen, trug er einen Mundschutz und klebte sich Klebeband auf seine Nase, um Leute zum Reden zu bringen.
Target in the dataset: Um auf Titelseiten zu gelangen, trug <v>er angeblich </v> einen Mundschutz und klebte sich Klebeband auf seine Nase, um Leute zum Reden zu bringen.
Superfluous Quotes
A few translations are wrapped in quotes that are not present in the original data. The raters have usually marked the quotes as addition or punctuation errors, which has slightly increased the error count for these types. This affects at least 135 out of 14180 samples for EN–DE and 25 out of 19994 samples for ZH–EN.
Example:
ID: Online-A.1574 | stv.tv.21636 | 17 | 77
Original source: "This review is focused on steps that can be taken to help aid enforcement agencies such as local authorities, as they use their powers to help keep communities safe."
Source in the dataset: ""This review is focused on steps that can be taken to help aid enforcement agencies such as local authorities, as they use their powers to help keep communities safe.""
Original translation: "Diese Überprüfung konzentriert sich auf Maßnahmen, die ergriffen werden können, um Durchsetzungsbehörden wie lokalen Behörden zu helfen, da sie ihre Befugnisse nutzen, um die Sicherheit der Gemeinschaften zu gewährleisten."
Target in the dataset: """Diese Überprüfung konzentriert sich auf Maßnahmen, die ergriffen werden können, um Durchsetzungsbehörden wie lokalen Behörden zu helfen, da sie ihre Befugnisse nutzen, um die Sicherheit der Gemeinschaften zu gewährleisten."""
I've tried this but looks like I'm still missing ~20% of the data to their references: https://www.kaggle.com/code/alvations/lightyear2
Hi, I saw in 2020 you published sentence scores and they were really helpful for me. Thanks for that :)
I couldn't find theses scores for 2021 in the repo. Could you provide them? Or do you have the script to calculate them from the MQM evaluations. It would be of great help, Thanks
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.