Task description
This task aims to examine how well automated systems can evaluate summaries from different languages. This task takes as input the summaries generated from automatic systems and humans in the Summarization Tasks of MultiLing 2015, but also in the Single document summarization tasks of 2015 and 2017 (when the latter is completed). The output should be a grading of the summaries. Ideally, we would want the automatic evaluation to maximally correlate to human judgement, thus the evaluation will be based on correlation measurement between estimated grades and human grades.
Data
To be announced
Results
To be announced
Dates
To be announced
Revision created 2899 days ago by George Giannakopoulos (Admin)
Revision created 2900 days ago by George Giannakopoulos (Admin)
Revision created 2900 days ago by George Giannakopoulos (Admin)