1 |
Multilingual Surface Realization Using Universal Dependency Trees
|
|
|
|
BASE
|
|
Show details
|
|
2 |
Proceedings of the 6th Workshop on Vision and Language (VL'17)
|
|
|
|
BASE
|
|
Show details
|
|
3 |
Effect of Data Annotation, Feature Selection and Model Choice on Spatial Description Generation in French
|
|
|
|
BASE
|
|
Show details
|
|
4 |
Proceedings of the 2016 Workshop on Vision and Language (VL’16)
|
|
|
|
BASE
|
|
Show details
|
|
6 |
Weakly supervised construction of a repository of iconic images
|
|
|
|
BASE
|
|
Show details
|
|
7 |
Proceedings of UCNLG+Eval: Language Generation and Evaluation
|
|
|
|
BASE
|
|
Show details
|
|
8 |
Unsupervised alignment of comparable data and text resources
|
|
|
|
BASE
|
|
Show details
|
|
9 |
Discrete vs. continuous rating scales for language evaluation in NLP
|
|
|
|
BASE
|
|
Show details
|
|
10 |
Comparing rating scales and preference judgements in language evaluation
|
|
|
|
BASE
|
|
Show details
|
|
11 |
Assessing the trade-off between system building cost and output quality in data-to-text generation
|
|
|
|
BASE
|
|
Show details
|
|
13 |
Construction of bilingual multimodal corpora of referring expressions in collaborative problem solving
|
|
|
|
BASE
|
|
Show details
|
|
14 |
Finding common ground: towards a surface realisation shared task
|
|
|
|
BASE
|
|
Show details
|
|
15 |
Generating referring expressions in context: the GREC shared task evaluation challenges
|
|
|
|
BASE
|
|
Show details
|
|
16 |
Extracting parallel fragments from comparable corpora for data-to-text generation
|
|
|
|
BASE
|
|
Show details
|
|
17 |
The GREC named entity generation challenge 2009; Overview and evaluation results
|
|
|
|
BASE
|
|
Show details
|
|
19 |
The GREC main subject reference generation challenge 2009: Overview and evaluation results
|
|
|
|
Abstract:
The GREC-MSR Task at Generation Challenges 2009 required participating systems to select coreference chains to the main subject of short encyclopaedic texts collected from Wikipedia. Three teams submitted one system each, and we additionally created four baseline systems. Systems were tested automatically using existing intrinsic metrics. We also evaluated systems extrinsically by applying coreference resolution tools to the outputs and measuring the success of the tools. In addition,systems were tested in an intrinsic evaluation involving human judges. This report describes the GREC-MSR Task and the evaluation methods applied, gives brief descriptions of the participating systems, and presents the evaluation results.
|
|
Keyword:
G400 Computing; Q100 Linguistics
|
|
URL: http://aclweb.org/anthology-new/W/W09/W09-2816.pdf http://eprints.brighton.ac.uk/6943/
|
|
BASE
|
|
Hide details
|
|
20 |
An investigation into the validity of some metrics for automatically evaluating Natural Language Generation systems
|
|
|
|
BASE
|
|
Show details
|
|
|
|