open research online methodologies in automati… · 1 introduction evaluation is a critical phase...

12
Open Research Online The Open University’s repository of research publications and other research outputs Evaluation methodologies in Automatic Question Generation 2013-2018 Conference or Workshop Item How to cite: Amidei, Jacopo; Piwek, Paul and Willis, Alistair (2018). Evaluation methodologies in Automatic Question Generation 2013-2018. In: Proceedings of The 11th International Natural Language Generation Conference, 5-8 Nov 2018, Tilburg, The Netherlands, pp. 307–317. For guidance on citations see FAQs . c 2018 Association for Computational Linguistics Version: Version of Record Copyright and Moral Rights for the articles on this site are retained by the individual authors and/or other copyright owners. For more information on Open Research Online’s data policy on reuse of materials please consult the policies page. oro.open.ac.uk

Upload: others

Post on 18-Jul-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Open Research Online methodologies in Automati… · 1 Introduction Evaluation is a critical phase for the development of Natural Language Generation (NLG) systems. It helps to improve

Open Research OnlineThe Open University’s repository of research publicationsand other research outputs

Evaluation methodologies in Automatic QuestionGeneration 2013-2018Conference or Workshop ItemHow to cite:

Amidei, Jacopo; Piwek, Paul and Willis, Alistair (2018). Evaluation methodologies in Automatic QuestionGeneration 2013-2018. In: Proceedings of The 11th International Natural Language Generation Conference, 5-8 Nov2018, Tilburg, The Netherlands, pp. 307–317.

For guidance on citations see FAQs.

c© 2018 Association for Computational Linguistics

Version: Version of Record

Copyright and Moral Rights for the articles on this site are retained by the individual authors and/or other copyrightowners. For more information on Open Research Online’s data policy on reuse of materials please consult the policiespage.

oro.open.ac.uk

Page 2: Open Research Online methodologies in Automati… · 1 Introduction Evaluation is a critical phase for the development of Natural Language Generation (NLG) systems. It helps to improve

Proceedings of The 11th International Natural Language Generation Conference, pages 307–317,Tilburg, The Netherlands, November 5-8, 2018. c©2018 Association for Computational Linguistics

307

Evaluation methodologies in Automatic Question Generation 2013-2018

Jacopo Amidei and Paul Piwek and Alistair WillisSchool of Computing and Communications

The Open UniversityMilton Keynes, UK

{jacopo.amidei, paul.piwek, alistair.willis}@open.ac.uk

Abstract

In the last few years Automatic QuestionGeneration (AQG) has attracted increasinginterest. In this paper we survey the evalu-ation methodologies used in AQG. Basedon a sample of 37 papers, our researchshows that the systems’ development hasnot been accompanied by similar develop-ments in the methodologies used for thesystems’ evaluation. Indeed, in the pa-pers we examine here, we find a wide va-riety of both intrinsic and extrinsic evalu-ation methodologies. Such diverse eval-uation practices make it difficult to reli-ably compare the quality of different gen-eration systems. Our study suggests that,given the rapidly increasing level of re-search in the area, a common frameworkis urgently needed to compare the perfor-mance of AQG systems and NLG systemsmore generally.

1 Introduction

Evaluation is a critical phase for the developmentof Natural Language Generation (NLG) systems.It helps to improve performance by highlightingweaknesses, and to identify new tasks to whichgeneration systems can be applied. Given thatgeneration systems and evaluation methodologiesshould be developed hand in hand, a system-atic study of evaluation methodologies for NLGshould take a central role in the effort of build-ing machines which are able to reach human-likelevels of linguistic communication. Such a studyshould investigate the current evaluation practicesused in various areas of NLG in order to seetheir weaknesses and suggest directions to im-prove them.

The aim of this paper is to analyze the evalu-ation methodologies used in Automatic QuestionGeneration (AQG) as a representative subtask ofNLG. To the best of our knowledge, since the in-troduction of the Question Generation Shared TaskEvaluation Challenge (QG-STEC) (Rus et al.,2010), no attempts have been made to introducea common framework for evaluation in AQG.

To approach this task, we examined the papersin the ACL anthology with a publication date be-tween the years 2013-2018 (more precisely Jan-uary 2013 to June 2018). Table 1 shows the distri-bution of the papers involved in the current studyacross this period. The ACL anthology websiterepresents a resource of inestimable value1 for thiswork.

Year of publication # papers2018 (Jan-June) 7 (so far)2017 132016 92015 52014 12013 2

Table 1: Number of papers per year describingquestion generation systems.

We used the single term question generation asthe search term with the search engine provided inthe ACL Anthology website. From the papers thatwere returned by this query, we focussed only onthose papers that were about question generationsystems. This gave us 37 papers to analyze, ofwhich 36 were published in conference proceed-ings and 1 was published in a journal. The num-ber of papers by year is given in Table 1 and illus-trated in Figure 1. Figure 1 indicates the rapid in-crease in publications in this area in recent years.

1http://aclweb.org/anthology/

Page 3: Open Research Online methodologies in Automati… · 1 Introduction Evaluation is a critical phase for the development of Natural Language Generation (NLG) systems. It helps to improve

308

Figure 1: Number of papers on AQG published byyear in the ACL anthology.

Note that this study of the literature was carriedout in June 2018, and so several major conferencesin this area (including ACL, INLG, EMNLP andCOLING) had not taken place.2

Publication type Journal or conference name # of papersConference proceed-ing

INLG 7

ACL 7NAACL-HLT 6Workshop on Innovative Useof NLP for Building Educa-tional Application

6

EMNLP 3IJCNLP 1EACL 1SIGDIAL 1COLING 1NLPTEA 1RANLP 1Workshop on RepresentationLearning for NLP

1

Journal Computational Linguistics 1

Table 2: Number of papers per conference pro-ceedings or journal.

Before looking more closely at the publicationsinvolved, let us introduce the AQG tasks studiedin these papers. AQG is the task

of automatically generating questionsfrom various inputs such as raw text,database, or semantic representation(Rus et al., 2008).

The above definition, adopted by the AQG com-munity, leaves room for researchers to decide whatkind of questions and input work with. FollowingPiwek and Boyer (2012) a particular AQG task canbe characterized by three aspects: the input, the

2A complete list of papers used in this study, as well asuseful information to reproduce the results presented in thepresent paper, can be found at the following link:

https://bit.ly/2IuPJIa

output, and finally the relationship between the in-put and the output. The 37 papers we analyzed canbe divided into the following three categories:

1. Input: text;Output: text;Relation: the output question is answered bythe input text or the output question asks aclarification question about the input text.

2. Input: knowledge base structured data (forexample triples 〈subject, object, subject/ob-ject relation〉);Output: text;Relation: the output question is answered bythe information structure in the input.

3. Input: image or image and text or image seg-mentation annotations;Output: text;Relation: the output question is answered bythe information pictured in the input.

For the sake of simplicity we will denote withText2Text the task expressed by category 1,Kb2Text the task expressed by category 2 and fi-nally Mm2Text the task expressed by category 3,where Mm is short for “Multi-modal”. Withineach category, we find papers with different aims.We show these in the following list, where thenumber in brackets shows how many papers fallinto that class:

1. Text2Text (30)

• Web searching (1)• Chatbot component (1)• Creation of comparative questions re-

lated to the input topic (1)• Clarification questions (1)• Question Answering (5)• Dataset creation purpose (1)• Educational purpose (9)• AQG general purposes (11)

2. Kb2Text (4)

• Question Answering (1)• Dataset creation purpose (1)• Educational purpose (1)• AQG general purposes (1)

3. Mm2Text (3)

Page 4: Open Research Online methodologies in Automati… · 1 Introduction Evaluation is a critical phase for the development of Natural Language Generation (NLG) systems. It helps to improve

309

• Data augmentation Visual Question An-swering (VQA) purpose (1)• AQG general purposes (2)

Regarding the papers in the Text2Text category, wefound some variety in the different types of output.Although in the majority of cases, the system’soutput was an interrogative sentence, there are 5papers in which the output is a “fill the gap” ques-tion, 3 papers where output is a multiple choicequestion (with its associated set of distractors) and3 papers in which the output is a question/answerpair. Also in both the Kb2Text and Mm2Text cate-gories there is 1 paper each in which the output isa question/answer pair. We also note that one pa-per in the Text2Text category developed a questiongenerator which takes a paragraph of text and anassociated answer as input. In this case, the gen-erated question must be answered by the answergiven in the input. We conclude this section byspecifying that AQG general purposes means thatthe system was not tied to a particular domain ortask-dependent setting, whereas Question Answer-ing means that the AQG system is developed inorder to be used in the Question Answering task.

2 Related Work

Two of the key references for evaluation in NLGare Krahmer and Theune (2010) and Gatt andKrahmer (2018). Both devote an entire sectionto evaluation. In particular, Section 7 of Gatt andKrahmer’s paper gives a helpful description of themethodologies used in NLG for the purpose ofevaluation, alongside examples and a discussionof the relevant problems.

Another highly relevant work is that of Gkatziaand Mahamood (2014). Gkatzia and Mahamoodstudied the use of evaluation methodologies forNLG, performing a study which analyzed a cor-pus of 79 conference and journal papers pub-lished between the years 2005-2014. Their resultsshow the increasing prevalence of automatic eval-uation over human evaluation and the prevalenceof intrinsic evaluation over extrinsic ones (we dis-cuss intrinsic and extrinsic methods in Section 3).Gkatzia and Mahamood also report that the eval-uation approaches are correlated with the publica-tion venue, so that papers published in the samejournal or conference tend to use the same evalu-ation methodologies. Our paper represents a con-tinuation and refinement of the Gkatzia and Ma-hamood paper, with our specific focus on AQG.

Regarding more specific work on AQG we referto Rakangor and Ghodasara (2015) and Le et al.(2014), both of which survey AQG, with the latterfocussing specifically on educational applicationsof AQG. For each paper considered, Rakangor andGhodasara present the methodology used, the gen-erated question type, the language of the generatedquestion, the evaluation methodologies and its re-sults. In contrast, Le et al. report on the educa-tional support type and the evaluation methodolo-gies, in one table and the question type and evalu-ation results in the other table. Although Le et al.present these results in two tables, the tables in facthave only one paper in common. In comparison,in this paper we focus all our attention on the eval-uation methodologies used. For this reason, wereport neither the systems’ specifications nor thesystems’ performance.

A final publication of importance to the cur-rent work is the report on the Question GenerationShared Task Evaluation Challenge (QG-STEC)(Rus et al., 2010). In the QG-STEC, two tasks, Aand B, were defined. Although both tasks sharedthe same output type, task A took a paragraph anda target question type as input, whereas task B tooka single sentence and a target question type as in-put. Both tasks were evaluated through a humanevaluation methodology, based on the 5 criteria:relevance, syntactic correctness and fluency, am-biguity, question type and variety. However, theInter Annotator Agreement (IAA) reached in theevaluation phase was low. An attempt to improvethe IAA for task B is described in Godwin and Pi-wek (2016), in which the authors define an inter-active process where the annotators discussed theiropinions about the criteria used in the evaluation.Although their method improves the IAA, the re-producibility of their results is not guaranteed.

3 Evaluation methodologies for AQG

In this section we present the findings of our anal-ysis. We focus our analysis on two dimensions: in-trinsic evaluation methodology and extrinsic eval-uation methodology.

Intrinsic evaluation methods measure the per-formance of a system by evaluating the system’soutput “in its own right, either against a refer-ence corpus or by eliciting human judgements ofquality” (Gatt and Belz, 2010, p. 264). For ex-ample, this could involve measuring the output’sgrammaticality and fluency. The prevailing intrin-

Page 5: Open Research Online methodologies in Automati… · 1 Introduction Evaluation is a critical phase for the development of Natural Language Generation (NLG) systems. It helps to improve

310

sic methods are human evaluation and automaticevaluation. In order to assess the quality of a gen-erated sentence, the former method uses humanjudgements, while the latter applies an algorithmthat automatically calculates a score, for exampleby checking the similarity between the generatedsentence and a set of reference sentences.

Extrinsic methods measure the performance ofa system by evaluating the system’s output withrespect to its ability to accomplish the task forwhich it was developed. An example of extrin-sic evaluation methods is that used to evaluate theSTOP system (Reiter et al., 2003). STOP gen-erates “short tailored smoking cessation letters,based on responses to a four-page smoking ques-tionnaire” (p. 41) with the aim of helping peopleto give up smoking. This system was evaluated“by recruiting 2553 smokers, sending 1/3 of themletters produced by STOP and the other 2/3 con-trol letters, and then measuring how many peoplein each group managed to stop smoking”3. In thiscase the system was evaluated in the real worldto see whether it has the desired effect, of help-ing people to quit smoking. The results showedthat there were no relevant differences between theSTOP letters and the control letters.

3.1 A general overviewTable 3 shows the evaluation methodologies usedin the papers that we examined. With respect to

Evaluation methodologies # of papersText2Text Kb2Text Mm2Text Total

Intrinsic humanonly

13 1 - 14

Intrinsic automaticonly

9 - 1 10

Extrinsic (human)only

2 - - 2

Intrinsic human &Intrinsic automatic

3 2 2 7

Intrinsic human &Extrinsic (human)

2 - - 2

Intrinsic automatic& Extrinsic (auto-matic)

1 - - 1

Intrinsic human &Intrinsic automatic& Extrinsic (auto-matic)

- 1 - 1

Table 3: Evaluation methodologies used.

the frequency of use of intrinsic compared to ex-trinsic methods, Table 3 confirms the trend iden-tified in Gkatzia and Mahamood (2014). Gkatziaand Mahamood found out that the 74.7% of thepapers used the intrinsic evaluation method. Inour analysis we found that 83% of the papers used

3See Ehud Reiter’s blog https://ehudreiter.com/2017/01/19/types-of-nlg-evaluation/

this methodology. However, we note that with re-spect to Gkatzia and Mahamood’s results, we havean inverted trend between the use of an extrinsicmethod compared to both intrinsic and extrinsic.Indeed, Gkatzia and Mahamood found that 15.2%of the papers used extrinsic methods, against the6% we get in our analysis, and 10.1% of the pa-pers used both methodologies, where our analysisshows that 11% of the papers use a combination ofboth.

Furthermore, our analysis confirms the trendbetween the use of automatic compared to humanintrinsic evaluation methodologies. In Gkatzia andMahamood (2014) the authors report that in 45.4%of the cases human evaluation is used, whereasin 38.2% of the cases automatic evaluation wereadopted. Similarly, our analysis shows that be-tween the papers that prefer intrinsic evaluationmethods, 45% used human evaluation, 32% usedautomatic evaluation and 23% used both humanand automatic evaluation.

Table 1 shows that in the period since 2016,there has been a considerable increase in the num-ber of publications in this area. It therefore makessense to ask whether this increase has been accom-panied with a change in the evaluation methodolo-gies used.

Evaluation methodologies # of papers2013-2015 2016-2018

Intrinsic human only 6 8Intrinsic automatic only 1 9Extrinsic (human) only - 2Intrinsic human & Intrinsic au-tomatic

1 6

Intrinsic human & Extrinsic (hu-man)

- 2

Intrinsic automatic & Extrinsic(automatic)

- 1

Intrinsic human & Intrinsic au-tomatic & Extrinsic (automatic)

- 1

Table 4: Variation of the evaluation methodolo-gies used between 2013 - 2015 and between 2016- 2018.

Table 4 shows how the range of evalua-tion methodologies used has changed. Betweenthe years 2013 - 2015 only intrinsic evaluationmethodologies were used – with 75% of papersusing human evaluation, 12.5% using automaticevaluation and 12.5% using both methodologies –for the years between 2016 - 2018 extrinsic evalu-ation methods have also been introduced. Indeed,although the majority of the papers in this period(79%) used intrinsic evaluation methods, 7% ofpapers used extrinsic evaluation methods and 14%

Page 6: Open Research Online methodologies in Automati… · 1 Introduction Evaluation is a critical phase for the development of Natural Language Generation (NLG) systems. It helps to improve

311

used both the methodologies. We can also see achange in the tendency to use intrinsic methods.Between the years 2016 - 2018, 35% of the papersused human evaluation (a decrease of 40% fromthe years between 2013 - 2015), 39% of the pa-pers used automatic evaluation (a 26.5% increaseon the years between 2013 - 2015) and 26% of thepapers used both methodologies (a 13.5% increaseon the years between 2013 - 2015).

3.2 Automatic evaluation

Table 5 presents a list of automatic metrics used inthe papers studied in the present research. Fromour analysis it turns out that the most used au-tomatic metric is BLEU followed by METEOR.Note that Table 5 only describes those that use thespecified metrics; other papers use metrics that aredefined for the specific aims described in the paperthat introduces them.

Evaluation methodologies # of papersText2Text Kb2Text Mm2Text Total

BLEU (Papineniet al., 2002)

8 3 2 13

METEOR (Baner-jee and Laviel,2005)

4 2 1 7

ROUGE (Lin andOch, 2004)

3 1 - 4

Precision 4 - - 4Recall 4 - - 4F1 4 - - 4Accuracy 2 0 1 3∆BLEU (Galleyet al., 2015)

- - 1 1

Embedding Greedy(Rus and Lintean.,2012)

- 1 - 1

Others 5 - - 5

Table 5: Automatic metrics used.

In our survey we found out that 31% of the pa-pers used just a single metric, whereas the other69% used more than one. The average is 2 metricsper paper, with a minimum of 1 metric (6 papers)and a maximum of 5 metrics (1 paper). In almost50% of cases (9 papers), 3 metrics were used. Wenoticed that only a single paper used an embed-ding based metric (see Sharma et al. (2017)). Ina majority of studies, word-overlap based metricswere used (see Sharma et al. (2017)).

In the last few years, many studies in NLGhave shed light on the correlation between hu-man judgement and automatic metrics. The re-sults, which have shown how this correspondenceis somewhat weak4, shed doubt on the feasibility

4For an in depth discussion of this point we refer to Reiterand Belz (2009) and to Gatt and Krahmer (2018), especiallysection 7.4.1 and the references presented there.

of using these metrics for evaluating the overallquality of a system.

To the best of our knowledge, the area of AQGis currently missing a study which aims to verifythe correlation between human judgement and au-tomatic metrics5. Such research would have twomerits: on one hand, this kind of meta-evaluationstudy would give a better characterisation of thegeneral problem. On the other hand, the researchcould provide guidance to researchers about whichmetric is most appropriate in evaluating a particu-lar model or system.

In conclusion, we believe that research in AQGwould benefit from a systematic study that aimsto clarify the relation between different evaluationmethodologies.

3.3 Human evaluation

Among the various human evaluation methodolo-gies, eliciting quality judgments is most common:human annotators are asked to assess the qualityof a question based on criteria such as question’sgrammaticality and fluency. Only two papers useda preference judgement methodology, in which thehuman annotators are asked either to assess pair-wise preference between questions or given a cou-ple of questions, one human generated and one au-tomatically generated, assess which one is auto-matically generated (or which one is the the hu-man generated). One of these papers also usedthe other methodology of eliciting quality judge-ments.

Quality judgment methodologies typically askannotators to use Likert or rating scales to recordtheir judgements. In our analysis, we found that56% of the papers used some kind of numericalscale. For example, human annotators were of-ten asked to assess the grammaticality of a ques-tion on a scale from 1 (worst) to 5 (best). On theother hand, 44% of the papers used a linguistic(or semantic) scale. In these cases, human annota-tors were typically asked to classify the questionsin some category such as coherent, somewhat co-herent or incoherent. The number of categoriesused in the Likert or rating scales by the papersthat adopted quality judgment methodologies areshown in Table 6.

Only three papers used more than one scale inthe evaluation. One of these uses a free scale in

5Yuan et al. (2017) raise some doubts about the capacityof BLEU to effectively measure the quality of systems usedin Text2Text tasks.

Page 7: Open Research Online methodologies in Automati… · 1 Introduction Evaluation is a critical phase for the development of Natural Language Generation (NLG) systems. It helps to improve

312

which the annotators have to choose a positive in-teger to count the inference steps necessary for an-swer a question.

Number of categories # of papersText2Text Kb2Text Mm2Text Total

2 6 - - 63 6 - 2 84 1 1 - 25 8 1 - 97 - 1 - 1

Table 6: Number of categories used in the Likertor rating scales.

Table 6 shows that the two most common num-ber of categories used in the Likert or rating scalesare 3 and 5. In a recent paper, Novikova et al.(2018) suggest that the use of a continuous scaleand relative assessments can improve the qualityof human judgments. Although in our study, wefound 2 papers that used relative assessment, wedid not find any papers that use a continuous scale.

Another interesting point is the number of anno-tators used in the evaluation. This number varies alot from paper to paper. We found a minimum of 1annotator (2 papers) to a maximum of 364 annota-tors (1 paper). Taking the papers which providedinformation on the number of annotators used (24papers), and removing five papers that used 53, 63,67, 81 and 364 annotators – these can be seen asoutliers – we found out that the average number ofannotators used was almost 4. The most commonnumber was two annotators, used by 29% (7 pa-pers) of the papers. 3 annotators were used by 17%(4 papers) and 4 annotators were used by 13% (3papers). The others paper used 5, 7, 8 or 10 anno-tators.

There is a similar breadth in the number of out-put questions used (that is, the questions gener-ated by the systems), and the criteria (that is, thequestion features to be checked) used in the eval-uation. The number of questions ranged from aminimum of 60 questions (1 paper) to a maximumof 2186 (1 paper). Amongst those papers whichactually provide this information (17 papers, or65%), we found out that the average number ofquestions used per paper is almost 493. 7 papers(27%) did not report this information, whereas 2papers (8%) report information about the amountof data from which the questions were generated,without giving the exact number of questions usedfor the evaluation.

Regarding the criteria used, we noticed that35% of the papers (8 studies) used an overall qual-

ity criterion, that is, a single criterion which wasused to evaluate the question’s overall quality. Onthe other hand, 52% of the papers (12 studies) usedspecific criteria, for example, question grammati-cality, question answerability, etc. A full list ofthese criteria is shown in Table 7. 13% of the pa-pers (3 studies) used both specific criteria and anoverall criterion. As Table 7 shows, there is a wideassortment of criteria used across the set of col-lected papers.

Criterion used # of papersText2Text Kb2Text Mm2Text Total

Grammaticality 7 - - 7Semantic correctness 4 - - 4Answer existence 3 - - 3Naturalness 2 1 - 3Question type 3 - - 3Clarity 3 - - 3Discriminator quality 3 - - 3Relevance 2 - - 2Correctness 2 - - 2Well-formedness 1 - - 1Key selection accuracy 1 - - 1Corrected retrieval 1 - - 1Fluency 1 - - 1Coherence 1 - - 1Timing 1 - - 1Inference step 1 - - 1Question diversity 1 - - 1Importance 1 - - 1Specificity 1 - - 1Predicate identification - 1 - 1Difficulty 1 - - 1Overall criterion 7 2 2 11

Table 7: Criteria used.

As we can see from Table 7, the specific cri-teria are mainly used in the Text2Text task. Justtwo criteria are used in the Kb2Text task and nonein the Mm2Text, where an overall quality criterionwas preferred. We note that some criteria, for ex-ample timing or importance, are specific to one ofthe aims of the paper in which they are used. In-deed, as shown in the introduction, we can finddifferent aims behind the papers’ motivations. Wenote that among the papers analyzed here, oftenonly little information is provided about the eval-uation guidelines6. We cannot exclude that, giventhe evaluation guidelines, some of the criteria pre-sented in Table 7 can be collapsed together. Thatis, it is possible that different researchers use dif-ferent names in order to check the same questionfeature. In order to have a better way to checkthe quality across systems, we suggest that re-searchers should publish the evaluation guidelinesused in the evaluation, as well as the quantitativeresults.

6Human evaluations are driven by some annotation guide-line which is a direct manifestation of some annotationscheme. Whereas the latter characterize the criteria to beevaluated, the first strictly define such criteria and suggesthow they should be evaluated.

Page 8: Open Research Online methodologies in Automati… · 1 Introduction Evaluation is a critical phase for the development of Natural Language Generation (NLG) systems. It helps to improve

313

Table 8 supplies an overview about the IAAreached in the human evaluations. We note that54% of the papers (14 studies) did not supply thisinformation. Only one of the two papers that usedpreference judgments reported the agreement be-tween evaluators. In that paper, Fleiss’ κ was usedto measure the IAA reached between 3 to 5 evalu-ators. The results, for three batches with differentevaluators and questions, were 0.242, 0.234 and0.182. Table 8 presents the IAA results reportedby the papers that used quality judgement meth-ods. Between the papers that reported this infor-mation, we found that the IAA was measured in26 cases and 9 of these were measured with twodifferent coefficients, for a total of 35 IAA values.The agreements were measured for specific crite-ria or for the overall quality criterion. In one casethe agreement over all the criteria was reported. It

Metric used for calculate IAA # of criteriameasured

Average Min. Max.

Cohen’s κ 14 0.46 0.10 0.80Krippendorff’s α 2 0.143 0.05 0.236Fleiss’s κ 4 0.45 0.33 0.62Pearson’s r 4 0.71 0.47 0.89Average measure 9 0.80 0.50 0.91k no better specified 2 0.085 0.08 0.09

Table 8: Measures of Inter-Annotator Agreement.

is notable that the agreement reached in the var-ious evaluations is generally quite low. Indeed,following Artstein and Poesio (2008), only agree-ment greater than or equal to 0.8 should be consid-ered. Quoting Artstein and Poesio, p. 591:

Both in our earlier work (Poesio andVieira 1998; Poesio 2004a) and in themore recent (Poesio and Artstein 2005)efforts we found that only values above0.8 ensures an annotation of reasonablequality. We therefore felt that if a thresh-old needs to be set, 0.8 is a good value.

Taking this as an appropriate quality threshold,among the papers that report IAA, very few eval-uations should be considered appropriate. Morespecifically, we found that only 23% (8 over 35values) of the evaluations reported IAA scores thatwere greater than or equal to 0.8.7

7Using the popular Krippendorff’s Kappa scales of inter-pretation (Krippendorff, 1980) – where any data annotationwith agreement in the interval [0.8, 1] should be consideredgood, agreement in the interval [0.67, 0.8) should be con-sidered tentative, and data annotation with agreement below0.67 should be discarded– we conclude that 43% (15 out of35) of the evaluations should be considered tentative.

Checking the agreement for number of annota-tors we found that in the case with 364 annotatorsthe IAA, measured for two criteria and a not betterspecified κ, was between 0.08 and 0.09. We foundonly 2 cases for 5 evaluators, which reported avalue of 0.05 for Krippendorff’s α and an averagemeasure of 0.89. Two papers used 4 annotators al-together: one reported a value of Krippendorff’sα of 0.236, with the other reporting a Pearson’s rof 0.71. Another paper used 3 evaluators and theFleiss’s κ to measure the IAA for 4 criteria. Theresults are reported in Table 8.

All other papers reporting an IAA measure werein evaluations that used 2 annotators. The resultsof these cases are shown in Table 8 highlightingCohen’s κ, the Average measure and Pearson’s r.8

There are sometimes attempts to design the ex-perimental methodology to improve the level ofIAA. In order to improve the agreement, one papercollapsed two score classes into one, whereas twopapers allowed a difference of one score betweenthe annotators rating. Two examples of the lattercase are the maximum value for Cohen’s κ and themaximum value for the average measure reportedin Table 8.

We conclude this section by noting that theproblem of a low IAA was present also in theShared Task Evaluation Challenge (QG-STEC)(Rus et al., 2010). In that case, an attempt to im-prove the IAA for task B was carried out by God-win and Piwek (2016). Godwin and Piwek definean interactive process in which the annotators candiscuss their opinions about the criteria used in theevaluation. At the end of the evaluation process,repeated three times with three annotators on dif-ferent data each time, they got high IAA with apeak of 0.94 for one of the five criteria used in theevaluation.

Although other papers (see for example Bayerland Paul (2011), Lommel et al. (2014) and HweeTou Ng and Foo (1999)) propose techniques whichaim to improve the IAA, in a recent paper (Amideiet al., 2018) we suggest thinking carefully aboutthis practice in the case of NLG tasks. Indeed, ifevaluation results have to inform generation sys-

8Regarding Pearson’s r, we should clarify that in the caseof 2 evaluators, IAA was measured for 3 criteria and not 4 asreported in table 8. However, because the Pearson’s r mea-sured for the fourth criterion was 0.71, that is the averagevalue, the Pearson’s r measure in the case of 2 evaluator isexactly the one shown in Table 8. Furthermore, for the aver-age measure there is a case with 5 annotators. Removing thatcase, the average for 2 annotators is 0.79.

Page 9: Open Research Online methodologies in Automati… · 1 Introduction Evaluation is a critical phase for the development of Natural Language Generation (NLG) systems. It helps to improve

314

tems developers of the extent to which they canimprove the communicative power of their sys-tems, the aim of attaining a higher IAA runs thedanger of biasing system developers towards ig-noring important aspects of human language. Anunchecked and unquestioned focus on the reduc-tion of disagreement among annotators runs thedanger of creating generation goals that rewardoutput that is more distant from, rather than closerto, natural human-like language.

3.4 Extrinsic evaluation

As shown in Table 3, extrinsic evaluation method-ologies are rare in the area. As reported by Gkatziaand Mahamood (2014) this is generally true forNLG tasks. Amongst the papers that have cho-sen to use this kind of evaluation technique, hu-man judges were used in 4 times out of the 6. Inthe papers where human judges were not used, theQuestion Generation (QG) system was tested asa component of a Question Answering (QA) sys-tem. The performance was evaluated by checkingthe difference between the QA system without theuse of the QG system against the performance ofthe QA system with the use of the QG system. Theaim of those papers was to improve QA systems bycreating more accurate question/answer pairs to beused for training purposes.

As a consequence of the different tasks in play,the other papers used humans in different ways.We can find tasks such as: answer the generatedquestions or use the generated questions in a webpage and then answer a survey about the utility ofthose questions. Or also: engage in a conversationwith a chatbot which involves a question-based di-alogue, and then rate the conversations.

Also in this case, the number of humans in-volved in the evaluation varies from paper to pa-per, ranging from 2 to 81. In contrast to the caseof intrinsic human evaluation, in this case the IAAis not reported. We note that human agreement inextrinsic evaluation is not as relevant as in the caseof intrinsic evaluation. Indeed, for intrinsic evalu-ations, agreement is required to have a reliabilityand validity measure of the evaluation scheme andguidelines (Artstein and Poesio, 2008). The agree-ment measure should gather evidence that differ-ent humans can make similar judgements aboutthe questions evaluated. This fact, following Krip-pendorff (2011) should allow us to answer thequestion of: “how much the resulting data can be

trusted to represent something real”? (page 1). Inhuman intrinsic evaluation, the agreement can beseen as a measure of the replicability of the results.For example, Carletta (1996, p. 1) wrote:

At one time, it was considered suffi-cient. . . to show examples based on theauthors’ interpretation. Research wasjudged according to whether or not thereader found the explanation plausible.Now, researchers are beginning to re-quire evidence that people besides theauthors themselves can understand andmake the judgments underlying the re-search reliably. This is a reasonable re-quirement because if researchers can’teven show that different people canagree about the judgments on whichtheir research is based, then there is nochance of replicating the research re-sults.

In the case of extrinsic methods, the evaluationaim is to check if the generated sentences fulfil thetask for which they were generated. To test this,humans need to use those sentences in real con-texts. Now, humans make use of the same toolsin different ways, and similarly they answer ques-tions in different ways. For this reason, it is not ex-pected that humans reach similar results in a realcontext of language use.

4 Discussion

Although systems and tools have been developedin the AQG area over the last few years, Table1 illustrates that this has not been accompaniedby similar improvements in evaluation methodolo-gies. Indeed, with the exception of the SharedTask Evaluation Challenge (QG -STEC) (Ruset al., 2010), no attempts have been undertaken tointroduce a common framework for evaluation thatallows for comparisons between systems.

We have seen that in human evaluation, differ-ent criteria and scales/categories are used. To ad-dress this, we recommend that researchers sharetheir evaluation guidelines, and work towardsadopting common guidelines that can used tocheck quality across systems. Furthermore, out ofthe papers examined here, the problem of evalua-tion validity emerges. In those studies where it hasactually been reported, the IAA is generally low.Also in this case we suggest researchers systemat-ically determine the IAA and share their results, as

Page 10: Open Research Online methodologies in Automati… · 1 Introduction Evaluation is a critical phase for the development of Natural Language Generation (NLG) systems. It helps to improve

315

well as ideas to attempt to understand and classifyany divergences between annotators.

Automatic evaluation can be thought of as atechnique to provide a way to standardize the eval-uation. Unfortunately, a comparison of humanand automatic evaluation is missing in the area.This makes it difficult to understand to what extentthe automatic metrics capture the systems’ quality.Lacking such comparison, and following Reiter(2018), we suggest considering metrics such asBLEU as tools for systems’ diagnostic more thanevaluation techniques able to measure the outputquality of the systems.

There is scope for more extrinsic evaluation,which can “provide useful insight of domains’need, and thus they provide better indications ofthe systems’ usefulness and utility” (Gkatzia andMahamood, 2014, p. 60). Unfortunately, extrinsicevaluations are not yet widely used.

Though the QG-STEC evaluation scheme hasonly limited uptake, with the much increased pop-ularity of AQG, it is timely to revisit and addressthe need for a shared evaluation scheme. The va-riety of evaluation methodologies, as brought tolight by the present work, demonstrates how diffi-cult it currently is to check question quality acrossgeneration systems. This prevents us from under-standing the actual contributions that are made bynew generation systems that are being introducedever more frequently.

We conclude this section with the following ob-servation. The problem of having a high degree ofvariation in methodologies is compounded by theuse of different datasets in the evaluation phase(see Table 9). The use of a common dataset forevaluation – as suggested by the Shared Task Eval-uation Campaign (STEC) (Gatt and Belz, 2010) –could remove bias coming from the training phase.This is particularly true for generation systems thatuse machine learning techniques. We note that thehigh variability in the dataset used in the evalua-tion phase is also due to the variation in the pa-pers’ motivations. However, Table 9 suggests thatthe aim of building a common framework for AQGtasks should involve creation of a dataset to beused only for evaluation purposes. If we want tounderstand the degree to which a system advancesthe state of the art, we need to compare differ-ent systems on the same dataset, or better, a setof datasets, of course, using the same evaluationmethodologies.

Tasks Dataset or source of test articlesText2Text SQuAD; MS-MARCO; Wik-

iQA; TriviaQA; TrecQA;Wikinews; Penn Treebank; QG-STEC datasets; StackExchange;Wikipedia; OMG! website;Project Gutemberg; Read-Works.org; Engarde corpus;CrunchBase; Newswire (Prop-Bank); textbook from OpenStaxand Saylor; not specify TOEFLbook; not specify science textbooks; not specify course Webpage; not specify news articlesnot specify teachers articles; 40people’s personal data.

Kb2Text Ontology documenting K-12 Bi-ology concepts; SimpleQues-tions; Freebase; WikiAnswers.

Mm2Text COCO-QA; COCO-VQA;IGCCrowd; Bing; COCO; Flickr.

Table 9: Dataset used.

An open evaluation platform in which re-searchers share their evaluation methodologiesand their results can be effective to compare thequality across systems. In such a platform, theshared evaluation methodologies, alongside somedatasets used only for evaluation, can be used byresearchers to test their systems’ performance andthe results can be recorded in the open platform.Another benefit of this platform could be to gener-ate an evolutionary process which allows the com-munity to select the evaluation methodologies thatare considered more effective.

5 Conclusion

In this paper we have analysed 37 papers whichwere about AQG. The aim of our work was tostudy the evaluation methodologies used in thearea. Our work confirms the conclusion of Gkatziaand Mahamood (2014) for NLG in general. InAQG we lack a standardised approach for evalu-ating generation systems. Indeed, our overviewshows a quite variegated evaluation landscapewhich prevents comparison of question qualityacross generation systems. A careful look at thepapers published in the AQG area in the last fiveyears shows how little attention has been given tothe evaluation methodology introduced in the QG-STEC. Given the ever-increasing number of publi-cations in the area, a common framework for test-ing the performance of generation systems is ur-gently needed.

Acknowledgments

We warmly thanks the anonymous reviewers fortheir helpful suggestions.

Page 11: Open Research Online methodologies in Automati… · 1 Introduction Evaluation is a critical phase for the development of Natural Language Generation (NLG) systems. It helps to improve

316

ReferencesJacopo Amidei, Paul Piwek, and Alistair Willis. 2018.

Rethinking the agreement in human evaluationtasks. In Proceedings of the 27th International Con-ference on Computational Linguistics, pages 3318–3329.

Ron Artstein and Massimo Poesio. 2008. Inter-coderagreement for computational linguistics. Computa-tional Linguistics, 34(4):555–596.

Satanjeev Banerjee and Alon Laviel. 2005. Meteor:An automatic metric for MT evaluation with im-proved correlation with human judgments. In Pro-ceedings of the ACL Workshop on Intrinsic and Ex-trinsic Evaluation Measures for Machine Transla-tion and/or Summarization, pages 65–72.

Petra S. Bayerl and Karsten I. Paul. 2011. What deter-mines inter-coder agreement in manual annotations?a meta-analytic investigation. Computational Lin-guistics, 37(4):699–725.

Jean Carletta. 1996. Assessing agreement on classi-fication tasks: The kappa statistic. ComputationalLinguistics, 22(2):249–254.

Michel Galley, Chris Brockett, Alessandro Sordoni,Yangfeng Ji, Michael Auli, Chris Quirk, Mar-garet Mitchell, Jianfeng Gao, and Bill Dolan. 2015.δbleu: A discriminative metric for generation taskswith intrinsically diverse targets. In Proceedingsof the 53rd Annual Meeting of the Association forComputational Linguistics and the 7th InternationalJoint Conference on Natural Language Processing,pages 26–31.

Albert Gatt and Anja Belz. 2010. Introducing sharedtasks to NLG: The TUNA shared task evaluationchallenges. in E. Krahmer and Mariet Theune(Eds.), Empirical Methods in Natural LanguageGeneration Springer-Verlag, Berlin Heidelberg.

Albert Gatt and Emiel Krahmer. 2018. Survey of thestate of the art in natural language generation: Coretasks, applications and evaluation. Journal of Artifi-cial Intelligence Research, 61(1):65–170.

Dimitra Gkatzia and Saad Mahamood. 2014. A snap-shot of NLG evaluation practices 2005 - 2014. InProceedings of the 15th European Workshop on Nat-ural Language Generation (ENLG), pages 57–60.

Keith Godwin and Paul Piwek. 2016. Collecting reli-able human judgements on machine-generated lan-guage: The case of the qg-stec data. In ProceedingsINLG16, pages 212–216.

Chung Yong Lim Hwee Tou Ng and Shou King Foo.1999. A case study on inter-annotator agreement forword sense disambiguation. In Proceedings of theACL SIGLEX Workshop: Standardizing Lexical Re-sources, pages 9–13.

Emiel Krahmer and Mariet Theune (Eds.). 2010. Em-pirical Methods in Natural Language Generation.Springer-Verlag, Berlin Heidelberg.

Klaus Krippendorff. 1980. Content Analysis: An In-troduction to Its Methodology. Sage Publications,Beverly Hills, CA.

Klaus Krippendorff. 2011. Computing Krip-pendorff’s alpha-reliability. Retrieved fromhttp://repository.upenn.edu.

Nguyen-Thinh Le, Tomoko Kojiri, and Niels Pinkwart.2014. Automatic Question Generation for Educa-tional Applications The State of Art. In: van Do T.,Thi H., Nguyen N. (eds) Advanced ComputationalMethods for Knowledge Engineering. Springer,Cham.

Chin-Yew Lin and Franz Josef Och. 2004. Automaticevaluation of machine translation quality using us-ing longest common subsequence and skip-bigramstatistics. In Proceedings ACL’04, pages 605–612.

Arle Lommel, Maja Popovic, and Aljoscha Burchardt.2014. Assessing inter-annotator agreement fortranslation error annotation. In: MTE: Workshopon Automatic and Manual Metrics for OperationalTranslation Evaluation, Reykjavik, Iceland.

Jekaterina Novikova, Ondrrej Dusek, and VerenaRieser. 2018. RankMe: Reliable human ratingsfor natural language generation. In Proceedings ofNAACL-HLT, pages 72–78.

Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. BLEU: a method for automaticevaluation of machine translation. In Proceedingsof the 40th Annual meeting of the Association forComputational Linguistics., pages 311–318.

Paul Piwek and Kristy. E. Boyer. 2012. Varieties ofquestion generation: Introduction to this special is-sue. dialogue and discourse. Dialogue and Dis-course, 3(2):1–9.

Sheetal Rakangor and Y. R. Ghodasara. 2015. Liter-ature review of automatic question generation sys-tems. International Journal of Scientific and Re-search Publications, 5(1):1–5.

Ehud Reiter. 2018. A structured review of the validityof bleu. Computational Linguistics, 44(3):393–401.

Ehud Reiter and Anja Belz. 2009. An investigation intothe validity of some metrics for automatically evalu-ating natural language generation systems. Compu-tational Linguistics, 35(4):529–558.

Ehud Reiter, Roma Robertson, and Liesl M.Osman.2003. Lessons from a failure: Generating tailoredsmoking cessation letters. Artificial Intelligence,144(1-2):41–58.

Page 12: Open Research Online methodologies in Automati… · 1 Introduction Evaluation is a critical phase for the development of Natural Language Generation (NLG) systems. It helps to improve

317

Vasile Rus, Zhiqiang Cai, and Art Graesser. 2008.Question generation: Example of a multi-year eval-uation campaign. In Rus, V. and A. Graesser(eds.), online Proceedings of 1st Question Genera-tion Workshop, pages 25–26.

Vasile Rus and Mihai C. Lintean. 2012. A compari-son of greedy and optimal assessment of natural lan-guage student input using word-to-word similaritymetrics. In In Proceedings of the Seventh Workshopon Building Educational Applications Using NLP.

Vasile Rus, Brendan Wyse, Paul Piwek, Mihai Lintean,Svetlana Stoyanchev, and Cristian Moldovan. 2010.The first question generation shared task evaluationchallenge. In Proceedings of the Sixth InternationalNatural Language Generation Conference (INLG),pages 7–9.

Shikhar Sharma, Layla El Asri, Hannes Schulz, andJeremie Zumer. 2017. Relevance of unsupervisedmetrics in task-oriented dialogue for evaluating nat-ural language generation. arXiv:1706.09799.

Xingdi Yuan, Tong Wang, Caglar Gulcehre, Alessan-dro Sordoni, Philip Bachman, and Saizheng Zhang.2017. Machine comprehension by text-to-text neu-ral question generation. In Proceedings of the2nd Workshop on Representation Learning for NLP,pages 15–25.