ner for south and south east asian languagesltrc.iiit.ac.in/ner-ssea-08/proc/nersseal.pdf ·...

121
IJCNLP-08 Workshop On NER for South and South East Asian Languages Proceedings of the Workshop 12 January 2008 IIIT, Hyderabad, India

Upload: others

Post on 18-Jul-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

  • IJCNLP-08 Workshop

    On

    NER for South and SouthEast Asian Languages

    Proceedings of the Workshop

    12 January 2008IIIT, Hyderabad, India

  • c©2008 Asian Federation of Natural Language Processing

    ii

  • Introduction

    Welcome to the IJCNLP Workshop on Named Entity Recognition for South and South East AsianLanguages, a meeting held in conjunction with the Third International Joint Conference on NaturalLanguage Processing at Hyderabad, India. The goal of this workshop is to ascertain the state of the artin Named Entity Recognition (NER) specifically for South and South East Asian (SSEA) languages.This workshop continues the work started in the NLPAI Machine Learning Contest 2007 which wasfocused on NER for South Asian languages. NER was selected this time for the contest as well as forthis workshop because it is one of the fundamental and most important problems in NLP for whichsystems with good accuracy have not been built so far for SSEA languages. The primary reason forthis is that the characteristics of SSEA languages relevant for NER are different in many respects fromEnglish, on which a lot of work has been done with a significant amount of success in the last fewyears. An introductory article further explains the background of and motivation for this workshop. Italso presents the results of an experiment on a reasonable baseline and compares the results obtainedby the participating teams with the results for this baseline.

    The workshop had two tracks: One track for regular research papers on NER for SSEA languagesand the second track on the lines of a shared task. The workshop attracted a lot of interest, especiallyfrom the South Asian region. Participation from most of the research centers in South Asia working onNER ensured that the workshop met its goal of ascertaining and advancing the state of the art in NERfor SSEA languages. Another major achievement was that a good quantity of named entity annotatedcorpus was created in five South Asian languages. The notable point about this effort was that this wasdone almost informally on a voluntary basis, without funding. This is an important point in the contextof SSEA languages because lack of annotated corpora has held back progress in many areas of NLP sofar in this region.

    Each paper was reviewed by three reviewers to ensure satisfactory quality of the selected papers.Another major feature of the workshop is that it includes two invited talks by senior researchers workingon the NER problem for South Asian languages. The only drawback of the workshop was that therewas no paper on any South East Asian language.

    We would like to thank the program committee members for all the hard work that they did duringthe reviewing process. We would also like to thank all the people involved in organizing the IJCNLPconference. We hope that this workshop will help in creating interest in NER for SSEA languages andwe will soon be able to achieve results comparable to those for languages like English.

    Rajeev Sangal, Dipti Misra Sharma and Anil Kumar Singh (Chairs)

    iii

  • Organizers:

    Rajeev Sangal, IIIT, Hyderabad, IndiaDipti Misra Sharma, IIIT, Hyderabad, Hyderabad, IndiaAnil Kumar Singh, IIIT, Hyderabad, India

    Program Committee:

    Rajeev Sangal, IIIT, Hyderabad, IndiaDekai Wu, The Hong Kong University of Science & Technology, Hong KongTed Pedersen, University of Minnesota, USADipti Misra Sharma, IIIT, Hyderabad, Hyderabad, IndiaVirach Sornlertlamvanich, TCL, NICT, ThailandAlexander Gelbukh, Center for Computing Research, National Polytechnic Institute, MexicoM. Sasikumar, CDAC, Mumbai, IndiaSudeshna Sarkar, Indian Institute of Technology, Kharagpur, IndiaThierry Poibeau, CNRS, FranceSobha L., AU-KBC, Chennai, IndiaTzong-Han Tsai, National Taiwan University, TaiwanPrasad Pingali, IIIT, Hyderabad, IndiaCanasai Kreungkrai, NICT, JapanManabu Sassano, Yahoo Japan Corporation, JapanKavi Narayana Murthy, University of Hyderabad, IndiaSivaji Bandyopadhyay, Jadavpur University, Kolkata, IndiaAnil Kumar Singh, IIIT, Hyderabad, Hyderabad, IndiaDoaa Samy, Universidad Autnoma de Madrid, SpainRatna Sanyal, IIIT, Allahabad, IndiaV. Sriram, IIIT, Hyderabad, Hyderabad, IndiaAnagha Kulkarni, Carnegie Mellon University, USASoma Paul, IIIT, Hyderabad, Hyderabad, IndiaSofia Galicia-Haro, National Autonomous University, MexicoGrigori Sidorov, National Polytechnic Institute, Mexico

    Special Acknowledgment:

    Samar Husain, IIIT, Hyderabad, IndiaHarshit Surana, IIIT, Hyderabad, India

    Invited Speakers:

    Sobha L., AU-KBC, Chennai, IndiaSivaji Bandyopadhyay, Jadavpur University, Kolkata, India

    v

  • Table of Contents

    Invited Talk: Named Entity Recognition: Different ApproachesSobha L . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

    Invited Talk: Multilingual Named Entity RecognitionSivaji Bandyopadhyay . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3

    Named Entity Recognition for South and South East Asian Languages: Taking StockAnil Kumar Singh . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

    A Hybrid Named Entity Recognition System for South and South East Asian LanguagesSujan Kumar Saha, Sanjay Chatterji, Sandipan Dandapat, Sudeshna Sarkar and Pabitra Mitra . 17

    Aggregating Machine Learning and Rule Based Heuristics for Named Entity RecognitionKarthik Gali, Harshit Surana, Ashwini Vaidya, Praneeth Shishtla and Dipti Misra Sharma. . . . .25

    Language Independent Named Entity Recognition in Indian LanguagesAsif Ekbal, Rejwanul Haque, Amitava Das, Venkateswarlu Poka and Sivaji Bandyopadhyay . . 33

    Named Entity Recognition for TeluguP Srikanth and Kavi Narayana Murthy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41

    Bengali Named Entity Recognition Using Support Vector MachineAsif Ekbal and Sivaji Bandyopadhyay . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51

    Domain Focused Named Entity Recognizer for Tamil Using Conditional Random FieldsVijayakrishna R and Sobha L . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59

    A Character n-gram Based Approach for Improved Recall in Indian Language NERPraneeth M Shishtla, Prasad Pingali and Vasudeva Varma . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67

    An Experiment on Automatic Detection of Named Entities in BanglaBidyut Baran Chaudhuri and Suvankar Bhattacharya . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75

    Hybrid Named Entity Recognition System for South and South East Asian LanguagesPraveen P and Ravi Kiran V . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83

    Named Entity Recognition for South Asian LanguagesAmit Goyal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89

    Named Entity Recognition for Indian LanguagesAnimesh Nayan, B. Ravi Kiran Rao, Pawandeep Singh, Sudip Sanyal and Ratna Sanyal . . . . . . 97

    Experiments in Telugu NER: A Conditional Random Field ApproachPraneeth M Shishtla, Karthik Gali, Prasad Pingali and Vasudeva Varma . . . . . . . . . . . . . . . . . . . . 105

    vii

  • viii

  • Workshop Program

    Saturday, January 12, 2008

    Session 1:

    09:00-9:30 Opening Remarks: Named Entity Recognition for South and South East AsianLanguages: Taking StockAnil Kumar Singh

    09:30-10:00 Invited Talk: Named Entity Recognition: Different ApproachesSobha L

    10:00-10:30 A Hybrid Named Entity Recognition System for South and South East Asian Lan-guagesSujan Kumar Saha, Sanjay Chatterji, Sandipan Dandapat, Sudeshna Sarkar andPabitra Mitra

    10:30-11:00 Break

    Session 2:

    11:00-11:30 Invited Talk: Multilingual Named Entity RecognitionSivaji Bandyopadhyay

    11:30-12:00 Aggregating Machine Learning and Rule Based Heuristics for Named Entity Recog-nitionKarthik Gali, Harshit Surana, Ashwini Vaidya, Praneeth Shishtla and Dipti MisraSharma

    12:00-12:30 Language Independent Named Entity Recognition in Indian LanguagesAsif Ekbal, Rejwanul Haque, Amitava Das, Venkateswarlu Poka and Sivaji Bandy-opadhyay

    12:30-14:00 Lunch

    Session 3:

    14:00-14:30 Named Entity Recognition for TeluguP Srikanth and Kavi Narayana Murthy

    14:30-15:00 Poster Display and Discussion

    An Experiment on Automatic Detection of Named Entities in BanglaBidyut Baran Chaudhuri and Suvankar Bhattacharya

    ix

  • Saturday, January 12, 2008 (continued)

    Hybrid Named Entity Recognition System for South and South East Asian LanguagesPraveen P and Ravi Kiran V

    Named Entity Recognition for South Asian LanguagesAmit Goyal

    Named Entity Recognition for Indian LanguagesAnimesh Nayan, B. Ravi Kiran Rao, Pawandeep Singh, Sudip Sanyal and Ratna Sanyal

    Experiments in Telugu NER: A Conditional Random Field ApproachPraneeth M Shishtla, Karthik Gali, Prasad Pingali and Vasudeva Varma

    15:30-16:00 Break

    Session 4:

    16:00-16:30 Bengali Named Entity Recognition Using Support Vector MachineAsif Ekbal and Sivaji Bandyopadhyay

    16:30-17:00 Domain Focused Named Entity Recognizer for Tamil Using Conditional Random FieldsVijayakrishna R and Sobha L

    17:00-17:30 A Character n-gram Based Approach for Improved Recall in Indian Language NERPraneeth M Shishtla, Prasad Pingali and Vasudeva Varma

    17:30-18:00 Closing Discussion

    x

  • Proceedings of the IJCNLP-08 Workshop on NER for South and South East Asian Languages, pages 1–2,Hyderabad, India, January 2008. c©2008 Asian Federation of Natural Language Processing

    Named Entity Recognition: Different Approaches

    Sobha, L AU-KBC Research Centre

    MIT Campus of Anna University Chennai-44

    [email protected]

    Abstract

    The talk deals with different approaches used for Named Entity recognition and how they are used in developing a robust Named Entity Recognizer. The talk includes the development of tagset for NER and manual annotation of text.

    1

  • 2

  • Proceedings of the IJCNLP-08 Workshop on NER for South and South East Asian Languages, pages 3–4,Hyderabad, India, January 2008. c©2008 Asian Federation of Natural Language Processing

    Multilingual Named Entity Recognition

    Sivaji Bandyopadhyay Computer Science & Engineering Department

    Jadavpur University Kolkata, INDIA.

    [email protected]

    Abstract

    The computational research aiming at automatically identifying named entities (NE) in texts forms a vast and heterogeneous pool of strategies, techniques and representations from hand-crafted rules towards ma-chine learning approaches. Hand-crafted rule based systems provide good performance at a relatively high system engineering cost. The availability of a large collection of annotated data is the prerequisite for us-ing supervised learning techniques. Semi-supervised and unsupervised learning techniques promise fast deployment for many NE types without the prerequisite of an annotated corpus. The main technique for semi-supervised learning is called bootstrapping and involves a small degree of supervision, such as a set of seeds, for starting the learning process. The typical approach in unsupervised learning is clustering where systems can try to gather NEs from clustered groups based on the similarity of context. The tech-niques rely on lexical resources (e.g., Wordnet), on lexical patterns and on statistics computed on a large unannotated corpus.

    In multilingual named entity recognition (NER), it must be possible to use the same method for many different languages and the extension to new languages must be easy and fast. Person names can be rec-ognized in text through a lookup procedure, by analyzing the local lexical context, by looking at part of a sequence of candidate words that is a known name component etc. Some organization names can be iden-tified by looking at contain organization-specific candidate words. Identification of place names necessar-ily involves lookup against a gazetteer, as most context markers are too weak and ambiguous. An important feature in multilingual person name detection is that the same person can be referred to by different name variants. The main reasons for these variations are: the reuse of name parts to avoid repeti-tion, morphological variants such as the added suffixes, spelling mistakes, adaptation of names to local spelling rules, transliteration differences due to different transliteration rules or different target languages etc.. Name variants can be found within the same language documents. The major challenges for looking up place names in a multilingual gazetteer are the following: place names are frequently homographic with common words or with person names, presence of a number of exonyms (foreign language equivalences), endonyms (local variants) and historical variants for many place names etc..

    Application of NER to multilingual document sets helps to find more and more accurate informa-tion on each NE, while at the same time rich in-formation about NEs is helpful and can even be a crucial ingredient for text analysis applications that cross the language barrier.

    3

  • 4

  • Proceedings of the IJCNLP-08 Workshop on NER for South and South East Asian Languages, pages 5–16,Hyderabad, India, January 2008. c©2008 Asian Federation of Natural Language Processing

    Named Entity Recognition for South and South East Asian Languages:Taking Stock

    Anil Kumar SinghLanguage Technologies Research Centre

    IIIT, Hyderabad, [email protected]

    Abstract

    In this paper we first present a brief discus-sion of the problem of Named Entity Recog-nition (NER) in the context of the IJCNLPworkshop on NER for South and South EastAsian (SSEA) languages1. We also present ashort report on the development of a namedentity annotated corpus in five South Asianlanguage, namely Hindi, Bengali, Telugu,Oriya and Urdu. We present some detailsabout a new named entity tagset used for thiscorpus and describe the annotation guide-lines. Since the corpus was used for a sharedtask, we also explain the evaluation mea-sures used for the task. We then presentthe results of our experiments on a baselinewhich uses a maximum entropy based ap-proach. Finally, we give an overview of thepapers to be presented at the workshop, in-cluding those from the shared task track. Wediscuss the results obtained by teams partic-ipating in the task and compare their resultswith the baseline results.

    1 Introduction

    One of the motivations for organizing a workshop(NERSSEAL-08) focused on named entities (NEs)was that they have a special status in Natural Lan-guage Processing (NLP) because they have someproperties which other elements of human languagesdo not have, e.g. they refer to specific things or con-cepts in the world and are not listed in the grammars

    1http://ltrc.iiit.ac.in/ner-ssea-08

    or the lexicons. Identifying and classifying them au-tomatically can help us in processing text becausethey form a significant portion of the types and to-kens occurring in a corpus. Also, because of theirvery nature, machine learning techniques have beenfound to be very useful in identifying them. In orderto use these machine learning techniques, we needcorpus annotated with named entities. In this paperwe describe such a corpus developed for five SouthAsian languages. These languages are Hindi, Ben-gali, Oriya, Telugu and Urdu.

    This paper also presents an overview of the workdone for the IJCNLP workshop on NER for SSEAlanguages. The workshop included two tracks. Thefirst track was for regular research papers, while thesecond was organized on the lines of a shared task.

    Fairly mature named entity recognition systemsare now available for European languages (Sang,2002; Sang and De Meulder, 2003), especially En-glish, and even for East Asian languages (Sassanoand Utsuro, 2000). However, for South and SouthEast Asian languages, the problem of NER is still farfrom being solved. Even though we can gain muchinsight from the methods used for English, there aremany issues which make the nature of the problemdifferent for SSEA languages. For example, theselanguages do not have capitalization, which is a ma-jor feature used by NER systems for European lan-guages.

    Another characteristic of these languages is thatmost of them use scripts of Brahmi origin, whichhave highly phonetic characteristics that could beutilized for multilingual NER. For some languages,there are additional issues like word segmentation

    5

  • (e.g. for Thai). Large gazetteers are not avail-able for most of these languages. There is alsothe problem of lack of standardization and spellingvariation. The number of frequently used words(common nouns) which can also be used as names(proper nouns) is very large for, unlike for Euro-pean languages where a larger proportion of the firstnames are not used as common words. For exam-ple, ‘Smith’, ‘John’, ‘Thomas’ and ‘George’ etc. arealmost always used as person names, but ‘Anand’,‘Vijay’, ‘Kiran’ and even ‘Manmohan’ can be (morethan often) used as common nouns. And the fre-quency with which they can be used as commonnouns as against person names is more or less unpre-dictable. The context might help in disambiguating,but this issue does make the problem much harderthan for English.

    Among other problems, one example is that of thevarious ways of representing abbreviations. Becauseof the alpha-syllabic nature of the SSEA scripts, ab-breviation can be expressed through a sequence ofletters or syllables. In the latter case, the syllablesare often combined together to form a pseudo-word,e.g. BAjapA (bhaajapaa) for Bharatiya Janata Partyor BJP.

    But most importantly, there is a serious lack oflabeled data for machine learning. As part of thisworkshop, we have tried to prepare some data but wewill need much more data for really accurate NERsystems.

    Since most of the South and South East Asian lan-guages are scarce in resources as well as tools, it isvery important that good systems for NER be avail-able, because many problems in information extrac-tion and machine translation (among others) are de-pendent on accurate NER.

    The need for a workshop specifically for SSEAlanguages was felt because the South and South EastAsian region has many major and numerous minorlanguages. In terms of the number of speakers thereare at least four in any list of top ten languages of theworld. For practical reasons, we focus only on themajor languages in the workshop (and in this paper).Most of the major languages belong to two families:Indo-European and Dravidian. There are a lot of dif-ferences among these languages, but there are a lotof similarities too, even across families (Emeneau,1956; Emeneau, 1980). For the reasons mentioned

    above, NER is perhaps more difficult for SSEA lan-guages than for European languages. For better orfor worse, there too many languages and too few re-sources. Moreover, these languages are also com-paratively less studied by researchers. However, wecan benefit from the similarities across these lan-guages to build multilingual systems so as to reducethe overall cost and effort required.

    All the issues mentioned above show that wemight need different methods for solving the NERproblem for SSEA languages. However, for com-paring the results of these different methods, we willneed a reasonably good baseline. A mature systemtuned for English but trained on SSEA language datacan become such a baseline. We will describe such abaseline in a later section. This baseline system hasbeen tested on the data provided for the shared task.We present the results for all five languages underthe settings required for the shared task.

    2 Related Work

    Various techniques have been used for solving theNER problem (Mikheev et al., 1999; Borthwick,1999; Cucerzan and Yarowsky, 1999; Chieu and Ng,2003; Klein et al., 2003; Kim and Woodland, 2000)ranging from naively using gazetteers to rules basedtechniques to purely statistical techniques, even hy-brid approaches. Several workshops consisting ofshared tasks (Sang, 2002; Sang and De Meulder,2003) have been held with specific focus on thisproblem. In this section we will mention some oftechniques used previously.

    Most of the approaches can be classified based onthe features they use, whether they are rule based ormachine learning based or hybrid approaches. Someof the commonly used features are:

    • Word form and part of speech (POS) tags

    • Orthographic features like capitalization, deci-mal, digits

    • Word type patterns

    • Conjunction of types like capitalization,quotes, functional words etc.

    • Bag of words

    • Trigger words like New YorkCity6

  • Tag Name DescriptionNEP Person Bob Dylan, Mohandas GandhiNED Designation General Manager, CommissionerNEO Organization Municipal CorporationNEA Abbreviation NLP, B.J.P.NEB Brand Pepsi, Nike (ambiguous)NETP Title-Person Mahatma, Dr., Mr.NETO Title-Object Pride and Prejudice, OthelloNEL Location New Delhi, ParisNETI Time 3rd September, 1991 (ambiguous)NEN Number 3.14, 4,500NEM Measure Rs. 4,500, 5 kgNETE Terms Maximum Entropy, Archeology

    Table 1: The named entity tagset used for the shared task

    • Affixes like Hyderabad, Rampur,Mehdipatnam, Lingampally

    • Gazetteer features: class in the gazetteer

    • Left and right context

    • Token length, e.g. the number of letters in aword

    • Previous history in the document or the corpus

    • Classes of preceding NEs

    The machine learning techniques tried for NERinclude the following:

    • Hidden Markov Models or HMM (Zhou andSu, 2001)

    • Decision Trees (Isozaki, 2001)

    • Maximum Entropy (Borthwick et al., 1998)

    • Support Vector Machines or SVM (Takeuchiand Collier, 2002)

    • Conditional Random Fields or CRF (Settles,2004)

    Different ways of classifying named entities havebeen used, i.e., there are more than one tagsets forNER. For example, the CoNLL 2003 shared task2

    had only four tags: persons, locations, organizations

    2http://www.cnts.ua.ac.be/conll2003/ner/

    and miscellaneous. On the other hand, MUC-63 hasa near ontology for information extraction purposes.In this (MUC-6) tagset, there are three4 main kindsof NEs: ENAMEX (persons, locations and organi-zations), TIMES (time expressions) and NUMEX(number expresssions).

    There has been some previous work on NERfor SSEA languages (McCallum and Li, 2003;Cucerzan and Yarowsky, 1999), but most of the timesuch work was an offshoot of the work done for Eu-ropean languages. Even including the current work-shop, the work on NER for SSEA languages is stillin the initial stages as the results reported by papersin this workshop clearly show.

    3 A New Named Entity Tagset

    The tagset being used for the NERSSEAL-08 sharedtask consists of more tags than the four tags usedfor the CoNLL 2003 shared task. The reason weopted for these tags was that we needed a slightlyfiner tagset for machine translation (MT). The ini-tial aim was to improve the performance of the MTsystem.

    As annotation progressed, we realized that therewere some problems that we had not anticipated.Some classes were hard to distinguish in some con-texts, making the task hard for annotators and bring-ing in inconsistencies. For example, it was not al-ways clear whether something should be marked as

    3http://cs.nyu.edu/cs/faculty/grishman/muc6.html4http://cs.nyu.edu/cs/faculty/grishman/NEtask20.book6.html

    7

  • Number or as Measure. Similarly for Time and Mea-sure. Another difficult class was that of (technical)terms. Is ’agriculture’ a term or not? If no (as mostpeople would say), is ’horticulture’ a term or not? Infact, Term was the most difficult class to mark.

    An option that we explored was to merge theabove mentioned confusable classes and ignore theTerm class. But we already had a relatively largecorpus marked up with these classes. If we mergedsome classes and ignored the Term class (which hada very large coverage and is definitely going to beuseful for MT), we would be throwing away a lotof information. And we also had some corpus an-notated by others which was based on a differenttagset. So some problems were inevitable. Finally,we decided to keep the original tagset, with onemodification. The initial tagset had only eleven tags.The problem was that there was one Title tag butit had two different meanings: ’Mr.’ is a Title, but’The Seven Year Itch’ is also a Title. This tag clearlyneeded to be split into two: Title-Person and Title-Object

    We should mention here that we considered usinganother tagset developed at AUKBC, Chennai. Thiswas based on ENAMEX, TIMEX and NUMEX. Thetotal number of tags in this tagset is more than a hun-dred and it is meant specifically for MT and only forcertain domains (health, tourism). Moreover, this isa tagset for entities in general, not just named enti-ties.

    The twelve tags in our tagset are briefly explainedin Table-1. In the next section we mention the con-straints under which the annotated corpus was cre-ated, using this tagset.

    4 Annotation Constraints

    The annotated corpus was created under severe con-straints. The annotation was to be for five languagesby different teams, sometimes with very little com-munication during the process of annotation. As aresult, there were many logistical problems.

    There were other practical constraints like the factthat this was not a funded project and all the workwas mainly voluntary. Another major constraint forall the languages except Hindi was time. There wasnot enough time for cross validation as the corpuswas required by a deadline. To keep annotation rea-

    sonably consistent, annotation guidelines were cre-ated and a common format was specified.

    5 Annotation Guidelines

    The annotation guidelines were of two kinds. Onewas meant for preparing training data through man-ual annotation. The other one was meant for prepar-ing reference data as well as for automatic annota-tion. The main guidelines for preparing the trainingdata are as follows:

    • Specificity: The most important criterion whiledeciding whether some expression is a namedentity or not is to see whether that expressionspecifies something definite and identifiable asif by a name or not. This decision will have tobe based on the context. For example, ’aanand’(in South Asian languages, where there is nocapitalization) is not a named entity in ’sabaaanand hii aanand hai’ (’There is bliss every-where’). But it is a named entity in ’aanandkaa yaha aakhiri saala hai’ (’Anand is in thelast year (of his studies)’). Number, Measureand Term may be seen as exceptions (see be-low).

    • Maximal Entity: Only the maximal entitieshave to be annotated for training data. Struc-ture of entities will not be annotated by theannotators, even though it has to be learnt bythe NER systems. For example, ’One HundredYears of Solitude’ has to be annotated as oneentity. ’One Hundred’ is not to be marked asa Number here, nor is ’One Hundred Years’ tobe made marked as a Measure in this case. Thepurpose of this guideline is to make the task ofannotation for several languages feasible, giventhe constraints.

    • Ambiguity: In cases where an entity can havetwo valid tags, the more appropriate one is tobe used. The annotator has to make the deci-sion in such cases. It is recommended that theannotation be validated by another person, oreven more preferably, two different annotatorshave to work on the same data independentlyand inconsistencies have to be resolved by anadjudicator. Abbreviation is an exception to theAmbiguity guideline (see below).

    8

  • Some other guidelines for specific tags are listedbelow:

    • Abbreviations: All abbreviations have to bemarked as Abbreviations, Even though everyabbreviation is also some other kind of namedentity. For example, APJ is an Abbreviation,but also a Person. IBM is also an Organiza-tion. Such ambiguity cannot be resolved fromthe context because it is due to the (wrong?)assumption that a named entity can have onlyone tag. Multiple annotations were not al-lowed. This is an exception to the third guide-line above.

    • Designation and Title-Person: An entity is aDesignation if it represents something formaland official status with certain responsibilities.If it is just something honorary, then it is aTitle-Object. For example, ’Event Coordina-tor’ or ’Research Assistant’ is a Designation,but ’Chakravarti’ or ’Mahatma’ are Titles.

    • Organization and Brand: The distinction be-tween these two has to be made based on thecontext. For example, ’Pepsi’ could mean anOrganization, but it is more likely to mean aBrand.

    • Time and Location: Whether something is tobe marked as Time or Location or not is to bedecided based on the Specificity guideline andthe context.

    • Number, Measure and Term: These three maynot be strictly named entities in the way a per-son name is. However, we have included thembecause they are different from other words ofthe language. For problems like machine trans-lation, they can be treated like named entities.For example, a Term is a word which can be di-rectly translated into some language if we havea dictionary of technical terms. Once we knowa word is a Term, there is likely to be less am-biguity about the intended sense of the word,unlike for other normal words.

    The second set of guidelines are different from thefirst set mainly in one respect: the corpus has to beannotated with not just the maximal NEs, but with

    all levels of NEs, i.e., nested NEs also have to bemarked.

    Nested entities were introduced because one ofthe requirements was that the corpus be useful forbuilding systems which can become parts of a ma-chine translation (MT) system. Nested entities canbe useful for MT systems because, quite often, partsof the entities can need to be translated, while theothers can just be transliterated. An example of anested named entity is ‘Mahatma Gandhi Interna-tional Hindi University’. This would be translatedin Hindi as mahaatmaa gaandhii antarraashtriyahindii vishvavidyaalaya. Only ‘International’ and‘University’ are to be translated, while the otherwords are to be transliterated. The nested named en-tities in this case are: ‘Mahatma’ (NETO), ‘Gandhi’(NEP), ‘Mahatma Gandhi’ (NEP), and ‘MahatmaGandhi International Hindi University’ (NEO).

    6 Named Entity Annotated Corpus

    For Hindi, Oriya and Telugu, all the annotation wasperformed at IIIT, Hyderabad. For Bengali, the cor-pus was developed at IIIT, Hyderabad and JadavpurUniversity (Ekbal and Bandyopadhyay, 2008b), Cal-cutta. For Urdu, annotation was performed atCRULP, Lahore (Hussain, 2008) and IIIT, Allahabd.Even though all the annotation was done by nativespeakers of respective languages, named entity an-notation was a new task for everyone involved. Thiswas because of practical constraints as explained inan earlier section.

    The corpus was divided into two parts, one fortraining and one for testing. The testing corpuswas annotated with nested named entities, while thetraining corpus was only annotated with ‘maximal’named entities.

    Since different teams were working on differentlanguages, in some cases even the same language,and also because most of the corpus was created onshort notice, each team made its own decisions re-garding the kind of corpus to be annotated. As a re-sult, the characteristics of the corpus differ widelyamong the five languages. The Hindi and Ben-gali (partly) text that was annotated was from themultilingual comparable corpus known as the CIIL(Central Institute of Indian Languages) corpus. TheOriya corpus was part of the Gyan Nidhi corpus.

    9

  • NE Hindi Bengali Oriya Telugu UrduTrn Tst Trn Tst Trn Tst Trn Tst Trn Tst

    NEP 4025 199 1299 728 2079 698 1757 330 365 145NED 935 61 185 11 67 216 87 77 98 41NEO 1225 44 264 20 87 200 86 12 155 40NEA 345 7 111 9 8 20 97 112 39 3NEB 5 0 22 0 11 1 1 6 9 18NETP 1 5 68 57 54 201 103 2 36 15NETO 964 88 204 46 37 28 276 118 4 147NEL 4089 211 634 202 525 564 258 751 1118 468NETI 1760 50 285 46 102 122 244 982 279 59NEN 6116 497 407 144 124 232 1444 391 310 47NEM 1287 17 352 146 280 139 315 53 140 40NETE 5658 843 1165 314 5 0 3498 138 30 4NEs 26432 2022 5000 1723 3381 2421 8178 3153 2584 1027Words 503179 32796 112845 38708 93173 27007 64026 8006 35447 12805Sentences 19998 2242 6030 1835 1801 452 5285 337 1508 498

    Trn: Training Data,Tst: Testing Data

    Table 2: Statistics about the corpus: counts of various named entity classes and the size of the corpus as thenumber of words and the number of sentences. Note that the values for the testing part are of nested NEs.Also, the number of sentences, especially in the case or Oriya is not accurate because the sentences were notcorrectly segmented as there was no automatic sentence splitter available for these languages and manualsplitting would have been too costly: without much benefit for the NER task.

    Both of these (CIIL and Gyan Nidhi) corpora con-sist of text from educational books written on vari-ous topics for common readers. The Urdu text waspartly news corpus. The same was the case with Tel-ugu, but the text for both these languages includedtext from other domains too.

    Admittedly, the texts selected for annotation werenot the ideal ones. For example, many documentshad very few named entities. Also, the distributionof domains as well as the classes of NEs was notrepresentative. The size of the annotated corporafor different languages is also widely varying, withHindi having the largest corpus and Urdu the small-est. However, this corpus is hopefully just a startingpoint for much more work in the near future.

    Some statistics about the annotated corpus aregiven in Table-2.

    7 Shared Task

    In the shared task, the contestants having their ownNER systems were given some annotated test data.The contestants had the freedom to use any tech-nique for NER, e.g. a purely rule based techniqueor a purely statistical technique.

    The contestants could build NER systems targetedfor a specific language, but they were required to re-

    port results for their systems on all the languagesfor which training data had been provided. Thiscondition was meant to provide a somewhat fairground for comparison of systems, since the amountof training data is different for different languages.

    The data released for the shared task has beenmade accessible to all for non-profit research word,not just for the shared task participants, with thehope others will contribute in improving this dataand adding to it.

    The task in this contest was different in one im-portant way. The NER systems also had to identifynested named entities. For example, in the sentenceThe Lal Bahadur Shastri National Academy of Ad-ministration is located in Mussoorie, ’Lal BahadurShastri’ is a Person, but ’Lal Bahadur Shastri Na-tional Academy of Administration’ is an Organiza-tion. In this case, the NER systems had to identifyboth ’Person’ and ’Organization’ in the given sen-tence.

    An evaluation script was also provided to evaluatethe performance of different systems in a uniformway.

    10

  • 8 Evaluation Measures

    As part of the evaluation process for the shared task,precision, recall and F-measure had to be calcu-lated for three cases: maximal named entities, nestednamed entities and lexical matches. Thus, therewere nine measures of performance:

    • Maximal Precision:Pm = cmrm

    • Maximal Recall:Rm = cmtm

    • Maximal F-Measure:Fm = 2∗Pm∗RmPm+Rm

    • Nested Precision:Pn = cnrn

    • Nested Recall:Rn = cntn

    • Nested F-Measure:Fn = 2PnRnPn+Rn

    • Lexical Precision:Pl =clrl

    • Lexical Recall:Rl =cltl

    • Lexical F-Measure:Fl =2PlRlPl+Rl

    wherec is the number of correctly retrieved (iden-tified) named entities,r is the total number of namedentities retrieved by the system being evaluated (cor-rect plus incorrect) andt is the total number ofnamed entities in the reference data.

    The participants were encouraged to report resultsfor specific classes of NEs. Evaluation was auto-matic and was against the manually prepared refer-ence data given to the participants. An evaluationscript for this purpose was also provided. This scriptassumes that there are single test and reference fileand the number and order of sentences is the same inboth. The format accepted by the evaluation script(which was also the format used for annotated data)was explained in an online tutorial5.

    9 Experiments on a Baseline

    For our baseline experiments, we used an opensource implementation of maximum entropy basedNatural Languages Processing tools which are partof the OpenNLP6 package. This package includes aname finder tool.

    5http://ltrc.iiit.ac.in/ner-ssea-08/NER-SAL-TUT.pdf6http://opennlp.sourceforge.net/

    This name finder was trained for all the twelveclasses of NEs and for all the five languages. Thetest data, which was the same as that given to theshared task participants, was run through this namefinder. Note that this NER tool is tuned for En-glish in terms of the features used, even though itwas trained on different SSEA languages in our case.Since the goal of the shared task was to encourageinvestigation of techniques (especially features) spe-cific to the SSEA languages, this fairly mature NERsystem (for English) could be used as a baselineagainst which to evaluate systems tuned (or speciallydesigned) for the five South Asian languages.

    The overall results of the baseline experiments areshown in Table-3. The performance on specific NEclasses is given in Table-4. It can be seen from thetables that the results are drastically low in compar-ison to the state of the art results reported for En-glish. These results clearly show that even a ma-chine learning based system cannot be directly usedfor SSEA languages even when it has been trainedwith annotated data for these languages.

    In the next section we present a brief overview ofthe papers selected for the workshop including theshared task papers.

    10 An Overview of the Papers

    In all, twelve papers were selected for the workshop,out of which four were in the shared task track. Sahaet al., who were able to achieve the best results inthe shared task, describe a hybrid system that ap-plies maximum entropy models, language specificrules, and gazetteers. For Hindi, the features theyutilized include orthographic features, informationabout suffixes and prefixes, morphological features,part of speech information, and information aboutthe surrounding words. They used rules for num-bers, measures and time classes. For designation,title-person and some terms (NETE), they built listsor gazetteers. They also used gazetteers for personand location. They did not use rules or gazetteers forOriya, Urdu and Telugu. To identify some kinds ofnested entities, they applied a set of rules.

    Gali et al. also combined machine learning withlanguage specific heuristics. In a separate section,they discussed at some length the issues relevant toNER for SSEA languages. Some of these have al-

    11

  • Measure → Precision Recall F-MeasureLanguage ↓ Pm Pn Pl Rm Rn Rl Fm Fn FlBengali 50.00 44.90 52.20 07.14 06.90 06.97 12.50 11.97 12.30Hindi 75.05 73.61 73.99 18.16 17.66 15.53 29.24 28.48 25.68Oriya 29.63 27.46 48.25 09.11 07.60 12.18 13.94 11.91 19.44Telugu 00.89 02.83 22.85 00.20 00.67 5.41 00.32 01.08 08.75Urdu 47.14 43.50 51.72 18.35 16.94 18.94 26.41 24.39 27.73

    m: Maximal, n: Nested, l: Lexical

    Table 3: Results for the experiments on a baseline for the fiveSouth Asian languages

    Bengali Hindi Oriya Telugu UrduNEP 06.62 26.23 28.48 00.00 04.39NED 00.00 12.20 00.00 00.00 00.00NEO 00.00 15.50 03.30 00.00 11.98NEA 00.00 00.00 00.00 00.00 00.00NEB NP NP 00.00 00.00 00.00NETP 00.00 NP 11.62 00.00 00.00NETO 00.00 05.92 04.08 00.00 00.00NEL 03.03 44.79 25.49 00.00 40.21NETI 34.00 47.41 22.38 01.51 38.38NEN 62.63 62.22 10.65 03.51 09.52NEM 13.61 24.39 08.03 00.71 07.15NETE 00.00 00.18 00.00 00.00 00.00

    NP: Not present in the reference data

    Table 4: Baseline results for specific named entity classes (F-Measures for nested lexical match)

    ready been mentioned, but two others are the ag-glutinative property of these (especially Dravidian)languages and the low accuracy of available part ofspeech taggers, particularly for nouns. They useda Conditional Random Fields (CRF) based methodfor machine learning and applied heuristics to takecare of the language specific issues. They also pointout that a very high percentage of NEs in the Hindicorpus were marked as NETE and machine learningfailed to take care of this class of NEs. This has beenvalidated by our results on the baseline too (Table-4) and is understandable because terms are hard toidentify even for humans.

    Ekbal et al. also used an approach based on CRFs.They also used some language specific features forHindi and Bengali. Srikanth and Murthy describethe results of their experiments on NER using CRFsfor Telugu. They concentrated only on person, placeand organization names and used newspaper text

    as the corpus. In this focused setting, they wereable to achieve overall F-measures between 80% and97% in various experiments. Chaudhuri and Bhat-tacharya also experimented on a news corpus forBengali using a three stage NER system. The threestages were based on an NE dictionary, rules andcontextual co-occurrence statistics. They only triedto identify the NEs, not classify them. For this task,they were able to achieve an overall F-measure of89.51%.

    Praveen and Ravi Kumar present the results ofexperiments (as part of the shared task) using twoapproaches: Hidden Markov Models (HMM) andCRF. Surprisingly, they obtained better results withHMM for all the five languages. Goyal described ex-periments using a CRF based model. He also usedpart of speech information. He experimented onlyon Hindi and was able to achieve results above 60%.One notable fact about this paper is that it also de-

    12

  • Language ↓ BL IK IH1 IH2 JUBengali 12.30 65.96 40.63 39.77 59.39Hindi 25.68 65.13 50.06 46.84 33.12Oriya 19.44 44.65 39.04 45.84 28.71Telugu 08.75 18.74 40.94 46.58 04.75Urdu 27.73 35.47 43.46 44.73 35.52

    Average 18.78 45.99 42.83 44.75 32.30

    BL: Baseline,IK: IIT KharagpurJU: Jadavpur University, Calcutta

    IH1: Karthik et al., IIIT HyderabadIH2: Praveen and Ravi Kiran, IIIT Hyderabad

    Table 5: Comparison of NER systems which participated in theNERSSEAL-08 shared task against a base-line that uses maximum entropy based name finder tuned for English but trained on data from five SouthAsian languages

    scribes experiments on the CoNLL 2003 shared taskdata for English, which shows that the significantlyhigher results for English are mainly due to the factthat the CoNLL 2003 data is already POS tagged andchunked with high accuracy. Goyal was also able toshow that capitalization is a major clue for English,either directly or indirectly (e.g., for accurate POStagging and chunking). He also indicated that thecharacteristics of the Hindi annotated corpus werepartly responsible for the low results on Hindi.

    Nayan et al. mainly describe how an NER systemcan benefit from approximate string matching basedon phonetic edit distance, both for a single language(to account for spelling variations) and for cross-lingual NER. Shishtla et al. (‘Experiments in Tel-ugu NER’) experimented only on Telugu and usedthe CoNLL shared task tagset. Using a CRF basedapproach, they were able to achieve an F-measureof 44.91%. Ekbal and Bandyopadhyay describe amethod based on Support Vector Machines (SVMs)for Bengali NER. On a news corpus and with sixteenNE classes, they were able to achieve an F-measureof 91.8%. Vijayakrishna and Sobha describe a CRFbased system for Tamil using 106 NE classes. Theirsystem is a multi-level system which gave an over-all F-measure of 80.44%. They also mention thattheir system achieved this level of performance on adomain focused corpus. Shishtla et al. (‘Charactern-gram Based Approach’) used a charactern-grambased method to identify NEs. They experimentedon Hindi as well as English and achieved F-measure

    values up to 45.48% for Hindi and 68.46% for En-glish.

    Apart from the paper presentations, the workshopwill also have two invited talks. The first one is titled“Named Entity Recognition: Different Approaches”by Sobha L. and the second one is “MultilingualNamed Entity Recognition” by Sivaji Bandyopad-hyay.

    11 Shared Task Results

    Five teams participated in the shared task. However,only four submitted papers for the workshop. All theteams tried to combine machine learning with somelanguage specific heuristics, at least for one of thelanguages. The results obtained by the four teamsare summarized in Table-5, which shows only the F-measure for lexical match. It can be seen from thetable that all the teams were able to get significantlybetter results than the baseline. Overall, the perfor-mance of the IIT Kharagpur team was the best, fol-lowed by the two teams from IIIT Hyderabad.

    Even though all the teams obtained results muchbetter than the baseline, it is still quite evident thatthe state of the art for NER for SSEA languagesleaves much to be desired. At around 46% max-imum F-measure on lexical matching, the resultsmean that the NER systems built so far for SSEAlanguages are not quite practically useful. But, afterthis workshop, we at least know where we stand andhow far we still have to go.

    However, it may be noted that the conditions for13

  • the shared task were very stringent compared to theprevious shared tasks on NER, e.g. neither the cor-pus was tagged with parts of speech or chunks, norwere good POS taggers or chunkers available forthe languages involved. This indicates that withprogress in building better resources and basic toolsfor these languages, the accuracy of NER systemsshould also increase. Already, some very high accu-racies are being reported under less stringent condi-tions, e.g. for domain focused NER.

    12 Conclusions

    We started by discussing the problem of NER forSouth and South East Asian languages and the moti-vations for organizing a workshop on this topic. Wealso described a named entity annotated corpus forfive South Asian languages used for this workshop.We presented some statistics about the corpus andalso the problems we encountered in getting the cor-pus annotated by teams located in distant places. Wealso presented a new named entity tagset that wasdeveloped for annotation of this corpus. Then wepresented the results for our experiments on a rea-sonable baseline. Finally we gave an overview ofthe papers selected for the NERSSEAL-08 work-shop and discussed the systems described in thesepapers and the results obtained, including those forthe shared task which was one of the two tracks inthe workshop.

    References

    Sivaji Bandyopadhyay. 2008. Invited talk: Multilin-gual named entity recognition. InProceedings of theIJCNLP-08 Workshop on NER for South and SouthEast Asian Languages, pages 15–17, Hyderabad, In-dia, January.

    A. Borthwick, J. Sterling, E. Agichtein, and R. Grishman.1998. Exploiting diverse knowledge sources via max-imum entropy in named entity recognition.Proceed-ings of the Sixth Workshop on Very Large Corpora,pages 152–160.

    A. Borthwick. 1999. A Maximum Entropy Approach toNamed Entity Recognition. Ph.D. thesis, New YorkUniversity.

    Bidyut Baran Chaudhuri and Suvankar Bhattacharya.2008. An experiment on automatic detection of namedentities in bangla. InProceedings of the IJCNLP-08

    Workshop on NER for South and South East AsianLanguages, pages 51–58, Hyderabad, India, January.

    H.L. Chieu and H.T. Ng. 2003. Named entity recognitionwith a maximum entropy approach.Proceedings ofthe seventh conference on Natural language learningat HLT-NAACL 2003-Volume 4, pages 160–163.

    S. Cucerzan and D. Yarowsky. 1999. Language indepen-dent named entity recognition combining morpholog-ical and contextual evidence.Proceedings of the JointSIGDAT Conference on EMNLP and VLC 1999, pages90–99.

    Asif Ekbal and Sivaji Bandyopadhyay. 2008a. Ben-gali named entity recognition using support vector ma-chine. In Proceedings of the IJCNLP-08 Workshopon NER for South and South East Asian Languages,pages 85–92, Hyderabad, India, January. Associationfor Computational Linguistics.

    Asif Ekbal and Sivaji Bandyopadhyay. 2008b. Devel-opment of bengali named entity tagged corpus and itsuse in ner systems. InProceedings of the Sixth Work-shop on Asian Language Resources, Hyderabad, India,January.

    Asif Ekbal, Rejwanul Haque, Amitava Das,Venkateswarlu Poka, and Sivaji Bandyopadhyay.2008. Language independent named entity recog-nition in indian languages. InProceedings of theIJCNLP-08 Workshop on NER for South and SouthEast Asian Languages, pages 33–40, Hyderabad,India, January.

    M. B. Emeneau. 1956. India as a linguistic area.Lin-guistics, 32:3-16.

    M. B. Emeneau. 1980.Language and linguistic area. Es-says by Murray B. Emeneau. Selected and introducedby Anwar S. Dil. Stanford University Press.

    Karthik Gali, Harshit Surana, Ashwini Vaidya, PraneethShishtla, and Dipti Misra Sharma. 2008. Aggregatingmachine learning and rule based heuristics for namedentity recognition. InProceedings of the IJCNLP-08Workshop on NER for South and South East AsianLanguages, pages 25–32, Hyderabad, India, January.

    Amit Goyal. 2008. Named entity recognition for southasian languages. InProceedings of the IJCNLP-08Workshop on NER for South and South East AsianLanguages, pages 63–70, Hyderabad, India, January.

    Sarmad Hussain. 2008. Resources for urdu languageprocessing. InProceedings of the Sixth Workshopon Asian Language Resources, Hyderabad, India, Jan-uary.

    Hideki Isozaki. 2001. Japanese named entity recogni-tion based on a simple rule generator and decision tree

    14

  • learning. InMeeting of the Association for Computa-tional Linguistics, pages 306–313.

    J.H. Kim and PC Woodland. 2000. A rule-based namedentity recognition system for speech input.Proc. ofICSLP, pages 521–524.

    D. Klein, J. Smarr, H. Nguyen, and C.D. Manning. 2003.Named entity recognition with character-level models.Proceedings of CoNLL, 3.

    Sobha L. 2008. Invited talk: Named entity recognition:Different approaches. InProceedings of the IJCNLP-08 Workshop on NER for South and South East AsianLanguages, pages 13–14, Hyderabad, India, January.

    A. McCallum and W. Li. 2003. Early results for namedentity recognition with conditional random fields, fea-ture induction and web-enhanced lexicons.SeventhConference on Natural Language Learning (CoNLL).

    A. Mikheev, M. Moens, and C. Grover. 1999. NamedEntity recognition without gazetteers.Proceedings ofthe ninth conference on European chapter of the Asso-ciation for Computational Linguistics, pages 1–8.

    Animesh Nayan, B. Ravi Kiran Rao, Pawandeep Singh,Sudip Sanyal, and Ratna Sanyal. 2008. Named entityrecognition for indian languages. InProceedings ofthe IJCNLP-08 Workshop on NER for South and SouthEast Asian Languages, pages 71–78, Hyderabad, In-dia, January. Association for Computational Linguis-tics.

    Praveen P and Ravi Kiran V. 2008. Hybrid named entityrecognition system for south and south east asian lan-guages. InProceedings of the IJCNLP-08 Workshopon NER for South and South East Asian Languages,pages 59–62, Hyderabad, India, January.

    Vijayakrishna R and Sobha L. 2008. Domain focusednamed entity recognizer for tamil using conditionalrandom fields. InProceedings of the IJCNLP-08Workshop on NER for South and South East AsianLanguages, pages 93–100, Hyderabad, India, January.Association for Computational Linguistics.

    Sujan Kumar Saha, Sanjay Chatterji, Sandipan Dandapat,Sudeshna Sarkar, and Pabitra Mitra. 2008. A hybridnamed entity recognition system for south and southeast asian languages. InProceedings of the IJCNLP-08 Workshop on NER for South and South East AsianLanguages, pages 17–24, Hyderabad, India, January.

    E.F.T.K. Sang and F. De Meulder. 2003. Introduction tothe CoNLL-2003 Shared Task: Language-IndependentNamed Entity Recognition.Development, 922:1341.

    Erik F. Tjong Kim Sang. 2002. Introduction to the conll-2002 shared task: Language-independentnamed entityrecognition. InProceedings of CoNLL-2002, pages155–158. Taipei, Taiwan.

    Manabu Sassano and Takehito Utsuro. 2000. Namedentity chunking techniques in supervised learning forjapanese named entity recognition. InProceedingsof the 18th conference on Computational linguistics,pages 705–711, Morristown, NJ, USA. Association forComputational Linguistics.

    B. Settles. 2004. Biomedical Named Entity RecognitionUsing Conditional Random Fields and Rich FeatureSets.log, 1:1.

    Praneeth M Shishtla, Karthik Gali, Prasad Pingali, andVasudeva Varma. 2008a. Experiments in telugu ner:A conditional random field approach. InProceed-ings of the IJCNLP-08 Workshop on NER for Southand South East Asian Languages, pages 79–84, Hyder-abad, India, January. Association for ComputationalLinguistics.

    Praneeth M Shishtla, Prasad Pingali, and VasudevaVarma. 2008b. A character n-gram based approachfor improved recall in indian language ner. InPro-ceedings of the IJCNLP-08 Workshop on NER forSouth and South East Asian Languages, pages 101–108, Hyderabad, India, January. Association for Com-putational Linguistics.

    P Srikanth and Kavi Narayana Murthy. 2008. Namedentity recognition for telugu. InProceedings of theIJCNLP-08 Workshop on NER for South and SouthEast Asian Languages, pages 41–50, Hyderabad, In-dia, January.

    K. Takeuchi and N. Collier. 2002. Use of support vec-tor machines in extended named entity recognition. InProceedings of the sixth Conference on Natural Lan-guage Learning (CoNLL-2002).

    G.D. Zhou and J. Su. 2001. Named entity recognitionusing an HMM-based chunk tagger.Proceedings ofthe 40th Annual Meeting on Association for Computa-tional Linguistics, pages 473–480.

    15

  • 16

  • Proceedings of the IJCNLP-08 Workshop on NER for South and South East Asian Languages, pages 17–24,Hyderabad, India, January 2008. c©2008 Asian Federation of Natural Language Processing

    A Hybrid Approach for Named Entity Recognition in Indian Lan guages

    Sujan Kumar Saha Sanjay Chatterji Sandipan DandapatIndian Institute of Technology Indian Institute of Technology Indian Institute of TechnologyKharagpur, West Bengal Kharagpur, West Bengal Kharagpur, West Bengal

    India - 721302 India - 721302 India - [email protected] [email protected] [email protected]

    Sudeshna Sarkar Pabitra MitraIndian Institute of Technology Indian Institute of TechnologyKharagpur, West Bengal Kharagpur, West Bengal

    India - 721302 India - [email protected] [email protected]

    Abstract

    In this paper we describe a hybrid systemthat applies Maximum Entropy model (Max-Ent), language specific rules and gazetteersto the task of Named Entity Recognition(NER) in Indian languages designed for theIJCNLP NERSSEAL shared task. Startingwith Named Entity (NE) annotated corporaand a set of features we first build a base-line NER system. Then some language spe-cific rules are added to the system to recog-nize some specific NE classes. Also we haveadded some gazetteers and context patternsto the system to increase the performance.As identification of rules and context pat-terns requires language knowledge, we wereable to prepare rules and identify contextpatterns for Hindi and Bengali only. For theother languages the system uses the MaxEntmodel only. After preparing the one-levelNER system, we have applied a set of rulesto identify the nested entities. The systemis able to recognize 12 classes of NEs with65.13% f-value in Hindi, 65.96% f-value inBengali and 44.65%, 18.74%, and 35.47%f-value in Oriya, Telugu and Urdu respec-tively.

    1 Introduction

    Named entity recognition involves locating and clas-sifying the names in text. NER is an important

    task, having applications in Information Extraction(IE), Question Answering (QA), Machine Transla-tion (MT) and in most other NLP applications.

    This paper presents a Hybrid NER system for In-dian languages which is designed for the IJCNLPNERSSEAL shared task competition, the goal ofwhich is to perform NE recognition on 12 typesof NEs - person, designation, title-person, organiza-tion, abbreviation, brand, title-object, location, time,number, measure and term.

    In this work we have identified suitable featuresfor the Hindi NER task. Orthography features, suf-fix and prefix information, morphology informa-tion, part-of-speech information as well as informa-tion about the surrounding words and their tags areused to develop a MaxEnt based Hindi NER sys-tem. Then we realized that the recognition of someclasses will be better if we apply class specific lan-guage rules in addition to the MaxEnt model. Wehave defined rules for time, measure and numberclasses. We made gazetteers based identification fordesignation, title-person and some terms. Also wehave used person and location gazetteers as featuresof MaxEnt for better identification of these classes.Finally we have built a module for semi-automaticextraction of context patterns and extracted contextpatterns for person, location, organization and title-object classes and these are added to the baselineNER system.

    The shared task was defined to build the NER sys-tems for 5 Indian languages - Hindi, Bengali, Oriya,Telugu and Urdu for which training data was pro-

    17

  • vided. Among these 5 languages only Bengali andHindi are known to us but we have no knowledge forother 3 languages. So we are unable to build rulesand extract context patterns for these languages. TheNER systems for these 3 languages contain onlythe baseline system i.e. the MaxEnt system. Alsoour baseline MaxEnt NER system uses morphologi-cal and parts-of-speech (POS) information as a fea-ture. Due to unavailability of morphological ana-lyzer and POS tagger for these 3 languages, these in-formation are not added to the systems. Among the3 languages, only for Oriya NER system we haveused small gazetteers for person, location and des-ignation extracted from the training data. For Ben-gali and Hindi the developed systems are completehybrid systems containing rules, gazetteers, contextpatterns and the MaxEnt model.

    The paper is organized as follows. A brief sur-vey of different techniques used for the NER taskin different languages and domains are presented inSection 2. Also a brief survey on nested NE recog-nition system is presented here. A discussion onthe training data is given in Section 3. The MaxEntbased NER system is described in Section 4. Vari-ous features used in NER are then discussed. Nextwe present the experimental results and related dis-cussions in Section 8. Finally Section 9 concludesthe paper.

    2 Previous Work

    A variety of techniques has been used for NER. Thetwo major approaches to NER are:

    1. Linguistic approaches.

    2. Machine Learning (ML) based approaches.

    The linguistic approaches typically use rules man-ually written by linguists. There are several rule-based NER systems, containing mainly lexicalizedgrammar, gazetteer lists, and list of trigger words,which are capable of providing 88%-92% f-measureaccuracy for English (Grishman, 1995; McDonald,1996; Wakao et al., 1996).

    The main disadvantages of these rule-based tech-niques are that these require huge experience andgrammatical knowledge of the particular languageor domain and these systems are not transferable toother languages or domains.

    ML based techniques for NER make use of alarge amount of NE annotated training data to ac-quire high level language knowledge. Several MLtechniques have been successfully used for the NERtask of which Hidden Markov Model (HMM) (Bikelet al., 1997), Maximum Entropy (MaxEnt) (Borth-wick, 1999), Conditional Random Field (CRF) (Liand Mccallum, 2004) are most common. Combina-tions of different ML approaches are also used. Sri-hari et al. (2000) combines MaxEnt, Hidden MarkovModel (HMM) and handcrafted rules to build anNER system.

    NER systems use gazetteer lists for identifyingnames. Both the linguistic approach (Grishman,1995; Wakao et al., 1996) and the ML based ap-proach (Borthwick, 1999; Srihari et al., 2000) usegazetteer lists.

    Linguistic approach uses handcrafted rules whichneeds skilled linguistics. Some recent approachestry to learn context patterns through ML which re-duce amount of manual labour. Talukder et al.(2006)combined grammatical and statistical techniques tocreate high precision patterns specific for NE extrac-tion. An approach to lexical pattern learning for In-dian languages is described by Ekbal and Bandopad-hyay (2007). They used seed data and annotated cor-pus to find the patterns for NER.

    The NER task for Hindi has been explored byCucerzan and Yarowsky in their language indepen-dent NER work which used morphological and con-textual evidences (Cucerzan and Yarowsky, 1999).They ran their experiment with 5 languages - Roma-nian, English, Greek, Turkish and Hindi. Amongthese the accuracy for Hindi was the worst. ForHindi the system achieved 41.70% f-value with avery low recall of 27.84% and about 85% precision.A more successful Hindi NER system was devel-oped by Wei Li and Andrew Mccallum (2004) usingConditional Random Fields (CRFs) with feature in-duction. They were able to achieve 71.50% f-valueusing a training set of size 340k words. In Hindithe maximum accuracy is achieved by Kumar andBhattacharyya, (2006). Their Maximum EntropyMarkov Model (MEMM) based model gives 79.7%f-value.

    All the NER systems described above are ableto detect one-level NEs. In recent years, the inter-est in detection of nested NEs has increased. Here

    18

  • we mention few attempts for nested NE detection.Zhou et al. (2004) described an approach to iden-tify cascaded NEs from biomedical texts. They de-tected the innermost NEs first and then they derivedrules to find the other NEs containing these as sub-strings. Another approach, described by McDonaldet al. (2005), uses structural multilevel classifica-tion to deal with overlapping and discontinuous enti-ties. B. Gu (2006) has treated the task of identifyingthe nested NEs a binary classification problem andsolved it using support vector machines. For eachtoken in nested NEs, they used two schemes to setits class label: labeling as the outermost entity or theinner entities.

    3 Training Data

    The data used for the training of the systems wasprovided. The annotated data uses Shakti StandardFormat (SSF). For our development we have con-verted the SSF format data into theIOB formattedtext in which aB − XXX tag indicates the firstword of an entity typeXXX andI−XXX is usedfor subsequent words of an entity. The tagO indi-cates the word is outside of a NE. The training datafor Hindi contains more than 5 lakh words, for Ben-gali about 160K words and about 93K, 64K and 36Kwords for Oriya, Telugu and Urdu respectively.

    In time of development we have observed thatthe training data, provided by the organizers of theshared task, contains several types of errors in NEtagging. These errors in the training corpora affectsbadly to the machine learning (ML) based models.But we have not made corrections of the errors inthe training corpora in time of our development. Allthe results shown in the paper are obtained using theprovided corpora without any modification in NEannotation.

    4 Maximum Entropy Based Model

    We have used MaxEnt model to build the baselineNER system. MaxEnt is a flexible statistical modelwhich assigns an outcome for each token based onits history and features. Given a set of features and atraining corpus, the MaxEnt estimation process pro-duces a model. For our development we have useda Java based open-nlp MaxEnt toolkit1 to get the

    1www.maxent.sourceforge.net

    probability values of a word belonging to each class.That is, given a sequence of words, the probabilityof each class is obtained for each word. To find themost probable tag corresponding to each word of asequence, we can choose the tag having the highestclass conditional probability value. But this methodis not good as it might result in an inadmissible as-signment.

    Some tag sequences should never happen. Toeliminate these inadmissible sequences we havemade some restrictions. Then we used a beamsearch algorithm with a beam of length 3 with theserestrictions.

    4.1 Features

    MaxEnt makes use of different features for identify-ing the NEs. Orthographic features (like capitaliza-tion, decimal, digits), affixes, left and right context(like previous and next words), NE specific triggerwords, gazetteer features, POS and morphologicalfeatures etc. are generally used for NER. In En-glish and some other languages, capitalization fea-tures play an important role as NEs are generallycapitalized for these languages. Unfortunately thisfeature is not applicable for the Indian languages.Also Indian person names are more diverse, lots ofcommon words having other meanings are also usedas person names. Li and Mccallum (2004) used theentire word text, character n-grams (n = 2, 3, 4),word prefix and suffix of lengths 2, 3 and 4, and 24Hindi gazetteer lists as atomic features in their HindiNER. Kumar and Bhattacharyya (2006) used wordfeatures (suffixes, digits, special characters), contextfeatures, dictionary features, NE list features etc. intheir MEMM based Hindi NER system. In the fol-lowing we have discussed about the features we haveidentified and used to develop the Indian languageNER systems.

    Static Word Feature: The previous and nextwords of a particular word are used as features. Thepreviousm words (wi−m...wi−1) to nextn words(wi+1...wi+n) can be considered. During our exper-iment different combinations of previous 4 to next 4words are used.

    Context Lists: Context words are defined as thefrequent words present in a word window for a par-ticular class. We compiled a list of the most frequentwords that occur within a window ofwi−3...wi+3

    19

  • of every NE class. For example, location con-text list contains the words like ‘jAkara2’ (go-ing to), ‘desha’ (country), ‘rAjadhAnI ’ (capital)etc. and person context list contains ‘kahA’ (say),‘pradhAnama.ntrI ’ (prime minister) etc. For agiven word, the value of this feature correspond-ing to a given NE type is set to 1 if the windowwi−3...wi+3 around thewi contains at last one wordfrom this list.

    Dynamic NE tag: Named Entity tags of the pre-vious words(ti−m...ti−1) are used as features.

    First Word: If the token is the first word of asentence, then this feature is set to1. Otherwise, itis set to0.

    Contains Digit: If a token ‘w’ contains digit(s)then the featureContainsDigit is set to 1.

    Numerical Word: For a token ‘w’ if the wordis a numerical word i.e. a word denoting a number(e.g.eka (one),do (two), tina (three) etc.) then thefeatureNumWord is set to 1.

    Word Suffix: Word suffix information is helpfulto identify the NEs. Two types of suffix featureshave been used. Firstly a fixed length word suffix ofthe current and surrounding words are used as fea-tures. Secondly we compiled lists of common suf-fixes of person and place names in Hindi. For ex-ample, ‘pura’, ‘ bAda’, ‘ nagara’ etc. are locationsuffixes. We used binary features corresponding tothe lists - whether a given word has a suffix from aparticular list.

    Word Prefix: Prefix information of a word mayalso be helpful in identifying whether it is a NE. Afixed length word prefix of current and surroundingwords are treated as features.

    Root Information of Word: Indian languagesare morphologically rich. Words are inflected in var-ious forms depending on its number, tense, person,case etc. Identification of NEs becomes difficult forthese inflections. The task becomes easier if insteadof the inflected words, corresponding root words arechecked whether these are NE or not. For that taskwe have used morphological analyzers for Hindi andBengali which are developed at IIT kharagpur.

    Parts-of-Speech (POS) Information: The POSof the current word and the surrounding words may

    2All Hindi words are written in italics using the ‘Itrans’transliteration

    be useful feature for NER. We have accessed toHindi and Bengali POS taggers developed at IITKharagpur which has accuracy about 90%. Thetagset of the tagger contains 28 tags. We have usedthe POS values of the current and surrounding to-kens as features.

    We realized that the detailed POS tagging is notvery relevant. Since NEs are noun phrases, the nountag is very relevant. Further the postposition follow-ing a name may give a clue to the NE type for Hindi.So we decided to use a coarse-grained tagset withonly three tags - nominal (Nom), postposition (PSP)and other (O).

    The POS information is also used by defining sev-eral binary features. An example is theNomPSPbinary feature. The value of this feature is definedto be 1 if the current token is nominal and the nexttoken is a PSP.

    5 Language Specific Rules

    After building of the MaxEnt model we have ob-served that only a small set of rules are able to iden-tify the classes like number, measure, time, more ef-ficiently than the MaxEnt based model. Then wehave tried to define the rules for these classes. Therule identification is done manually and requires lan-guage knowledge. We have defined the requiredrules for Bengali and Hindi but we are unable to dothe same for other 3 languages as the languages areunknown to us. In the following we have mentionedsome example rules which are defined and used inour system.

    • IF ((Wi is a number or numeric word) AND(Wi+1 is an unit))THEN (Wi Wi+1) bigram is ameasure NE.

    • IF ((Wi is a number or numeric word) AND(Wi+1 is a month-name) AND (Wi+2 is a 4digit number))THEN (Wi Wi+1 Wi+2) trigram is atime NE.

    • IF ((Wi denotes a day of a week) AND (Wi+1is a number or numeric word) AND (Wi+2 is amonth name))THEN (Wi Wi+1 Wi+2) trigram is atime NE.

    We have defined 36 rules in total for time, mea-sure and number classes. These rules use some lists

    20

  • which are built. These lists contain correspond-ing entries both in the target language and in En-glish. For example the months names list containsthe names according to the English calender and thenames according to the Indian calender. In the fol-lowing we have mentioned the lists we have pre-pared for the rule-based module.

    • Names of months.

    • Names of seasons.

    • Days of a week.

    • Names of units.

    • Numerical words.

    5.1 Semi-automatic Extraction of ContextPatterns

    Similar to the rules defined for time, measure anddate classes, if efficient context patterns (CP) canbe extracted for a particular class, these can helpin identification of NEs of the corresponding class.But extraction of CP requires huge labour if donemanually. We have developed a module for semi-automatically extraction of context patterns. Thismodule makes use of the most frequent entities ofa particular class asseed for that class and finds thesurrounding tokens of theseed to extract effectivepatterns. We mark a pattern as ‘effective’ if the pre-cision of the pattern is very high. Precision of a pat-tern is defined as the ratio of correct identificationand the total identification when the pattern is usedto identify NEs of a particular type from a text.

    For our task we have extracted patterns for per-son, location, organization and title-object classes.These patterns are able to identify the NEs of a spe-cific classes but detection of NE boundary is notdone properly by the patterns. For boundary detec-tion we have added some heuristics and used POSinformation of the surrounding words. The patternsfor a particular class may identify the NEs of otherclasses also. For example the patterns for identify-ing person names may also identify the designationor title-persons. These need to be handled carefullyat the time of using patterns. In the following someexample patterns are listed which are able to identifyperson names for Hindi.

    • ne kahA ki

    • kA kathana he.n

    • mukhyama.ntrI Aja

    • ne apane gra.ntha

    • ke putra

    6 Use of Gazetteer Lists

    Lists of names of various types are helpful in nameidentification. Firstly we have prepared the lists us-ing the training corpus. But these are not sufficient.Then we have compiled some specialized name listsfrom different web sources. But the names in theselists are in English, not in Indian languages. So wehave transliterated these English name lists to makethem useful for our NER task.

    Using transliteration we have constructed severallists. Which are, month name and days of the week,list of common locations, location names list, firstnames list, middle names list, surnames list etc.

    The lists can be used in name identification in var-ious ways. One way is to check whether a token is inany list. But this approach is not good as it has somelimitations. Some words may present in two or moregazetteer lists. Confusions arise to make decisionsfor these words. Some words are in gazetteer listsbut sometimes these are used in text as not-name en-tity. We have used these gazetteer lists as features ofMaxEnt. We have prepared several binary featureswhich are defined as whether a given word is in aparticular list.

    7 Detection of Nested Entities

    The training corpora used for the models, are notannotated as nested. The maximal entities are an-notated in the training corpus. For detection of thenested NEs, we have derived some rules. For exam-ple, if a particular word is a number or numeric wordand is a part of a NE type other than ‘number’, thenwe have made the nesting. Again, if any common lo-cation identifier word like,jilA (district), shahara(town) etc. is a part of a ‘location’ entity then wehave nested there. During one-level NE identifica-tion, we have generated lists for all the identified lo-cation and person names. Then we have searchedother NEs containing these as substring to make the

    21

  • nesting. After preparing the one-level NER system,we have applied the derived rules on it to identifythe nested entities.

    8 Evaluation

    The accuracies of the system are measured in termsof the f-measure, which is the weighted harmonicmean of precision and recall. Nested, maximal andlexical accuracies are calculated separately. Thetest data for all the five languages are provided.The size of the shared task test files are: Hindi- 38,704 words, Bengali - 32,796 words, Oriya -26,988 words, Telugu - 7,076 words and Urdu -12,805 words.

    We have already mentioned that after preparinga one-level NER system, the rule-based module isused to modify it to a nested one. A number of ex-periments are conducted considering various combi-nations of features to identify the best feature set forIndian language NER task. It is very difficult andtime consuming to conduct experiments for all thelanguages. During the development we have con-ducted all the experiments on Hindi and Bengali. Wehave prepared a development test data composed of24,265 words for Hindi and 10,902 word for Ben-gali and accuracies of the system are tested on thedevelopment data. The details of the experiments onHindi data for the best feature selection is describedin the following section.

    8.1 Best Feature Set Selection

    The performance of the system on the Hindi datausing various features are presented in Table 1.They are summarized below. While experimentingwith static word features, we have observed that awindow of previous two words to next two words(Wi−2...Wi+2) gives best results. But when sev-eral other features are combined then smaller win-dow (Wi−1...Wi+1) performs better. Similarly wehave experimented with suffixes of different lengthsand observed that the suffixes of length≤ 2 givesthe best result for the Hindi NER task. In usingPOS information, we have observed that the coarse-grained POS tagger information is more effectivethan the finer-grained POS values. The most in-teresting fact we have observed that more complexfeatures do not guarantee to achieve better results.

    For example, a feature set combined with currentand surrounding words, previous NE tag and fixedlength suffix information, gives a f-value 64.17%.But when prefix information are added the f-valuedecreased to 63.73%. Again when the context listsare added to the feature set containing words, previ-ous tags, suffix information, digit information andthe NomPSP binary feature, the accuracy has de-creased to 67.33% from 68.0%.

    Feature OverallF-value

    Word, NE Tag 58.92Word, NE Tag, Suffix(≤ 2) 64.17Word, NE Tag, Suffix(≤ 2),Prefix

    63.73

    Word, NE Tag, Digit, Suffix 66.61Word, NE Tag, Context List 63.57Word, NE Tag, POS (full) 61.28Word, NE Tag, Suffix(≤ 2),Digit, NomPSP

    68.60

    Word, NE Tag, Suffix(≤ 2),Digit, Context List, NomPSP

    67.33

    Word, NE Tag, Suffix (≤2), Digit, NomPSP, Linguis-tic Rules

    73.40

    Word, NE Tag, Suffix(≤ 2),Digit, NomPSP, Gazetteers

    72.08

    Word, NE Tag, Suffix (≤2), Digit, NomPSP, Linguis-tic Rules, Gazetteers

    74.53

    Table 1: Hindi development set f-values for differentfeatures

    The feature set containing words, previoustags, suffix information, digit information and theNomPSP binary feature is the identified best featureset without linguistic rules and gazetteer informa-tion. Then we have added the linguistic rules, pat-terns and gazetteer information to the system and thechanges in accuracies are shown in the table.

    8.2 Results on the Test Data

    The best identified feature set is used for the de-velopment of the NER systems for all the five lan-guages. We have already mentioned that for onlyfor Bengali and Hindi we have added linguistic rules

    22

  • and gazetteer lists in the MaxEnt based NER sys-tems. The accuracy of the system on the shared tasktest data for all the languages are shown in Table 2.

    Lan-guage

    Type Preci-sion

    Recall F-measure

    BengaliMaximal 52.92 68.07 59.54Nested 55.02 68.43 60.99Lexical 62.30 70.07 65.96

    HindiMaximal 75.19 58.94 66.08Nested 79.58 58.61 67.50Lexical 82.76 53.69 65.13

    OriyaMaximal 21.17 26.92 23.70Nested 27.73 28.13 27.93Lexical 51.51 39.40 44.65

    TeluguMaximal 10.47 9.64 10.04Nested 22.05 13.16 16.48Lexical 25.23 14.91 18.74

    UrduMaximal 26.12 29.69 27.79Nested 27.99 29.21 28.59Lexical 37.58 33.58 35.47

    Table 2: Accuracy of the system for all languages

    The accuracies of Oriya, Telugu and Urdu lan-guages are poor compared to the other two lan-guages. The reasons are POS information, mor-phological information, language specific rules andgazetteers are not used for these languages. Also thesize of training data for these languages are smaller.To mention, for Urdu, size of the training data is onlyabout 36K words which is very small to train a Max-Ent model.

    It is mentioned that we have prepared a set of ruleswhich are capable of identifying the nested NEs.Once the one-level NER system has built, we haveapplied the rules on it. In Table 3 we have shownthe f-values of each class after addition of the nestedrules. The detailed results for all languages are notshown. In the table we have shown only the resultsof Bengali and Hindi.

    For both the languages ‘title-person’ and ‘desig-nation’ classes are suffering from poor accuracies.The reason is, in the training data and also in theannotated test data, these classes contains many an-notation errors. Also the classes being closely re-lated to each other, the system fails to distinguishthem properly. The detection of the ‘term’ class is

    Hindi BengaliClass Maximal Nested Maximal NestedPerson 70.87 71.00 77.45 79.09Desig-nation

    48.98 59.81 26.32 26.32

    Organi-zation

    47.22 47.22 41.43 71.43

    Abbre-viation

    - 72.73 51.61 51.61

    Brand - - - -Title-person

    - 60.00 5.19 47.61

    Title-object

    41.32 40.98 72.97 72.97

    Location 86.02 87.02 76.27 76.27Time 67.42 67.42 56.30 56.30Number 84.59 85.13 40.65 40.65Measure 59.26 55.17 62.50 62.50Term 48.91 50.51 43.67 43.67

    Table 3: Comparison of maximal and nested f-values for different classes of Hindi and Bengali

    very difficult. In the test files amount of ‘term’ en-tity is large, for Bengali - 434 and for Hindi - 1080,so the poor accuracy of the class affects badly to theoverall accuracy. We have made rule-based identi-fication for ‘number’, ‘measure’ and ‘time’ classes;the accuracies of these classes proves that the rulesneed to be modified to achieve better accuracy forthese classes. Also the accuracy of the ‘organiza-tion’ class is not high, because amount of organiza-tion entities is not sufficient in the training corpus.We have achieved good results for other two mainclasses - ‘person’ and ‘location’.

    8.3 Comparison with Other Shared TaskSystems

    The comparison of the accuracies of our systemand other shared task systems is given in Table 4.From the comparison we can see that our systemhas achieved the best accuracies for most of the lan-guages.

    9 Conclusion

    We have prepared a MaxEnt based system for theNER task in Indian languages. We have also added

    23

  • Lan-guage

    Our S2 S6 S7

    Bengali 65.96 39.77 40.63 59.39Hindi 65.13 46.84 50.06 33.12Oriya 44.65 45.84 39.04 28.71Telugu 18.74 46.58 40.94 4.75Urdu 35.47 44.73 43.46 35.52

    Table 4: Comparison of our lexical f-measure accu-racies with the systems : S2 - Praveen P.(2008), S6 -Gali et al.(2008) and S7 - Ekbal et al.(2008)

    rules and gazetteers for Bengali and Hindi. Also ourderived rules need to be modified for improvementof the system. We have not made use of rules andgazetteers for Oriya, Telugu and Urdu. As the sizeof training data is not much for these 3 languages,rules and gazetteers would be effective. We haveexperimented with MaxEnt model only, other MLmethods like HMM, CRF or MEMM may be ableto give better accuracy. We have not worked muchon the detection of nested NEs. Proper detection ofnested entities may lead to further improvement ofperformance and is under investigation.

    References

    Bikel Daniel M., Miller Scott, Schwartz Richard andWeischedel Ralph. 1997. Nymble: A High Perfor-mance Learning Name-finder. InProceedings of theFifth Conference on Applied Natural Language Pro-cessing, 194–201.

    Borthwick Andrew. 1999. A Maximum Entropy Ap-proach to Named Entity Recognition.Ph.D. thesis,Computer Science Department, New York University.

    Cucerzan Silviu and Yarowsky David. 1999. LanguageIndependent Named Entity Recognition CombiningMorphological and Contextual Evidence. InProceed-ings of the Joint SIGDAT Conference on EMNLP andVLC 1999, 90–99.

    Ekbal A. and Bandyopadhyay S. 2007. Lexical PatternLearning from Corpus Data for Named Entity Recog-nition. In Proceedings of International Conference onNatural Language Processing (ICON), 2007.

    Ekbal A., Haque R., Das A., Poka V. and Bandyopad-hyay S. 2008. Language Independent Named EntityRecognition in Indian Languages InProceedings ofIJCNLP workshop on NERSSEAL. (Accepted)

    Gali K., Surana H., Vaidya A., Shishtla P. and MisraSharma D. 2008. Named Entity Recognition throughCRF Based Machine Learning and Language SpecificHeuristics InProceedings of IJCNLP workshop onNERSSEAL. (Accepted)

    Grishman Ralph. 1995. The New York University Sys-tem MUC-6 or Where’s the syntax? InProceedings ofthe Sixth Message Understanding Conference.

    Gu B. 2006. Recognizing Nested Named Entities in GE-NIA corpus. InProceedings of the BioNLP Workshopon Linking Natural Language Processing and Biologyat HLT-NAACL 06, pages 112-113.

    Kumar N. and Bhattacharyya Pushpak. 2006. NamedEntity Recognition in Hindi using MEMM. InTechni-cal Report, IIT Bombay, India..

    Li Wei and McCallum Andrew. 2004. Rapid Develop-ment of Hindi Named Entity Recognition using Condi-tional Random Fields and Feature Induction (Short Pa-per). InACM Transactions on Computational Logic.

    McDonald D. 1996. Internal and external evidence in theidentification and semantic categorization of propernames. InB. Boguraev and J. Pustejovsky, editors,Corpus Processing for Lexical Acquisition, 21–39.

    McDonald R., Crammer K. and Pereira F. 2005. Flexibletext segmentation with structured multilabel classifica-tion. In Proceedings of EMNLP05.

    Praveen P. 2008. Hybrid Named Entity Recogni-tion System for South-South East Indian Languages.InProceedings of IJCNLP workshop on NERSSEAL.(Accepted)

    Srihari R., Niu C. and Li W. 2000. A Hybrid Approachfor Named Entity and Sub-Type Tagging. InProceed-ings of the sixth conference on Applied natural lan-guage processing.

    Talukdar Pratim P., Brants T., Liberman M., and PereiraF. 2006. A context pattern induction methodfor named entity extraction. InProceedings of theTenth Conference on Computational Natural Lan-guage Learning (CoNLL-X).

    Wakao T., Gaizauskas R. and Wilks Y. 1996. Evaluationof an algorithm for the recognition and classificationof proper names. InProceedings of COLING-96.

    Zhou G., Zhang J., Su J., Shen D. and Tan C. 2004.Recognizing Names in Biomedical Texts: a MachineLearning Approach. Bioinformatics, 20(7):1178-1190.

    24

  • Proceedings of the IJCNLP-08 Workshop on NER for South and South East Asian Languages, pages 25–32,Hyderabad, India, January 2008. c©2008 Asian Federation of Natural Language Processing

    Aggregating Machine Learning and Rule Based Heuristics for Named Entity Recognition

    Karthik Gali, Harshit Surana, Ashwini Vaidya, Praneeth Shishtla and Dipti Misra Sharma

    Language Technologies Research Centre, International Institute of Information Technology,

    Hyderabad, India. [email protected], [email protected],

    [email protected], [email protected], [email protected]

    Abstract

    This paper, submitted as an entry for the NERSSEAL-2008 shared task, describes a system build for Named Entity Recognition for South and South East Asian Languages. Our paper combines machine learning techniques with language specific heuris-tics to model the problem of NER for In-dian languages. The system has been tested on five languages: Telugu, Hindi, Bengali, Urdu and Oriya. It uses CRF (Conditional Random Fields) based machine learning, followed by post processing which in-volves using some heuristics or rules. The system is specifically tuned for Hindi and Telugu, we also report the results for the other four languages.

    1 Introduction Named Entity Recognition (NER) is a task that seeks to locate and classify entities (‘atomic ele-ments’) in a text into predefined categories such as the names of persons, organizations, locations, ex-pressions of times, quantities, etc. It can be viewed as a two stage process: 1. Identification of entity boundaries 2. Classification into the correct category

    For example, if “Mahatma Gandhi” is a named

    entity in the corpus, it is necessary to identify the beginning and the end of this entity in the sentence. Following this step, the entity must be classified

    into the predefined category, which is NEP (Named Entity Person) in this case.

    This task is the precursor for many natural lan-guage processing applications. It has been used in Question Answering (Toral et al, 2005) as well as Machine Translation (Babych et al, 2004).

    The NERSSEAL contest has used 12 c