text
stringlengths
0
203
employed memory-based learning (De Meulder and
Daelemans, 2003; Hendrickx and Van den Bosch,
2003). Transformation-based learning (Florian et
al., 2003), Support Vector Machines (Mayfield et al.,
2003) and Conditional Random Fields (McCallum
and Li, 2003) were applied by one system each.
Combination of different learning systems has
proven to be a good method for obtaining excellent
results. Five participating groups have applied system combination. Florian et al. (2003) tested different methods for combining the results of four systems and found that robust risk minimization worked
best. Klein et al. (2003) employed a stacked learning system which contains Hidden Markov Models,
Maximum Entropy Models and Conditional Markov
Models. Mayfield et al. (2003) stacked two learners
and obtained better performance. Wu et al. (2003)
applied both stacking and voting to three learners.
Munro et al. (2003) employed both voting and bagging for combining classifiers.
3.2
Features
The choice of the learning approach is important for
obtaining a good system for recognizing named entities. However, in the CoNLL-2002 shared task we
found out that choice of features is at least as important. An overview of some of the types of features
chosen by the shared task participants, can be found
in Table 3.
All participants used lexical features (words) except for Whitelaw and Patrick (2003) who implemented a character-based method. Most of the systems employed part-of-speech tags and two of them
have recomputed the English tags with better taggers (Hendrickx and Van den Bosch, 2003; Wu et al.,
2003). Othographic information, affixes, gazetteers
and chunk information were also incorporated in
most systems although one group reports that the
available chunking information did not help (Wu et
al., 2003) Other features were used less frequently.
Table 3 does not reveal a single feature that would
be ideal for named entity recognition.
3.3
External resources
Eleven of the sixteen participating teams have attempted to use information other than the training
data that was supplied for this shared task. All included gazetteers in their systems. Four groups examined the usability of unannotated data, either for
extracting training instances (Bender et al., 2003;
Hendrickx and Van den Bosch, 2003) or obtaining
extra named entities for gazetteers (De Meulder and
Daelemans, 2003; McCallum and Li, 2003). A reasonable number of groups have also employed unannotated data for obtaining capitalization features for
words. One participating team has used externally
trained named entity recognition systems for English
as a part in a combined system (Florian et al., 2003).
Table 4 shows the error reduction of the systems
Zhang
Florian
Chieu
Hammerton
Carreras (a)
Hendrickx
De Meulder
Bender
Curran
McCallum
Wu
G
+
+
+
+
+
+
+
+
+
+
+
U
+
+
+
+
-
E
+
-
English
19%
27%
17%
22%
12%
7%
8%
3%
1%
?
?
German
15%