next up previous
Next: Error Analysis Up: Learning Rules and Their Previous: Second Experiment: Using Prior


Shared Task CoNLL'00

Table 10 presents the complete results of the shared task of CoNLL'00 (chunking). The task itself and the data are described in [Sang(2000a)]. ALLiS provides state-of-the art results. All the systems do not use exactly the same features, in particular the size of the left context varies from three words (ALLiS), five in [Sang(2000b)], six in [Koeling(2000)], until eight in some cases in [van Halteren(2000)]. A comparison using exactly the same context would discriminate more properly the different approaches. We can note that, even with the smallest context (three word), ALLiS performs better than some systems with larger contexts like Maximum Entropy used in [Koeling(2000)].

Note also that three out of four systems performing better than ALLiS use a system combination or voting approach. The only ``simple'' system performing better than ALLiS is the one using the Support Vector Machine [, see][]svm_conll00. They use a context of two words after and before the current word, a context larger than this used by ALLiS, but they also add in the left context information about the chunk tag. This information is dynamically computed during the tagging step since the chunk tags are not given by the test data. Such information can not be currently used by ALLiS, and its integration would require a modification not implemented for the time being, but scheduled. Experiments done by others participants show that this information improves results.

[Nerbonne et al.(2001)Nerbonne, Belz, Cancedda, Déjean, Hammerton, Koeling, Konstantopoulos, Osborne, Thollard, and Tjong Kim Sang] presents other systems that were applied on based-NP chunking. We can note that ALLiS outperforms decision tree, well known for handling noisy data [, see][]Quinlan_86.


Table 10: CoNLL'00 shared task: chunking
test data precision recall F$_{\beta=1}$
Kudoh and Matsumoto 93.45 93 51 93.48
Van Halteren 93.13 93.51 93.32
Tjong Kim Sang 94.04 91.00 92.50
Zhou, Tey and Su 91.99 92.25 92.12
ALLiS 91.87 92.31 92.09
Koeling 92.08 91.86 91.97
Osborne 91.65 92.33 91.64
Veenstra and Van den Bosch 91.05 92.03 91.54
Pla, Molina and Prieto 90.63 88.25 85.76
Johansson 86.24 88.25 87.23
Vilain and Day 88.82 82.91 85.76
baseline 72.58 82.14 77.07




Subsections
next up previous
Next: Error Analysis Up: Learning Rules and Their Previous: Second Experiment: Using Prior
Hammerton J. 2002-03-13