We provide a complete overview of the results of the HIPE-2022 evaluation campaign. 🏁
For more information about the tasks and challenges, see this schema and the participation guidelines pages 13-18.
Team keys (in numerical order)
Team key | Team name | Team affiliation |
---|---|---|
team1 | HISTeria | Ludwig-Maximilians-Universität and Bayerische Staatsbibliothek München, Munich, Germany |
team2 | L3i | La Rochelle University, La Rochelle, France |
team3 | WLV | University of Wolverhampton, Wolverhampton, UK |
team4 | aauzh | Machine Learning MA class, Zurich University, Switzerland |
team5 | SBB | Berlin State Library, Berlin, Germany |
We provide an overview table of the preliminary results of the runs submitted by the teams.
It also includes a neural baseline created by the organizers. A non-neural CRF-based baseline will be added soon for NERC.
- Date: 24.05.2022.
- Bundles: 1 to 5
- HIPE-2022 data version: v2.1-test-all-unmasked
- HIPE scorer version: v2.0
- The current results for NEL can still change as we may extend the list of equivalent wikidata IDs.
- Detailed results for all systems can be found in the corresponding .tsv file (link provided below each table).
- Detailed results for each team's runs are sent privately.
- System name composition is:
teamID_bundle_dataset_lang_run
. - F1 scores of 0.0 are excluded from the table.
- Results are ordered by F1 scores.
-
NERC and Entity Linking (EL) are evaluated in terms of macro and micro Precision, Recall, F1-measure. Here only micro is reported.
-
NERC evaluation includes strict (exact boundary matching) and fuzzy (boundary overlap) scenarios.
-
For EL, the definition of strict and fuzzy regimes differs. In terms of boundaries, EL is always evaluated according to overlapping boundaries (what is of interest is the capacity to provide the correct link rather than the correct boundaries). EL strict regime considers only the system's top link prediction (NIL or QID), while the fuzzy/relaxed regime expands system predictions with a set of historically related entity QIDs. We additionally report F@1/3/5 in the .tsv files.
Relevant bundles: 1-4
NERC-Coarse hipe2020 German strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle3_hipe2020_de_1 | 0.794 | 0.784 | 0.805 | 923 | 254 | 224 |
2 | team2_bundle1_hipe2020_de_2 | 0.784 | 0.780 | 0.787 | 903 | 255 | 244 |
3 | team2_bundle1_hipe2020_de_1 | 0.774 | 0.757 | 0.792 | 908 | 292 | 239 |
4 | team4_bundle4_hipe2020_de_1 | 0.725 | 0.716 | 0.735 | 843 | 335 | 304 |
5 | neurbsl_bundle3_hipe2020_de_1 | 0.703 | 0.665 | 0.746 | 856 | 432 | 291 |
6 | team4_bundle4_hipe2020_de_2 | 0.695 | 0.677 | 0.714 | 819 | 391 | 328 |
See ranking-hipe2020-de-coarse-micro-strict-all.tsv for full details.
NERC-Coarse hipe2020 German fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle3_hipe2020_de_1 | 0.876 | 0.865 | 0.888 | 1018 | 159 | 129 |
2 | team2_bundle1_hipe2020_de_2 | 0.874 | 0.870 | 0.878 | 1007 | 151 | 140 |
3 | team2_bundle1_hipe2020_de_1 | 0.872 | 0.853 | 0.892 | 1023 | 177 | 124 |
4 | team4_bundle4_hipe2020_de_1 | 0.822 | 0.812 | 0.833 | 956 | 222 | 191 |
5 | team4_bundle4_hipe2020_de_2 | 0.804 | 0.783 | 0.826 | 947 | 263 | 200 |
6 | neurbsl_bundle3_hipe2020_de_1 | 0.793 | 0.750 | 0.842 | 966 | 322 | 181 |
See ranking-hipe2020-de-coarse-micro-fuzzy-all.tsv for full details.
NERC-Coarse hipe2020 English strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_en_1 | 0.620 | 0.624 | 0.617 | 277 | 167 | 172 |
2 | team2_bundle1_hipe2020_en_2 | 0.612 | 0.604 | 0.619 | 278 | 182 | 171 |
3 | team4_bundle4_hipe2020_en_1 | 0.513 | 0.538 | 0.490 | 220 | 189 | 229 |
4 | neurbsl_bundle3_hipe2020_en_1 | 0.477 | 0.432 | 0.532 | 239 | 314 | 210 |
5 | team3_bundle4_hipe2020_en_1 | 0.414 | 0.400 | 0.430 | 193 | 290 | 256 |
See ranking-hipe2020-en-coarse-micro-strict-all.tsv for full details.
NERC-Coarse hipe2020 English fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_en_1 | 0.788 | 0.793 | 0.784 | 352 | 92 | 97 |
2 | team2_bundle1_hipe2020_en_2 | 0.781 | 0.772 | 0.791 | 355 | 105 | 94 |
3 | team4_bundle4_hipe2020_en_1 | 0.692 | 0.726 | 0.661 | 297 | 112 | 152 |
4 | neurbsl_bundle3_hipe2020_en_1 | 0.623 | 0.564 | 0.695 | 312 | 241 | 137 |
5 | team3_bundle4_hipe2020_en_1 | 0.603 | 0.582 | 0.626 | 281 | 202 | 168 |
See ranking-hipe2020-en-coarse-micro-fuzzy-all.tsv for full details.
NERC-Coarse hipe2020 French strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_fr_1 | 0.808 | 0.786 | 0.831 | 1329 | 361 | 271 |
2 | team2_bundle1_hipe2020_fr_2 | 0.804 | 0.782 | 0.827 | 1323 | 368 | 277 |
3 | team2_bundle3_hipe2020_fr_1 | 0.798 | 0.775 | 0.823 | 1316 | 381 | 284 |
4 | neurbsl_bundle3_hipe2020_fr_1 | 0.757 | 0.730 | 0.785 | 1256 | 464 | 344 |
5 | team4_bundle4_hipe2020_fr_2 | 0.696 | 0.718 | 0.675 | 1080 | 425 | 520 |
6 | team4_bundle4_hipe2020_fr_1 | 0.678 | 0.700 | 0.657 | 1051 | 451 | 549 |
7 | team3_bundle4_hipe2020_fr_1 | 0.674 | 0.640 | 0.712 | 1139 | 640 | 461 |
See ranking-hipe2020-fr-coarse-micro-strict-all.tsv for full details.
NERC-Coarse hipe2020 French fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_fr_2 | 0.907 | 0.883 | 0.933 | 1493 | 198 | 107 |
2 | team2_bundle1_hipe2020_fr_1 | 0.904 | 0.880 | 0.929 | 1487 | 203 | 113 |
3 | team2_bundle3_hipe2020_fr_1 | 0.901 | 0.875 | 0.928 | 1485 | 212 | 115 |
4 | neurbsl_bundle3_hipe2020_fr_1 | 0.866 | 0.836 | 0.899 | 1438 | 282 | 162 |
5 | team3_bundle4_hipe2020_fr_1 | 0.808 | 0.767 | 0.853 | 1365 | 414 | 235 |
6 | team4_bundle4_hipe2020_fr_2 | 0.800 | 0.825 | 0.776 | 1242 | 263 | 358 |
7 | team4_bundle4_hipe2020_fr_1 | 0.798 | 0.824 | 0.773 | 1237 | 265 | 363 |
See ranking-hipe2020-fr-coarse-micro-fuzzy-all.tsv for full details.
NERC-Coarse newseye German strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | neurbsl_bundle3_newseye_de_1 | 0.477 | 0.429 | 0.537 | 1291 | 1721 | 1111 |
2 | team4_bundle4_newseye_de_2 | 0.408 | 0.395 | 0.421 | 1012 | 1550 | 1390 |
3 | team4_bundle4_newseye_de_1 | 0.395 | 0.396 | 0.394 | 947 | 1444 | 1455 |
See ranking-newseye-de-coarse-micro-strict-all.tsv for full details.
NERC-Coarse newseye German fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | neurbsl_bundle3_newseye_de_1 | 0.570 | 0.512 | 0.642 | 1542 | 1470 | 860 |
2 | team4_bundle4_newseye_de_2 | 0.495 | 0.480 | 0.512 | 1229 | 1333 | 1173 |
3 | team4_bundle4_newseye_de_1 | 0.479 | 0.481 | 0.478 | 1149 | 1242 | 1253 |
See ranking-newseye-de-coarse-micro-fuzzy-all.tsv for full details.
NERC-Coarse newseye French strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team4_bundle4_newseye_fr_2 | 0.656 | 0.655 | 0.657 | 1661 | 876 | 869 |
2 | neurbsl_bundle3_newseye_fr_1 | 0.654 | 0.634 | 0.676 | 1710 | 989 | 820 |
3 | team4_bundle4_newseye_fr_1 | 0.648 | 0.673 | 0.625 | 1582 | 768 | 948 |
See ranking-newseye-fr-coarse-micro-strict-all.tsv for full details.
NERC-Coarse newseye French fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team4_bundle4_newseye_fr_2 | 0.786 | 0.785 | 0.787 | 1992 | 545 | 538 |
2 | neurbsl_bundle3_newseye_fr_1 | 0.779 | 0.755 | 0.805 | 2037 | 662 | 493 |
3 | team4_bundle4_newseye_fr_1 | 0.772 | 0.801 | 0.744 | 1883 | 467 | 647 |
See ranking-newseye-fr-coarse-micro-fuzzy-all.tsv for full details.
NERC-Coarse newseye Swedish strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | neurbsl_bundle3_newseye_sv_1 | 0.651 | 0.588 | 0.728 | 440 | 308 | 164 |
2 | team4_bundle4_newseye_sv_1 | 0.643 | 0.686 | 0.604 | 365 | 167 | 239 |
3 | team4_bundle4_newseye_sv_2 | 0.636 | 0.673 | 0.603 | 364 | 177 | 240 |
See ranking-newseye-sv-coarse-micro-strict-all.tsv for full details.
NERC-Coarse newseye Swedish fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | neurbsl_bundle3_newseye_sv_1 | 0.747 | 0.675 | 0.836 | 505 | 243 | 99 |
2 | team4_bundle4_newseye_sv_1 | 0.746 | 0.797 | 0.702 | 424 | 108 | 180 |
3 | team4_bundle4_newseye_sv_2 | 0.742 | 0.786 | 0.704 | 425 | 116 | 179 |
See ranking-newseye-sv-coarse-micro-fuzzy-all.tsv for full details.
NERC-Coarse newseye Finnish strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | neurbsl_bundle3_newseye_fi_1 | 0.644 | 0.605 | 0.687 | 475 | 310 | 216 |
2 | team4_bundle4_newseye_fi_1 | 0.567 | 0.618 | 0.524 | 362 | 224 | 329 |
3 | team4_bundle4_newseye_fi_2 | 0.556 | 0.592 | 0.524 | 362 | 250 | 329 |
See ranking-newseye-fi-coarse-micro-strict-all.tsv for full details.
NERC-Coarse newseye Finnish fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | neurbsl_bundle3_newseye_fi_1 | 0.760 | 0.715 | 0.812 | 561 | 224 | 130 |
2 | team4_bundle4_newseye_fi_1 | 0.670 | 0.730 | 0.619 | 428 | 158 | 263 |
3 | team4_bundle4_newseye_fi_2 | 0.640 | 0.681 | 0.603 | 417 | 195 | 274 |
See ranking-newseye-fi-coarse-micro-fuzzy-all.tsv for full details.
NERC-Coarse letemps French strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | neurbsl_bundle3_letemps_fr_1 | 0.661 | 0.595 | 0.744 | 757 | 516 | 260 |
2 | team4_bundle4_letemps_fr_1 | 0.644 | 0.589 | 0.710 | 722 | 503 | 295 |
3 | team4_bundle4_letemps_fr_2 | 0.622 | 0.557 | 0.704 | 716 | 570 | 301 |
4 | team3_bundle4_letemps_fr_1 | 0.618 | 0.581 | 0.659 | 670 | 483 | 347 |
See ranking-letemps-fr-coarse-micro-strict-all.tsv for full details.
NERC-Coarse letemps French fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | neurbsl_bundle3_letemps_fr_1 | 0.711 | 0.639 | 0.800 | 814 | 459 | 203 |
2 | team4_bundle4_letemps_fr_1 | 0.701 | 0.642 | 0.773 | 786 | 439 | 231 |
3 | team4_bundle4_letemps_fr_2 | 0.681 | 0.610 | 0.771 | 784 | 502 | 233 |
4 | team3_bundle4_letemps_fr_1 | 0.666 | 0.627 | 0.711 | 723 | 430 | 294 |
See ranking-letemps-fr-coarse-micro-fuzzy-all.tsv for full details.
NERC-Coarse sonar German strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team4_bundle4_sonar_de_1 | 0.529 | 0.512 | 0.548 | 258 | 246 | 213 |
2 | team4_bundle4_sonar_de_2 | 0.516 | 0.486 | 0.550 | 259 | 274 | 212 |
3 | neurbsl_bundle3_sonar_de_1 | 0.307 | 0.267 | 0.361 | 170 | 467 | 301 |
See ranking-sonar-de-coarse-micro-strict-all.tsv for full details.
NERC-Coarse sonar German fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team4_bundle4_sonar_de_2 | 0.695 | 0.655 | 0.741 | 349 | 184 | 122 |
2 | team4_bundle4_sonar_de_1 | 0.693 | 0.671 | 0.718 | 338 | 166 | 133 |
3 | neurbsl_bundle3_sonar_de_1 | 0.471 | 0.410 | 0.554 | 261 | 376 | 210 |
See ranking-sonar-de-coarse-micro-fuzzy-all.tsv for full details.
NERC-Coarse topres19th English strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team4_bundle4_topres19th_en_1 | 0.787 | 0.816 | 0.760 | 902 | 204 | 285 |
2 | team4_bundle4_topres19th_en_2 | 0.781 | 0.761 | 0.802 | 952 | 299 | 235 |
3 | neurbsl_bundle3_topres19th_en_1 | 0.764 | 0.747 | 0.782 | 928 | 315 | 259 |
4 | team3_bundle4_topres19th_en_1 | 0.740 | 0.712 | 0.771 | 915 | 371 | 272 |
See ranking-topres19th-en-coarse-micro-strict-all.tsv for full details.
NERC-Coarse topres19th English fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team4_bundle4_topres19th_en_1 | 0.838 | 0.869 | 0.810 | 961 | 145 | 226 |
2 | team4_bundle4_topres19th_en_2 | 0.829 | 0.807 | 0.851 | 1010 | 241 | 177 |
3 | neurbsl_bundle3_topres19th_en_1 | 0.816 | 0.798 | 0.836 | 992 | 251 | 195 |
4 | team3_bundle4_topres19th_en_1 | 0.796 | 0.765 | 0.829 | 984 | 302 | 203 |
See ranking-topres19th-en-coarse-micro-fuzzy-all.tsv for full details.
NERC-Coarse ajmc German strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle3_ajmc_de_2 | 0.934 | 0.946 | 0.921 | 352 | 20 | 30 |
2 | team1_bundle4_ajmc_de_1 | 0.913 | 0.930 | 0.898 | 343 | 26 | 39 |
3 | team1_bundle4_ajmc_de_2 | 0.912 | 0.905 | 0.919 | 351 | 37 | 31 |
4 | team2_bundle3_ajmc_de_1 | 0.908 | 0.913 | 0.903 | 345 | 33 | 37 |
5 | neurbsl_bundle3_ajmc_de_1 | 0.818 | 0.792 | 0.846 | 323 | 85 | 59 |
See ranking-ajmc-de-coarse-micro-strict-all.tsv for full details.
NERC-Coarse ajmc German fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle3_ajmc_de_2 | 0.952 | 0.965 | 0.940 | 359 | 13 | 23 |
2 | team1_bundle4_ajmc_de_2 | 0.945 | 0.938 | 0.953 | 364 | 24 | 18 |
3 | team1_bundle4_ajmc_de_1 | 0.937 | 0.954 | 0.921 | 352 | 17 | 30 |
4 | team2_bundle3_ajmc_de_1 | 0.934 | 0.939 | 0.929 | 355 | 23 | 27 |
5 | neurbsl_bundle3_ajmc_de_1 | 0.873 | 0.846 | 0.903 | 345 | 63 | 37 |
See ranking-ajmc-de-coarse-micro-fuzzy-all.tsv for full details.
NERC-Coarse ajmc English strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team1_bundle4_ajmc_en_2 | 0.854 | 0.826 | 0.885 | 308 | 65 | 40 |
2 | team2_bundle1_ajmc_en_1 | 0.850 | 0.824 | 0.876 | 305 | 65 | 43 |
3 | team2_bundle1_ajmc_en_2 | 0.841 | 0.831 | 0.851 | 296 | 60 | 52 |
4 | team1_bundle4_ajmc_en_1 | 0.819 | 0.783 | 0.859 | 299 | 83 | 49 |
5 | neurbsl_bundle3_ajmc_en_1 | 0.736 | 0.680 | 0.802 | 279 | 131 | 69 |
See ranking-ajmc-en-coarse-micro-strict-all.tsv for full details.
NERC-Coarse ajmc English fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team1_bundle4_ajmc_en_2 | 0.910 | 0.879 | 0.943 | 328 | 45 | 20 |
2 | team1_bundle4_ajmc_en_1 | 0.899 | 0.859 | 0.943 | 328 | 54 | 20 |
3 | team2_bundle1_ajmc_en_1 | 0.894 | 0.868 | 0.922 | 321 | 49 | 27 |
4 | team2_bundle1_ajmc_en_2 | 0.884 | 0.874 | 0.894 | 311 | 45 | 37 |
5 | neurbsl_bundle3_ajmc_en_1 | 0.828 | 0.766 | 0.902 | 314 | 96 | 34 |
See ranking-ajmc-en-coarse-micro-fuzzy-all.tsv for full details.
NERC-Coarse ajmc French strict (literal sense) [NE-COARSE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team1_bundle4_ajmc_fr_2 | 0.842 | 0.834 | 0.850 | 306 | 61 | 54 |
2 | team1_bundle4_ajmc_fr_1 | 0.833 | 0.820 | 0.847 | 305 | 67 | 55 |
3 | team2_bundle3_ajmc_fr_2 | 0.826 | 0.810 | 0.842 | 303 | 71 | 57 |
4 | team2_bundle3_ajmc_fr_1 | 0.798 | 0.780 | 0.817 | 294 | 83 | 66 |
5 | neurbsl_bundle3_ajmc_fr_1 | 0.741 | 0.707 | 0.778 | 280 | 116 | 80 |
See ranking-ajmc-fr-coarse-micro-strict-all.tsv for full details.
NERC-Coarse ajmc French fuzzy (literal sense) [NE-COARSE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team1_bundle4_ajmc_fr_1 | 0.888 | 0.874 | 0.903 | 325 | 47 | 35 |
2 | team1_bundle4_ajmc_fr_2 | 0.880 | 0.872 | 0.889 | 320 | 47 | 40 |
3 | team2_bundle3_ajmc_fr_2 | 0.872 | 0.856 | 0.889 | 320 | 54 | 40 |
4 | team2_bundle3_ajmc_fr_1 | 0.860 | 0.841 | 0.881 | 317 | 60 | 43 |
5 | neurbsl_bundle3_ajmc_fr_1 | 0.825 | 0.788 | 0.867 | 312 | 84 | 48 |
See ranking-ajmc-fr-coarse-micro-fuzzy-all.tsv for full details.
Relevant bundles: 1, 3
NERC-Fine hipe2020 German strict (literal sense) [NE-FINE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle3_hipe2020_de_1 | 0.718 | 0.691 | 0.747 | 857 | 384 | 290 |
2 | team2_bundle1_hipe2020_de_1 | 0.689 | 0.658 | 0.724 | 830 | 431 | 317 |
3 | team2_bundle1_hipe2020_de_2 | 0.682 | 0.657 | 0.710 | 814 | 425 | 333 |
4 | neurbsl_bundle3_hipe2020_de_1 | 0.625 | 0.584 | 0.673 | 772 | 550 | 375 |
See ranking-hipe2020-de-fine-micro-strict-all.tsv for full details.
NERC-Fine hipe2020 German fuzzy (literal sense) [NE-FINE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle3_hipe2020_de_1 | 0.807 | 0.776 | 0.840 | 963 | 278 | 184 |
2 | team2_bundle1_hipe2020_de_1 | 0.785 | 0.749 | 0.824 | 945 | 316 | 202 |
3 | team2_bundle1_hipe2020_de_2 | 0.783 | 0.754 | 0.814 | 934 | 305 | 213 |
4 | neurbsl_bundle3_hipe2020_de_1 | 0.706 | 0.659 | 0.759 | 871 | 451 | 276 |
See ranking-hipe2020-de-fine-micro-fuzzy-all.tsv for full details.
NERC-Fine hipe2020 German strict, nested entities [NE-NESTED-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle3_hipe2020_de_1 | 0.522 | 0.714 | 0.411 | 30 | 12 | 43 |
2 | team2_bundle1_hipe2020_de_2 | 0.457 | 0.750 | 0.329 | 24 | 8 | 49 |
3 | team2_bundle1_hipe2020_de_1 | 0.393 | 0.618 | 0.288 | 21 | 13 | 52 |
See ranking-hipe2020-de-fine-micro-strict-all.tsv for full details.
NERC-Fine hipe2020 German fuzzy, nested entities [NE-NESTED-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle3_hipe2020_de_1 | 0.539 | 0.738 | 0.425 | 31 | 11 | 42 |
2 | team2_bundle1_hipe2020_de_2 | 0.476 | 0.781 | 0.342 | 25 | 7 | 48 |
3 | team2_bundle1_hipe2020_de_1 | 0.393 | 0.618 | 0.288 | 21 | 13 | 52 |
See ranking-hipe2020-de-fine-micro-fuzzy-all.tsv for full details.
NERC-Fine hipe2020 French strict (literal sense) [NE-FINE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_fr_1 | 0.740 | 0.702 | 0.782 | 1251 | 531 | 349 |
2 | team2_bundle1_hipe2020_fr_2 | 0.736 | 0.697 | 0.779 | 1246 | 541 | 354 |
3 | team2_bundle3_hipe2020_fr_1 | 0.720 | 0.679 | 0.767 | 1227 | 581 | 373 |
4 | neurbsl_bundle3_hipe2020_fr_1 | 0.708 | 0.685 | 0.733 | 1172 | 538 | 428 |
See ranking-hipe2020-fr-fine-micro-strict-all.tsv for full details.
NERC-Fine hipe2020 French fuzzy (literal sense) [NE-FINE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_fr_1 | 0.826 | 0.784 | 0.873 | 1397 | 385 | 203 |
2 | team2_bundle1_hipe2020_fr_2 | 0.825 | 0.782 | 0.873 | 1397 | 390 | 203 |
3 | team2_bundle3_hipe2020_fr_1 | 0.812 | 0.765 | 0.865 | 1384 | 424 | 216 |
4 | neurbsl_bundle3_hipe2020_fr_1 | 0.795 | 0.769 | 0.822 | 1315 | 395 | 285 |
See ranking-hipe2020-fr-fine-micro-fuzzy-all.tsv for full details.
NERC-Fine hipe2020 French strict, nested entities [NE-NESTED-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_fr_2 | 0.377 | 0.390 | 0.366 | 30 | 47 | 52 |
2 | team2_bundle1_hipe2020_fr_1 | 0.366 | 0.394 | 0.341 | 28 | 43 | 54 |
3 | team2_bundle3_hipe2020_fr_1 | 0.320 | 0.301 | 0.341 | 28 | 65 | 54 |
See ranking-hipe2020-fr-fine-micro-strict-all.tsv for full details.
NERC-Fine hipe2020 French fuzzy, nested entities [NE-NESTED-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_fr_2 | 0.403 | 0.416 | 0.390 | 32 | 45 | 50 |
2 | team2_bundle1_hipe2020_fr_1 | 0.392 | 0.423 | 0.366 | 30 | 41 | 52 |
3 | team2_bundle3_hipe2020_fr_1 | 0.389 | 0.366 | 0.415 | 34 | 59 | 48 |
See ranking-hipe2020-fr-fine-micro-fuzzy-all.tsv for full details.
NERC-Fine letemps French strict (literal sense) [NE-FINE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | neurbsl_bundle3_letemps_fr_1 | 0.615 | 0.564 | 0.676 | 688 | 531 | 329 |
See ranking-letemps-fr-fine-micro-strict-all.tsv for full details.
NERC-Fine letemps French fuzzy (literal sense) [NE-FINE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | neurbsl_bundle3_letemps_fr_1 | 0.654 | 0.600 | 0.719 | 731 | 488 | 286 |
See ranking-letemps-fr-fine-micro-fuzzy-all.tsv for full details.
NERC-Fine ajmc German strict (literal sense) [NE-FINE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle3_ajmc_de_2 | 0.906 | 0.915 | 0.898 | 343 | 32 | 39 |
2 | team2_bundle3_ajmc_de_1 | 0.880 | 0.860 | 0.901 | 344 | 56 | 38 |
3 | neurbsl_bundle3_ajmc_de_1 | 0.818 | 0.819 | 0.817 | 312 | 69 | 70 |
See ranking-ajmc-de-fine-micro-strict-all.tsv for full details.
NERC-Fine ajmc German fuzzy (literal sense) [NE-FINE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle3_ajmc_de_2 | 0.933 | 0.941 | 0.924 | 353 | 22 | 29 |
2 | team2_bundle3_ajmc_de_1 | 0.905 | 0.885 | 0.927 | 354 | 46 | 28 |
3 | neurbsl_bundle3_ajmc_de_1 | 0.865 | 0.866 | 0.864 | 330 | 51 | 52 |
See ranking-ajmc-de-fine-micro-fuzzy-all.tsv for full details.
NERC-Fine ajmc English strict (literal sense) [NE-FINE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_ajmc_en_1 | 0.798 | 0.754 | 0.848 | 295 | 96 | 53 |
2 | team2_bundle1_ajmc_en_2 | 0.781 | 0.745 | 0.822 | 286 | 98 | 62 |
3 | neurbsl_bundle3_ajmc_en_1 | 0.664 | 0.600 | 0.744 | 259 | 173 | 89 |
See ranking-ajmc-en-fine-micro-strict-all.tsv for full details.
NERC-Fine ajmc English fuzzy (literal sense) [NE-FINE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_ajmc_en_1 | 0.847 | 0.801 | 0.899 | 313 | 78 | 35 |
2 | team2_bundle1_ajmc_en_2 | 0.847 | 0.807 | 0.891 | 310 | 74 | 38 |
3 | neurbsl_bundle3_ajmc_en_1 | 0.749 | 0.676 | 0.839 | 292 | 140 | 56 |
See ranking-ajmc-en-fine-micro-fuzzy-all.tsv for full details.
NERC-Fine ajmc French strict (literal sense) [NE-FINE-LIT-micro-strict-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle3_ajmc_fr_2 | 0.669 | 0.646 | 0.694 | 250 | 137 | 110 |
2 | team2_bundle3_ajmc_fr_1 | 0.645 | 0.623 | 0.669 | 241 | 146 | 119 |
3 | neurbsl_bundle3_ajmc_fr_1 | 0.545 | 0.526 | 0.567 | 204 | 184 | 156 |
See ranking-ajmc-fr-fine-micro-strict-all.tsv for full details.
NERC-Fine ajmc French fuzzy (literal sense) [NE-FINE-LIT-micro-fuzzy-TIME-ALL-LED-ALL
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle3_ajmc_fr_2 | 0.728 | 0.703 | 0.756 | 272 | 115 | 88 |
2 | team2_bundle3_ajmc_fr_1 | 0.710 | 0.685 | 0.736 | 265 | 122 | 95 |
3 | neurbsl_bundle3_ajmc_fr_1 | 0.639 | 0.616 | 0.664 | 239 | 149 | 121 |
See ranking-ajmc-fr-fine-micro-fuzzy-all.tsv for full details.
Relevant bundles: 1, 2
End-to-end EL hipe2020 German strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_de_2 | 0.449 | 0.446 | 0.451 | 517 | 641 | 630 |
2 | team2_bundle1_hipe2020_de_1 | 0.447 | 0.438 | 0.458 | 525 | 675 | 622 |
See ranking-hipe2020-de-nel-micro-fuzzy-all.tsv for full details.
End-to-end EL hipe2020 German relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_de_2 | 0.464 | 0.462 | 0.466 | 535 | 623 | 612 |
2 | team2_bundle1_hipe2020_de_1 | 0.463 | 0.453 | 0.473 | 543 | 657 | 604 |
See ranking-hipe2020-de-nel-micro-fuzzy-relaxed-all.tsv for full details.
End-to-end EL hipe2020 English strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_en_2 | 0.469 | 0.463 | 0.474 | 213 | 247 | 236 |
2 | team2_bundle1_hipe2020_en_1 | 0.468 | 0.471 | 0.465 | 209 | 235 | 240 |
See ranking-hipe2020-en-nel-micro-fuzzy-all.tsv for full details.
End-to-end EL hipe2020 English relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_en_2 | 0.469 | 0.463 | 0.474 | 213 | 247 | 236 |
2 | team2_bundle1_hipe2020_en_1 | 0.468 | 0.471 | 0.465 | 209 | 235 | 240 |
See ranking-hipe2020-en-nel-micro-fuzzy-relaxed-all.tsv for full details.
End-to-end EL hipe2020 French strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_fr_1 | 0.560 | 0.546 | 0.576 | 922 | 768 | 678 |
2 | team2_bundle1_hipe2020_fr_2 | 0.558 | 0.543 | 0.574 | 918 | 773 | 682 |
See ranking-hipe2020-fr-nel-micro-fuzzy-all.tsv for full details.
End-to-end EL hipe2020 French relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_hipe2020_fr_1 | 0.578 | 0.563 | 0.594 | 951 | 739 | 649 |
2 | team2_bundle1_hipe2020_fr_2 | 0.576 | 0.560 | 0.592 | 947 | 744 | 653 |
See ranking-hipe2020-fr-nel-micro-fuzzy-relaxed-all.tsv for full details.
End-to-end EL ajmc English strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_ajmc_en_2 | 0.030 | 0.022 | 0.044 | 8 | 348 | 175 |
2 | team2_bundle1_ajmc_en_1 | 0.029 | 0.022 | 0.044 | 8 | 362 | 175 |
See ranking-ajmc-en-nel-micro-fuzzy-all.tsv for full details.
End-to-end EL ajmc English relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle1_ajmc_en_2 | 0.030 | 0.022 | 0.044 | 8 | 348 | 175 |
2 | team2_bundle1_ajmc_en_1 | 0.029 | 0.022 | 0.044 | 8 | 362 | 175 |
See ranking-ajmc-en-nel-micro-fuzzy-relaxed-all.tsv for full details.
Relevant bundles: 5
EL-only hipe2020 German strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_hipe2020_de_1 | 0.506 | 0.603 | 0.435 | 499 | 328 | 648 |
2 | team5_bundle5_hipe2020_de_2 | 0.499 | 0.663 | 0.400 | 459 | 233 | 688 |
3 | team2_bundle5_hipe2020_de_1 | 0.481 | 0.481 | 0.481 | 552 | 595 | 595 |
See ranking-hipe2020-de-nel-only-micro-fuzzy-all.tsv for full details.
EL-only hipe2020 German relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_hipe2020_de_1 | 0.525 | 0.626 | 0.452 | 518 | 309 | 629 |
2 | team5_bundle5_hipe2020_de_2 | 0.517 | 0.686 | 0.414 | 475 | 217 | 672 |
3 | team2_bundle5_hipe2020_de_1 | 0.497 | 0.497 | 0.497 | 570 | 577 | 577 |
See ranking-hipe2020-de-nel-only-micro-fuzzy-relaxed-all.tsv for full details.
EL-only hipe2020 English strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle5_hipe2020_en_1 | 0.546 | 0.546 | 0.546 | 245 | 204 | 204 |
2 | team5_bundle5_hipe2020_en_2 | 0.393 | 0.503 | 0.323 | 145 | 143 | 304 |
3 | team5_bundle5_hipe2020_en_1 | 0.380 | 0.481 | 0.314 | 141 | 152 | 308 |
See ranking-hipe2020-en-nel-only-micro-fuzzy-all.tsv for full details.
EL-only hipe2020 English relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle5_hipe2020_en_1 | 0.546 | 0.546 | 0.546 | 245 | 204 | 204 |
2 | team5_bundle5_hipe2020_en_2 | 0.393 | 0.503 | 0.323 | 145 | 143 | 304 |
3 | team5_bundle5_hipe2020_en_1 | 0.380 | 0.481 | 0.314 | 141 | 152 | 308 |
See ranking-hipe2020-en-nel-only-micro-fuzzy-relaxed-all.tsv for full details.
EL-only hipe2020 French strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle5_hipe2020_fr_1 | 0.602 | 0.602 | 0.602 | 963 | 637 | 637 |
2 | team5_bundle5_hipe2020_fr_2 | 0.596 | 0.707 | 0.515 | 824 | 341 | 776 |
3 | team5_bundle5_hipe2020_fr_1 | 0.562 | 0.664 | 0.487 | 779 | 394 | 821 |
See ranking-hipe2020-fr-nel-only-micro-fuzzy-all.tsv for full details.
EL-only hipe2020 French relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team2_bundle5_hipe2020_fr_1 | 0.620 | 0.620 | 0.620 | 992 | 608 | 608 |
2 | team5_bundle5_hipe2020_fr_2 | 0.616 | 0.730 | 0.532 | 851 | 314 | 749 |
3 | team5_bundle5_hipe2020_fr_1 | 0.582 | 0.688 | 0.504 | 807 | 366 | 793 |
See ranking-hipe2020-fr-nel-only-micro-fuzzy-relaxed-all.tsv for full details.
EL-only newseye German strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_newseye_de_1 | 0.444 | 0.522 | 0.387 | 929 | 849 | 1473 |
2 | team5_bundle5_newseye_de_2 | 0.393 | 0.520 | 0.316 | 759 | 701 | 1643 |
See ranking-newseye-de-nel-only-micro-fuzzy-all.tsv for full details.
EL-only newseye German relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_newseye_de_1 | 0.455 | 0.535 | 0.396 | 951 | 827 | 1451 |
2 | team5_bundle5_newseye_de_2 | 0.405 | 0.536 | 0.326 | 782 | 678 | 1620 |
See ranking-newseye-de-nel-only-micro-fuzzy-relaxed-all.tsv for full details.
EL-only newseye French strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_newseye_fr_1 | 0.431 | 0.534 | 0.361 | 914 | 797 | 1616 |
2 | team5_bundle5_newseye_fr_2 | 0.430 | 0.528 | 0.363 | 919 | 821 | 1611 |
See ranking-newseye-fr-nel-only-micro-fuzzy-all.tsv for full details.
EL-only newseye French relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_newseye_fr_1 | 0.435 | 0.539 | 0.364 | 922 | 789 | 1608 |
2 | team5_bundle5_newseye_fr_2 | 0.433 | 0.531 | 0.365 | 924 | 816 | 1606 |
See ranking-newseye-fr-nel-only-micro-fuzzy-relaxed-all.tsv for full details.
EL-only sonar German strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_sonar_de_1 | 0.517 | 0.616 | 0.446 | 210 | 131 | 261 |
2 | team5_bundle5_sonar_de_2 | 0.503 | 0.634 | 0.416 | 196 | 113 | 275 |
See ranking-sonar-de-nel-only-micro-fuzzy-all.tsv for full details.
EL-only sonar German relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_sonar_de_1 | 0.517 | 0.616 | 0.446 | 210 | 131 | 261 |
2 | team5_bundle5_sonar_de_2 | 0.505 | 0.638 | 0.418 | 197 | 112 | 274 |
See ranking-sonar-de-nel-only-micro-fuzzy-relaxed-all.tsv for full details.
EL-only topres19th English strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_topres19th_en_2 | 0.651 | 0.778 | 0.559 | 664 | 190 | 523 |
2 | team5_bundle5_topres19th_en_1 | 0.649 | 0.786 | 0.552 | 655 | 178 | 532 |
See ranking-topres19th-en-nel-only-micro-fuzzy-all.tsv for full details.
EL-only topres19th English relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_topres19th_en_2 | 0.654 | 0.781 | 0.562 | 667 | 187 | 520 |
2 | team5_bundle5_topres19th_en_1 | 0.650 | 0.789 | 0.553 | 657 | 176 | 530 |
See ranking-topres19th-en-nel-only-micro-fuzzy-relaxed-all.tsv for full details.
EL-only ajmc German strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_ajmc_de_2 | 0.503 | 0.712 | 0.389 | 74 | 30 | 116 |
2 | team5_bundle5_ajmc_de_1 | 0.471 | 0.608 | 0.384 | 73 | 47 | 117 |
See ranking-ajmc-de-nel-only-micro-fuzzy-all.tsv for full details.
EL-only ajmc German relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_ajmc_de_2 | 0.503 | 0.712 | 0.389 | 74 | 30 | 116 |
2 | team5_bundle5_ajmc_de_1 | 0.471 | 0.608 | 0.384 | 73 | 47 | 117 |
See ranking-ajmc-de-nel-only-micro-fuzzy-relaxed-all.tsv for full details.
EL-only ajmc English strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_ajmc_en_2 | 0.381 | 0.578 | 0.284 | 52 | 38 | 131 |
2 | team5_bundle5_ajmc_en_1 | 0.376 | 0.580 | 0.279 | 51 | 37 | 132 |
See ranking-ajmc-en-nel-only-micro-fuzzy-all.tsv for full details.
EL-only ajmc English relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_ajmc_en_2 | 0.381 | 0.578 | 0.284 | 52 | 38 | 131 |
2 | team5_bundle5_ajmc_en_1 | 0.376 | 0.580 | 0.279 | 51 | 37 | 132 |
See ranking-ajmc-en-nel-only-micro-fuzzy-relaxed-all.tsv for full details.
EL-only ajmc French strict @1 (literal sense) [NEL-LIT-micro-fuzzy-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_ajmc_fr_1 | 0.470 | 0.621 | 0.378 | 82 | 50 | 135 |
2 | team5_bundle5_ajmc_fr_2 | 0.469 | 0.617 | 0.378 | 82 | 51 | 135 |
See ranking-ajmc-fr-nel-only-micro-fuzzy-all.tsv for full details.
EL-only ajmc French relaxed @1 (literal sense) [NEL-LIT-micro-fuzzy-relaxed-TIME-ALL-LED-ALL-@1
]
Rank | System | F1 | Precision | Recall | TP | FP | FN |
---|---|---|---|---|---|---|---|
1 | team5_bundle5_ajmc_fr_1 | 0.464 | 0.614 | 0.373 | 81 | 51 | 136 |
2 | team5_bundle5_ajmc_fr_2 | 0.463 | 0.609 | 0.373 | 81 | 52 | 136 |
See ranking-ajmc-fr-nel-only-micro-fuzzy-relaxed-all.tsv for full details.
- Multilingual Newspaper Challenge (MNC)
- Multilingual Classical Commentary Challenge (MCC)
- Global Adaptation Challenge (GAC)
- GAC: Overall ranking
- GAC: Ranking overview for nel-only-relaxed
- GAC: Per dataset view for nel-only-relaxed
- GAC: Ranking overview for nel-relaxed
- GAC: Per dataset view for nel-relaxed
- GAC: Ranking overview for nerc-coarse-fuzzy
- GAC: Per dataset view for nerc-coarse-fuzzy
- GAC: Ranking overview for nerc-fine+nested-fuzzy
- GAC: Per dataset view for nerc-fine+nested-fuzzy
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
MNC | 1 | 460 | team4 |
MNC | 2 | 440 | team2 |
MNC | 3 | 330 | team5 |
MNC | 4 | 150 | team3 |
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
MCC:EL-ONLY | 1 | 330 | team5 |
MCC:EL-ONLY | 2 | 140 | team2 |
See mnc-challenge-nel-only-relaxed-challenge-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM | DATASET | LANGUAGE | F1 | System |
---|---|---|---|---|---|---|---|
MCC:EL-ONLY | 1 | 50 | team5 | hipe2020 | de | 0.525 | team5_bundle5_hipe2020_de_1 |
MCC:EL-ONLY | 2 | 40 | team2 | hipe2020 | de | 0.497 | team2_bundle5_hipe2020_de_1 |
MCC:EL-ONLY | 1 | 50 | team2 | hipe2020 | en | 0.546 | team2_bundle5_hipe2020_en_1 |
MCC:EL-ONLY | 2 | 40 | team5 | hipe2020 | en | 0.393 | team5_bundle5_hipe2020_en_2 |
MCC:EL-ONLY | 1 | 50 | team2 | hipe2020 | fr | 0.620 | team2_bundle5_hipe2020_fr_1 |
MCC:EL-ONLY | 2 | 40 | team5 | hipe2020 | fr | 0.616 | team5_bundle5_hipe2020_fr_2 |
MCC:EL-ONLY | 1 | 50 | team5 | newseye | de | 0.455 | team5_bundle5_newseye_de_1 |
MCC:EL-ONLY | 1 | 50 | team5 | newseye | fr | 0.435 | team5_bundle5_newseye_fr_1 |
MCC:EL-ONLY | 1 | 50 | team5 | sonar | de | 0.517 | team5_bundle5_sonar_de_1 |
MCC:EL-ONLY | 1 | 50 | team5 | topres19th | en | 0.654 | team5_bundle5_topres19th_en_2 |
See mnc-challenge-nel-only-relaxed-dataset-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
MNC:EL | 1 | 150 | team2 |
See mnc-challenge-nel-relaxed-challenge-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM | DATASET | LANGUAGE | F1 | System |
---|---|---|---|---|---|---|---|
MNC:EL | 1 | 50 | team2 | hipe2020 | de | 0.464 | team2_bundle1_hipe2020_de_2 |
MNC:EL | 1 | 50 | team2 | hipe2020 | en | 0.469 | team2_bundle1_hipe2020_en_2 |
MNC:EL | 1 | 50 | team2 | hipe2020 | fr | 0.578 | team2_bundle1_hipe2020_fr_1 |
See mnc-challenge-nel-relaxed-dataset-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
MNC:NERC-COARSE | 1 | 460 | team4 |
MNC:NERC-COARSE | 2 | 150 | team2 |
MNC:NERC-COARSE | 3 | 150 | team3 |
See mnc-challenge-nerc-coarse-fuzzy-challenge-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM | DATASET | LANGUAGE | F1 | System |
---|---|---|---|---|---|---|---|
MNC:NERC-COARSE | 1 | 50 | team2 | hipe2020 | de | 0.876 | team2_bundle3_hipe2020_de_1 |
MNC:NERC-COARSE | 2 | 40 | team4 | hipe2020 | de | 0.822 | team4_bundle4_hipe2020_de_1 |
MNC:NERC-COARSE | 1 | 50 | team2 | hipe2020 | en | 0.788 | team2_bundle1_hipe2020_en_1 |
MNC:NERC-COARSE | 2 | 40 | team4 | hipe2020 | en | 0.692 | team4_bundle4_hipe2020_en_1 |
MNC:NERC-COARSE | 3 | 30 | team3 | hipe2020 | en | 0.603 | team3_bundle4_hipe2020_en_1 |
MNC:NERC-COARSE | 1 | 50 | team2 | hipe2020 | fr | 0.907 | team2_bundle1_hipe2020_fr_2 |
MNC:NERC-COARSE | 2 | 40 | team3 | hipe2020 | fr | 0.808 | team3_bundle4_hipe2020_fr_1 |
MNC:NERC-COARSE | 3 | 30 | team4 | hipe2020 | fr | 0.800 | team4_bundle4_hipe2020_fr_2 |
MNC:NERC-COARSE | 1 | 50 | team4 | letemps | fr | 0.701 | team4_bundle4_letemps_fr_1 |
MNC:NERC-COARSE | 2 | 40 | team3 | letemps | fr | 0.666 | team3_bundle4_letemps_fr_1 |
MNC:NERC-COARSE | 1 | 50 | team4 | newseye | de | 0.495 | team4_bundle4_newseye_de_2 |
MNC:NERC-COARSE | 1 | 50 | team4 | newseye | fi | 0.670 | team4_bundle4_newseye_fi_1 |
MNC:NERC-COARSE | 1 | 50 | team4 | newseye | fr | 0.786 | team4_bundle4_newseye_fr_2 |
MNC:NERC-COARSE | 1 | 50 | team4 | newseye | sv | 0.746 | team4_bundle4_newseye_sv_1 |
MNC:NERC-COARSE | 1 | 50 | team4 | sonar | de | 0.695 | team4_bundle4_sonar_de_2 |
MNC:NERC-COARSE | 1 | 50 | team4 | topres19th | en | 0.838 | team4_bundle4_topres19th_en_1 |
MNC:NERC-COARSE | 2 | 40 | team3 | topres19th | en | 0.796 | team3_bundle4_topres19th_en_1 |
See mnc-challenge-nerc-coarse-fuzzy-dataset-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
MCC | 1 | 180 | team2 |
MCC | 2 | 150 | team5 |
MCC | 3 | 140 | team1 |
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
MCC:EL-ONLY | 1 | 150 | team5 |
See mcc-challenge-nel-only-relaxed-challenge-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM | DATASET | LANGUAGE | F1 | System |
---|---|---|---|---|---|---|---|
MCC:EL-ONLY | 1 | 50 | team5 | ajmc | de | 0.503 | team5_bundle5_ajmc_de_2 |
MCC:EL-ONLY | 1 | 50 | team5 | ajmc | en | 0.381 | team5_bundle5_ajmc_en_2 |
MCC:EL-ONLY | 1 | 50 | team5 | ajmc | fr | 0.464 | team5_bundle5_ajmc_fr_1 |
See mcc-challenge-nel-only-relaxed-dataset-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
MCC:EL | 1 | 50 | team2 |
See mcc-challenge-nel-relaxed-challenge-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM | DATASET | LANGUAGE | F1 | System |
---|---|---|---|---|---|---|---|
MCC:EL | 1 | 50 | team2 | ajmc | en | 0.030 | team2_bundle1_ajmc_en_2 |
See mcc-challenge-nel-relaxed-dataset-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
MCC:NERC-COARSE | 1 | 140 | team1 |
MCC:NERC-COARSE | 2 | 130 | team2 |
See mcc-challenge-nerc-coarse-fuzzy-challenge-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM | DATASET | LANGUAGE | F1 | System |
---|---|---|---|---|---|---|---|
MCC:NERC-COARSE | 1 | 50 | team2 | ajmc | de | 0.952 | team2_bundle3_ajmc_de_2 |
MCC:NERC-COARSE | 2 | 40 | team1 | ajmc | de | 0.945 | team1_bundle4_ajmc_de_2 |
MCC:NERC-COARSE | 1 | 50 | team1 | ajmc | en | 0.910 | team1_bundle4_ajmc_en_2 |
MCC:NERC-COARSE | 2 | 40 | team2 | ajmc | en | 0.894 | team2_bundle1_ajmc_en_1 |
MCC:NERC-COARSE | 1 | 50 | team1 | ajmc | fr | 0.888 | team1_bundle4_ajmc_fr_1 |
MCC:NERC-COARSE | 2 | 40 | team2 | ajmc | fr | 0.872 | team2_bundle3_ajmc_fr_2 |
See mcc-challenge-nerc-coarse-fuzzy-dataset-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
GAC | 1 | 890 | team2 |
GAC | 2 | 480 | team5 |
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
GAC:EL-ONLY | 1 | 480 | team5 |
GAC:EL-ONLY | 2 | 140 | team2 |
See gac-challenge-nel-only-relaxed-challenge-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM | DATASET | LANGUAGE | F1 | System |
---|---|---|---|---|---|---|---|
GAC:EL-ONLY | 1 | 50 | team5 | ajmc | de | 0.503 | team5_bundle5_ajmc_de_2 |
GAC:EL-ONLY | 1 | 50 | team5 | ajmc | en | 0.381 | team5_bundle5_ajmc_en_2 |
GAC:EL-ONLY | 1 | 50 | team5 | ajmc | fr | 0.464 | team5_bundle5_ajmc_fr_1 |
GAC:EL-ONLY | 1 | 50 | team5 | hipe2020 | de | 0.525 | team5_bundle5_hipe2020_de_1 |
GAC:EL-ONLY | 2 | 40 | team2 | hipe2020 | de | 0.497 | team2_bundle5_hipe2020_de_1 |
GAC:EL-ONLY | 1 | 50 | team2 | hipe2020 | en | 0.546 | team2_bundle5_hipe2020_en_1 |
GAC:EL-ONLY | 2 | 40 | team5 | hipe2020 | en | 0.393 | team5_bundle5_hipe2020_en_2 |
GAC:EL-ONLY | 1 | 50 | team2 | hipe2020 | fr | 0.620 | team2_bundle5_hipe2020_fr_1 |
GAC:EL-ONLY | 2 | 40 | team5 | hipe2020 | fr | 0.616 | team5_bundle5_hipe2020_fr_2 |
GAC:EL-ONLY | 1 | 50 | team5 | newseye | de | 0.455 | team5_bundle5_newseye_de_1 |
GAC:EL-ONLY | 1 | 50 | team5 | newseye | fr | 0.435 | team5_bundle5_newseye_fr_1 |
GAC:EL-ONLY | 1 | 50 | team5 | sonar | de | 0.517 | team5_bundle5_sonar_de_1 |
GAC:EL-ONLY | 1 | 50 | team5 | topres19th | en | 0.654 | team5_bundle5_topres19th_en_2 |
See gac-challenge-nel-only-relaxed-dataset-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
GAC:EL | 1 | 200 | team2 |
See gac-challenge-nel-relaxed-challenge-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM | DATASET | LANGUAGE | F1 | System |
---|---|---|---|---|---|---|---|
GAC:EL | 1 | 50 | team2 | ajmc | en | 0.030 | team2_bundle1_ajmc_en_2 |
GAC:EL | 1 | 50 | team2 | hipe2020 | de | 0.464 | team2_bundle1_hipe2020_de_2 |
GAC:EL | 1 | 50 | team2 | hipe2020 | en | 0.469 | team2_bundle1_hipe2020_en_2 |
GAC:EL | 1 | 50 | team2 | hipe2020 | fr | 0.578 | team2_bundle1_hipe2020_fr_1 |
See gac-challenge-nel-relaxed-dataset-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
GAC:NERC-COARSE | 1 | 300 | team2 |
See gac-challenge-nerc-coarse-fuzzy-challenge-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM | DATASET | LANGUAGE | F1 | System |
---|---|---|---|---|---|---|---|
GAC:NERC-COARSE | 1 | 50 | team2 | ajmc | de | 0.952 | team2_bundle3_ajmc_de_2 |
GAC:NERC-COARSE | 1 | 50 | team2 | ajmc | en | 0.894 | team2_bundle1_ajmc_en_1 |
GAC:NERC-COARSE | 1 | 50 | team2 | ajmc | fr | 0.872 | team2_bundle3_ajmc_fr_2 |
GAC:NERC-COARSE | 1 | 50 | team2 | hipe2020 | de | 0.876 | team2_bundle3_hipe2020_de_1 |
GAC:NERC-COARSE | 1 | 50 | team2 | hipe2020 | en | 0.788 | team2_bundle1_hipe2020_en_1 |
GAC:NERC-COARSE | 1 | 50 | team2 | hipe2020 | fr | 0.907 | team2_bundle1_hipe2020_fr_2 |
See gac-challenge-nerc-coarse-fuzzy-dataset-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM |
---|---|---|---|
GAC:NERC-FINE+NESTED | 1 | 250 | team2 |
See gac-challenge-nerc-fine+nested-fuzzy-challenge-team-ranking.tsv for full details.
CHALLENGE | RANK | POINTS | TEAM | DATASET | LANGUAGE | F1 | System |
---|---|---|---|---|---|---|---|
GAC:NERC-FINE+NESTED | 1 | 50 | team2 | ajmc | de | 0.933 | team2_bundle3_ajmc_de_2 |
GAC:NERC-FINE+NESTED | 1 | 50 | team2 | ajmc | en | 0.847 | team2_bundle1_ajmc_en_1 |
GAC:NERC-FINE+NESTED | 1 | 50 | team2 | ajmc | fr | 0.728 | team2_bundle3_ajmc_fr_2 |
GAC:NERC-FINE+NESTED | 1 | 50 | team2 | hipe2020 | de | 0.673 | team2_bundle3_hipe2020_de_1 |
GAC:NERC-FINE+NESTED | 1 | 50 | team2 | hipe2020 | fr | 0.614 | team2_bundle1_hipe2020_fr_2 |
See gac-challenge-nerc-fine+nested-fuzzy-dataset-team-ranking.tsv for full details.