voxel a benchmark dataset for
play

VoxEL: A Benchmark Dataset for Multilingual Entity Linking Henry - PowerPoint PPT Presentation

VoxEL: A Benchmark Dataset for Multilingual Entity Linking Henry Rosales-M endez, Aidan Hogan and Barbara Poblete University of Chile { hrosales,ahogan,bpoblete } @dcc.uchile.cl October 10, 2018 ISWC 2018 - The 17th Internationl


  1. VoxEL: A Benchmark Dataset for Multilingual Entity Linking † Henry Rosales-M´ endez, Aidan Hogan and Barbara Poblete University of Chile { hrosales,ahogan,bpoblete } @dcc.uchile.cl October 10, 2018 † ISWC 2018 - The 17th Internationl Semantic Web Conference

  2. Example

  3. Example - Entity Recognition

  4. Example - Entity Disambiguation

  5. Applications • Semantic Search • Semantic Annotations • Relation Extraction • Topic Extraction

  6. Name Variations in Entity Linking Michael Joseph Jackson Michael J. Jackson King of Pop

  7. Name Variations in Entity Linking Michael Jackson

  8. Multilingual Entity Linking - English

  9. Multilingual Entity Linking - Italian

  10. Multilingual Entity Linking - Spanish

  11. Datasets

  12. Datasets

  13. Goals 1 Create a benchmark dataset for multilingual Entity Linking

  14. Curated source: VoxEurop

  15. Example - Any other entity?

  16. Example - Any other entity?

  17. Example - Any other entity?

  18. Example - Any other entity?

  19. Example annotations produced by four EL systems

  20. Example annotations produced by four EL systems

  21. Example annotations produced by four EL systems Aida

  22. Example annotations produced by four EL systems Babelfy Aida

  23. Example annotations produced by four EL systems Babelfy Aida DBpedia Spotlight

  24. Example annotations produced by four EL systems Babelfy Aida T agME DBpedia Spotlight

  25. • What should Entity Linking link?

  26. Datasets

  27. Datasets

  28. Goals 1 Create a benchmark dataset for multilingual Entity Linking

  29. Goals 1 Create a benchmark dataset for multilingual Entity Linking 2 Create two versions of the dataset: strict and relaxed.

  30. Strict verison: class-based definition

  31. Strict verison: class-based definition

  32. Strict verison: class-based definition ?

  33. Relaxed version: Knowledge Base definition

  34. Creation of VoxEL dataset • It is based on curated text from five languages. • Same sentences by each corresponding document. • Same annotations by each corresponding sentence. • Revision process.

  35. Summary

  36. Summary

  37. Summary

  38. Experiments 1 GERBIL Evaluation of state-of-the-art approaches

  39. Experiments

  40. Experiments

  41. Experiments DE EN Avg. of Micro F1 ES .66 .65 FR .61 .60 .59 .60 .58 .58 .57 IT .50.47 .40 .39 .34 .34.33.32 .34 .34 .30 .28 .27 .27 .22 .19 Babefy TAGME THD Babefy DB-sp FREME r s (a) Results of the Relaxed version of VoxEL .86 .81 .81 .78 .76 .75 .74 .74 .72 .71 .72.70.71 Avg. of Micro F1 .70 .71 .65 .64 .64 .60 .60 .54 .53 .49.51 .50 Babefy TAGME THD Babefy DB-sp FREME r s (b) Results of the Strict version of VoxEL

  42. Experiments DE EN Avg. of Micro F1 ES .66 .65 FR .61 .60 .59 .60 .58 .58 .57 IT .50.47 .40 .39 .34 .34.33.32 .34 .34 .30 .28 .27 .27 .22 .19 Babefy TAGME THD Babefy DB-sp FREME r s (a) Results of the Relaxed version of VoxEL .86 .81 .81 .78 .76 .75 .74 .74 .72 .71 .72.70.71 Avg. of Micro F1 .70 .71 .65 .64 .64 .60 .60 .54 .53 .49.51 .50 Babefy TAGME THD Babefy DB-sp FREME r s (b) Results of the Strict version of VoxEL

  43. Experiments DE EN Avg. of Micro F1 ES .66 .65 FR .61 .60 .59 .60 .58 .58 .57 IT .50.47 .40 .39 .34 .34.33.32 .34 .34 .30 .28 .27 .27 .22 .19 Babefy TAGME THD Babefy DB-sp FREME r s (a) Results of the Relaxed version of VoxEL .86 .81 .81 .78 .76 .75 .74 .74 .72 .71 .72.70.71 Avg. of Micro F1 .70 .71 .65 .64 .64 .60 .60 .54 .53 .49.51 .50 Babefy TAGME THD Babefy DB-sp FREME r s (b) Results of the Strict version of VoxEL

  44. Experiments DE EN Avg. of Micro F1 ES .66 .65 FR .61 .60 .59 .60 .58 .58 .57 IT .50.47 .40 .39 .34 .34.33.32 .34 .34 .30 .28 .27 .27 .22 .19 Babefy TAGME THD Babefy DB-sp FREME r s (a) Results of the Relaxed version of VoxEL .86 .81 .81 .78 .76 .75 .74 .74 .72 .71 .72.70.71 Avg. of Micro F1 .70 .71 .65 .64 .64 .60 .60 .54 .53 .49.51 .50 Babefy TAGME THD Babefy DB-sp FREME r s (b) Results of the Strict version of VoxEL

  45. Experiments 1 GERBIL Evaluation of state-of-the-art approaches

  46. Experiments 1 GERBIL Evaluation of state-of-the-art approaches 2 Evaluate the performance of state-of-the-art approaches using machine translation.

  47. Experiments Input T ext EN ES FR IT DE FR DE EN ES IT System Configuration

  48. Experiments Input T ext EN ES FR IT DE FR DE EN ES IT System Configuration

  49. Experiments Input T ext EN ES FR IT DE FR DE EN ES IT System Configuration

  50. Experiments Input T ext EN ES FR IT DE FR DE EN ES IT System Configuration

  51. Experiments Calibrated Translation Avg. of Micro F1 English .55 .52 .51 .47 .46 .45 .43 .40 .41 .41 .39 .39 .33 .31 .31 .31 .30 .24 Babefy DB-sp FREME TAGME THD Babefy r s (a) Results of the Relaxed version of VoxEL Avg. of Micro F1 .71 .71 .70 .71 .63 .60 .60 .57 .57 .55 .53 .53 .35 .33 .33 .32 .30 .27 Babefy DB-sp FREME TAGME THD Babefy r s (b) Results of the Strict version of VoxEL

  52. Conclusion Our main contribution is VoxEL (https://dx.doi.org/10.6084/m9.figshare.6539675) • Most systems perform (much) better for English. • Machine Translation could be an option to address multilingual domains in Entity Linking.

  53. Poster P20: Machine Translation vs. Multilingual Approaches for Entity Linking EN IT ES

  54. VoxEL: A Benchmark Dataset for Multilingual Entity Linking † Henry Rosales-M´ endez, Aidan Hogan and Barbara Poblete University of Chile { hrosales,ahogan,bpoblete } @dcc.uchile.cl October 10, 2018 † ISWC 2018 - The 17th Internationl Semantic Web Conference

Recommend


More recommend