Artikel in einem Konferenzbericht,

Large-Scale Evaluation of Keyphrase Extraction Models

, , und .
Proceedings of the ACM/IEEE Joint Conference on Digital Libraries in 2020, Seite 271-278. ACM, (August 2020)
DOI: 10.1145/3383583.3398517

Zusammenfassung

Keyphrase extraction models are usually evaluated under different, not directly comparable, experimental setups. As a result, it remains unclear how well proposed models actually perform, and how they compare to each other. In this work, we address this issue by presenting a systematic large-scale analysis of state-of-the-art keyphrase extraction models involving multiple benchmark datasets from various sources and domains. Our main results reveal that state-of-the-art models are in fact still challenged by simple baselines on some datasets. We also present new insights about the impact of using author- or reader-assigned keyphrases as a proxy for gold standard, and give recommendations for strong baselines and reliable benchmark datasets.

Tags

Nutzer

  • @brusilovsky
  • @dblp

Kommentare und Rezensionen