Home > Research > Publications & Outputs > A Comparative Study of Evaluation Metrics for L...

Electronic data

Links

Text available via DOI:

View graph of relations

A Comparative Study of Evaluation Metrics for Long-Document Financial Narrative Summarization with Transformers

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
Publication date21/06/2023
Host publicationNatural Language Processing and Information Systems - 28th International Conference on Applications of Natural Language to Information Systems, NLDB 2023, Proceedings
EditorsElisabeth Métais, Farid Meziane, Warren Manning, Stephan Reiff-Marganiec, Vijayan Sugumaran
Place of PublicationCham
PublisherSpringer
Pages391-403
Number of pages13
ISBN (print)9783031353192
<mark>Original language</mark>English
Event28th International Conference on Applications of Natural Language to Information Systems, NLDB 2023 - Derby, United Kingdom
Duration: 21/06/202323/06/2023

Conference

Conference28th International Conference on Applications of Natural Language to Information Systems, NLDB 2023
Country/TerritoryUnited Kingdom
CityDerby
Period21/06/2323/06/23

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume13913 LNCS
ISSN (Print)0302-9743
ISSN (electronic)1611-3349

Conference

Conference28th International Conference on Applications of Natural Language to Information Systems, NLDB 2023
Country/TerritoryUnited Kingdom
CityDerby
Period21/06/2323/06/23

Abstract

There are more than 2,000 listed companies on the UK’s London Stock Exchange, divided into 11 sectors who are required to communicate their financial results at least twice in a single financial year. UK annual reports are very lengthy documents with around 80 pages on average. In this study, we aim to benchmark a variety of summarisation methods on a set of different pre-trained transformers with different extraction techniques. In addition, we considered multiple evaluation metrics in order to investigate their differing behaviour and applicability on a dataset from the Financial Narrative Summarisation (FNS 2020) shared task, which is composed of annual reports published by firms listed on the London Stock Exchange and their corresponding summaries. We hypothesise that some evaluation metrics do not reflect true summarisation ability and propose a novel BRUGEscore metric, as the harmonic mean of ROUGE-2 and BERTscore. Finally, we perform a statistical significance test on our results to verify whether they are statistically robust, alongside an adversarial analysis task with three different corruption methods.