TY - GEN
T1 - SUMMN
T2 - 60th Annual Meeting of the Association for Computational Linguistics, ACL 2022
AU - Zhang, Yusen
AU - Ni, Ansong
AU - Mao, Ziming
AU - Wu, Chen Henry
AU - Zhu, Chenguang
AU - Deb, Budhaditya
AU - Awadallah, Ahmed H.
AU - Radev, Dragomir
AU - Zhang, Rui
N1 - Publisher Copyright:
© 2022 Association for Computational Linguistics.
PY - 2022
Y1 - 2022
N2 - Text summarization helps readers capture salient information from documents, news, interviews, and meetings. However, most state-of-the-art pretrained language models (LM) are unable to efficiently process long text for many summarization tasks. In this paper, we propose SUMMN, a simple, flexible, and effective multi-stage framework for input texts that are longer than the maximum context length of typical pretrained LMs. SUMMN first splits the data samples and generates a coarse summary in multiple stages and then produces the final fine-grained summary based on it. Our framework can process input text of arbitrary length by adjusting the number of stages, while keeping the LM input size fixed. Moreover, it can deal with both single-source documents and dialogues, and it can be used on top of different backbone abstractive summarization models. To the best of our knowledge, SUMMN is the first multi-stage split-then-summarize framework for long input summarization. Our experiments demonstrate that SUMMN outperforms previous state-of-the-art methods by improving ROUGE scores on three long meeting summarization datasets AMI, ICSI, and QMSum, two long TV series datasets from SummScreen, and a long document summarization dataset GovReport. Our data and code are available at https://github.com/psunlpgroup/Summ-N.
AB - Text summarization helps readers capture salient information from documents, news, interviews, and meetings. However, most state-of-the-art pretrained language models (LM) are unable to efficiently process long text for many summarization tasks. In this paper, we propose SUMMN, a simple, flexible, and effective multi-stage framework for input texts that are longer than the maximum context length of typical pretrained LMs. SUMMN first splits the data samples and generates a coarse summary in multiple stages and then produces the final fine-grained summary based on it. Our framework can process input text of arbitrary length by adjusting the number of stages, while keeping the LM input size fixed. Moreover, it can deal with both single-source documents and dialogues, and it can be used on top of different backbone abstractive summarization models. To the best of our knowledge, SUMMN is the first multi-stage split-then-summarize framework for long input summarization. Our experiments demonstrate that SUMMN outperforms previous state-of-the-art methods by improving ROUGE scores on three long meeting summarization datasets AMI, ICSI, and QMSum, two long TV series datasets from SummScreen, and a long document summarization dataset GovReport. Our data and code are available at https://github.com/psunlpgroup/Summ-N.
UR - http://www.scopus.com/inward/record.url?scp=85139155737&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85139155737&partnerID=8YFLogxK
U2 - 10.18653/v1/2022.acl-long.112
DO - 10.18653/v1/2022.acl-long.112
M3 - Conference contribution
AN - SCOPUS:85139155737
T3 - Proceedings of the Annual Meeting of the Association for Computational Linguistics
SP - 1592
EP - 1604
BT - ACL 2022 - 60th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers)
A2 - Muresan, Smaranda
A2 - Nakov, Preslav
A2 - Villavicencio, Aline
PB - Association for Computational Linguistics (ACL)
Y2 - 22 May 2022 through 27 May 2022
ER -