Skip to content

Ravoxsg/awesome-two-stage-summarization

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

48 Commits
 
 

Repository files navigation

Two-stage abstractive summarization

Standard abstractive summarization

Traditionally, abstractive summarization models are fine-tuned in a single-stage. For instance, one can use a BART, T5 or PEGASUS pre-trained checkpoint, then fine-tune it on the desired dataset (e.g, CNN/DailyMail). Fine-tuning is done by Maximum Likelihood Estimation (MLE) and negative log-likelihood (NLL) loss, maximizing the probability that the model assigns to the (unique) ground-truth summary paired with the source document. During training, teacher forcing is used, while during inference, auto-regressive decoding is used.

Why "two-stage" abstractive summarization

The above process is not ideal for two main reasons:

  1. The model only optimizes for a single summary, while in practice the search space is huge and there exists plenty of high-quality summaries.
  2. There is a large discrepancy between teacher forcing during training, and auto-regressive decoding during inference, known as the exposure bias. At inference, the model has no mechanism to realize it's decoding in a wrong direction, and cannot change course. Besides, it is never trained to build upon its previous predictions.

Due to these limitations, since around 2021, researchers have started training abstractive summarization in two stages. While the 1st stage is the same (MLE with teacher forcing), the 2nd stage training process differs. The overall goal of the 2nd-stage is to calibrate the generation, so that the model assigns higher probability to summary candidates which actually have a higher performance (e.g., ROUGE or BERTScore with regards to the target). A lot of 2nd stage models operate at the sequence level (in contrast to the 1st stage where the loss is at the token level).

Two-stage abstractive summarization techniques

We attempt a broad categorization of two-stage abstractive summarization training techniques:

  • Reinforcement learning: GOLD [5]
  • Using guidance from another model to improve the current model: GSum [6]
  • Meta-learning to learn from different systems: RefSum [7]
  • Contrastive learning (most popular type of approach):
    • Contrastive loss: SeqCo [10]
    • Ranking loss: ConSum [8], SimCLS [9], BRIO [12], SLiC [13], BalSum [15]
    • Binary cross-entropy loss: SummaReranker [11]
    • Expected reward: SLiC [13]
    • Mix of ranking loss + coarse-grained contrastive learning: SimMCS [16]
  • Fusion of several summary candidates: SummaFusion [14]
  • Linear regression with summary candidates features: SummScore [17]

We highlight that some of these models just train the 2nd stage model, which cannot generate summaries by itself and must also rely on 1st stage model summary generation at inference.
This is the case of: SimCLS [9], SummaReranker [11], BRIO-ctr [12], BalSum [15], SummScore [17]
While other approaches perform multi-task learning, where the model is still fine-tuned with the NLL loss.
This is the case of: ConSum [8], SeqCo [10], BRIO-mul [12], SLiC [13], SummaFusion [14], SimMCS [16]

Results on CNN-DailyMail

We first show results with baseline (1st stage) models:

Model R-1 R-2 R-L
T5 (large) [1] 42.50 20.68 39.75
BART (large) [2] 44.16 21.28 40.90
PEGASUS (large) [3] 44.17 21.47 41.11
ProphetNet [4] 43.68 20.64 40.72

And now results with 2-stage models:

Model Base Model R-1 R-2 R-L
GOLD [5] BART 45.40 22.01 42.25
GSum [6] BART 45.94 22.32 42.48
RefSum [7] _ 46.18 22.36 42.91
ConSum [8] PEGASUS 44.53 21.54 41.57
SimCLS [9] BART 46.67 22.15 43.54
SeqCo [10] BART 45.02 21.80 41.75
SummaReranker [11] PEGASUS 47.16 22.55 43.87
BRIO [12] BART 47.78 23.55 44.57
SLiC [13] PEGASUS 47.97 24.18 44.88
BalSum [15] BART 46.58 22.33 43.49
SimMCS [16] BART 48.38 24.17 44.79

We see that 2nd-stage methods have moved the SOTA from ~44 R-1 to ~48 R-1 in a couple years.

References

  1. Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, Peter J. Liu. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer (T5). JMLR 2020.
  2. Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov, Luke Zettlemoyer. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. ACL 2020.
  3. Jingqing Zhang, Yao Zhao, Mohammad Saleh, Peter J. Liu. PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization. ICML 2020.
  4. Weizhen Qi, Yu Yan, Yeyun Gong, Dayiheng Liu, Nan Duan, Jiusheng Chen, Ruofei Zhang, Ming Zhou. ProphetNet: Predicting Future N-gram for Sequence-to-Sequence Pre-training. EMNLP Findings 2020.
  5. Richard Yuanzhe Pang, He He. Text Generation by Learning from Demonstrations (GOLD). ICLR 2021.
  6. Zi-Yi Dou, Pengfei Liu, Hiroaki Hayashi, Zhengbao Jiang, Graham Neubig. GSum: A General Framework for Guided Neural Abstractive Summarization. NAACL 2021.
  7. Yixin Liu, Zi-Yi Dou, Pengfei Liu. RefSum: Refactoring Neural Summarization. NAACL 2021.
  8. Shichao Sun, Wenjie Li. Alleviating Exposure Bias via Contrastive Learning for Abstractive Text Summarization (ConSum).
  9. Yixin Liu, Pengfei Liu. SimCLS: A Simple Framework for Contrastive Learning of Abstractive Summarization. ACL 2021.
  10. Shusheng Xu, Xingxing Zhang, Yi Wu, Furu Wei. Sequence Level Contrastive Learning for Text Summarization (SeqCo). AAAI 2022.
  11. Mathieu Ravaut, Shafiq Joty, Nancy F. Chen. SummaReranker: A Multi-Task Mixture-of-Experts Re-ranking Framework for Abstractive Summarization. ACL 2022.
  12. Yixin Liu, Pengfei Liu, Dragomir Radev, Graham Neubig. BRIO: Bringing Order to Abstractive Summarization. ACL 2022.
  13. Yao Zhao, Misha Khalman, Rishabh Joshi, Shashi Narayan, Mohammad Saleh, Peter J Liu. Calibrating Sequence Likelihood Improves Conditional Language Generation (SLiC). ICLR 2023.
  14. Mathieu Ravaut, Shafiq Joty, Nancy F. Chen. Towards Summary Candidates Fusion (SummaFusion). EMNLP 2022.
  15. Jeewoo Sul and Yong Suk Choi. Balancing Lexical and Semantic Quality in Abstractive Summarization (BalSum). ACL 2023.
  16. Jiawen Xie, Qi Su, Shaoting Zhang, Xiaofan Zhang. Alleviating Exposure Bias via Multi-level Contrastive Learning and Deviation Simulation in Abstractive Summarization (SimMCS). ACL 2023.
  17. Mathieu Ravaut, Shafiq Joty, Nancy Chen. Unsupervised Summarization Re-ranking (SummScore). ACL Findings 2023.

About

A curated list of papers about two-stage summarization.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published