📄

PEGASUS

Official

GitHub - google-research/pegasus
Pre-training with Extracted Gap-sentences for Abstractive SUmmarization Sequence-to-sequence models, or PEGASUS, uses self-supervised objective Gap Sentences Generation (GSG) to train a transformer encoder-decoder model. The paper can be found on arXiv. ICML 2020 accepted.
https://github.com/google-research/pegasus
PEGASUS: A State-of-the-Art Model for Abstractive Text Summarization
Students are often tasked with reading a document and producing a summary (for example, a book report) to demonstrate both reading comprehension and writing ability. This abstractive text summarization is one of the most challenging tasks in natural language processing, involving understanding of long passages, information compression, and language generation.
https://ai.googleblog.com/2020/06/pegasus-state-of-art-model-for.html

Summary

They propose pre-training large Transformer-based encoder-decoder models. The important sentences are removed/masked from an input document and are generated together as one output sequence from the remaining sentences (similar to extractive summary). Recent work leveraging pre-trained Transformer-based sequence-to-sequence models has extended the success in NLP tasks to text generation, including abstractive summarization.

Architecture

Pre-Training Task

Masking whole sentences from a document and generating these gap-sentences from the rest of the document works well as a pre-training objective for downstream summarization tasks. Choosing putatively important sentences outperforms lead or randomly selected ones. This self-supervised objective is called Gap Sentences Generation (GSG).

We select and mask whole sentences from documents, and concatenate the gap-sentences into a pseudo-summary. The corresponding position of each selected gap sentence is replaced by a mask token to inform the model. Gap Sentences Ratio (GSR), refers to the number of selected gap sentences to the total number of sentences in the document.

We select sentences that appear to be important to the document, based on 3 primary strategies:

We apply Masked Language Model (MLM) to train the Transformer encoder as the sole pre-training objective or along with GSG. However, MLM does not improve downstream tasks at large number of pre-training steps.

We consider two large text corpora:

Downstream Tasks

We use public abstractive summarization datasets accessible through:

TensorFlow Datasets
The datasets documented here are from Note: HEAD and so not all are available in the current tensorflow-datasets package. They are all accessible in our nightly package tfds-nightly. See our getting-started guide for a quick introduction. for ex in tfds.load('cifar10', split='train'): ...
https://www.tensorflow.org/datasets/catalog/overview

Experiments

There are several paragraph-ideal-model examples appended to the paper.

Performance

Further Readings

How to Perform Abstractive Summarization with PEGASUS
Through the article, we hope that you can adapt and adopt PEGASUS for your abstractive summarization tasks. To put it analogously, we no longer need to work like a horse and can instead ride on PEGASUS to achieve great heights and results 😉.
https://towardsdatascience.com/how-to-perform-abstractive-summarization-with-pegasus-3dd74e48bafb
PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization
Understanding current SOTA Summarization System Text summarization is the task of shortening a given text yet retaining the essential gist of the entire document. Summarization is usually seen to be broadly divided into two categories Extractive and Abstractive. Extractive summarization can be...
https://medium.com/analytics-vidhya/pegasus-pre-training-with-extracted-gap-sentences-for-abstractive-summarization-acb238aa1096
Summarization Using Pegasus Model with the Transformers Library
PEGASUS is the latest state-of-the-art model for abstractive summarization open-sourced by Google, recently in June 2020. It stands for Pre-training with Extracted Gap-sentences for Abstractive SUmmarization Sequence-to-sequence models. For more details on the model, refer to the paper on arXiv .
https://pub.towardsai.net/summarization-using-pegasus-model-with-the-transformers-library-553cd0dc5c2
Abstractive Summarization Using Pegasus - Turbolab Technologies
In the last article, we have seen how to perform extractive summarization of some text, which selects important sentences and gives them out by ranking them, without changing any text. While they are suitable for some cases, they do not achieve the sophistication of human-like summaries.
https://turbolab.in/abstractive-summarization-using-pegasus/
PEGASUS: Google's State of the Art Abstractive Summarization Model
The ability to summarize evaluates one's understanding of a given piece of text or a language. Perhaps the best test of a man's intelligence is his capacity for making a summary - Lytton Strachey Hence, summarization is a fairly significant concept in NLP.
https://towardsdatascience.com/pegasus-google-state-of-the-art-abstractive-summarization-model-627b1bbbc5ce?gi=00cf4c1329a3