Abstract
Recently, contrastive learning has been extended from visual representation to summarization tasks. Abstractive summarization aims to generate a short description for a document while retaining significant information. At present, the methods of contrastive learning summarization focus on modeling the global semantics of source documents, targets and candidate summaries to maximize their similarities. However, they ignore the influence of sentence semantics in the source document. In this paper, we propose a sentence-level semantic similarity distance-aware contrastive learning method (SSDCL), which integrates the semantic similarity distance between summaries and sentences of source documents into the contrastive loss in the form of soft weights. Therefore, our model maximize the similarity between summaries and salient information, while minimizing the similarity between summaries and noise. We conducted extensive experiments on CNN/Daily Mail and XSum datasets to verify our model. The experimental results show that the proposed method achieved remarkable performance over the baseline and many advanced methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Chen, S., Zhou, J., Sun, Y., et al.: An information minimization based contrastive learning model for unsupervised sentence embeddings learning. In: COLING, pp. 4821–4831 (2022)
Chen, X., Xie, S., He, K.: An empirical study of training self-supervised vision transformers. In: ICCV, pp. 9640–9649 (2021)
Chopra, S., Hadsell, R., LeCun, Y.: Learning a similarity metric discriminatively, with application to face verification. In: CVPR, pp. 539–546 (2005)
Choubey, P.K., Fabbri, A., Vig, J., et al.: CaPE: contrastive parameter ensembling for reducing hallucination in abstractive summarization. In: ACL, pp. 10755–10773 (2023)
Feng, J., Long, J., Han, C., et al.: RepSum: a general abstractive summarization framework with dynamic word embedding representation correction. Comput. Speech Lang. 80, 101491 (2023)
Hermann, K.M., Kočiský, T., Grefenstette, E., et al.: Teaching machines to read and comprehend. In: NIPS, pp. 1693–1701 (2015)
Hou, C., Li, Z., Tang, Z., et al.: Multiple instance relation graph reasoning for cross-modal hash retrieval. Knowl.-Based Syst. 256, 109891 (2022)
Hsu, W.T., Lin, C.K., Lee, M.Y., et al.: A unified model for extractive and abstractive summarization using inconsistency loss. In: ACL, pp. 132–141 (2018)
Lebanoff, L., Song, K., Dernoncourt, F., et al.: Scoring sentence singletons and pairs for abstractive summarization. In: ACL, pp. 2175–2189 (2019)
Lewis, M., Liu, Y., Goyal, N., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: ACL, pp. 7871–7880 (2020)
Li, Z., Peng, Z., Tang, S., et al.: Text summarization method based on double attention pointer network. IEEE Access 8, 11279–11288 (2020)
Li, Z., Sun, Y., Zhu, J., et al.: Improve relation extraction with dual attention-guided graph convolutional networks. Neural Comput. Appl. 33, 1773–1784 (2021)
Lin, C.Y.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74–81 (2004)
Liu, Y., Zhu, C., Zeng, M.: End-to-end segmentation-based news summarization. In: ACL, pp. 544–554 (2022)
Liu, Y., Liu, P.: SimCLS: a simple framework for contrastive learning of abstractive summarization. In: ACL, pp. 1065–1072 (2021)
Liu, Y., Liu, P., Radev, D., et al.: BRIO: bringing order to abstractive summarization. In: ACL, pp. 2890–2903 (2022)
Liu, Y., Jia, Q., Zhu, K.: Length control in abstractive summarization by pretraining information selection. In: ACL, pp. 6885–6895 (2022)
Nallapati, R., Zhou, B., dos Santos, C., et al.: Abstractive text summarization using sequence-to-sequence RNNs and Beyond. In: CoNLL, pp. 280–290 (2016)
Narayan, S., Cohen, S.B., Lapata, M.: Don’t give me the details, Just the Summary! Topic-Aware convolutional neural networks for extreme summarization. In: EMNLP, pp. 1797–1807 (2018)
Pang, R.Y., He, H.: Text generation by learning from demonstrations. In: ICLR, pp. 1–22 (2021)
Rush, A.M., Chopra, S., Weston, J.: A neural attention model for abstractive sentence summarization. In: EMNLP, pp. 379–389 (2015)
See, A., Liu, P.J., Manning, C.D.: Get to the point: summarization with pointer-generator networks. In: ACL, pp. 1073–1083 (2017)
Shazeer, N., Stern, M.: Adafactor: adaptive learning rates with sublinear memory cost. In: ICML, pp. 4596–4604 (2018)
Sun, S., Li, W.: Alleviating exposure bias via contrastive learning for abstractive text summarization. arXiv preprint arXiv:2108.11846, pp. 1–6 (2021)
Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: NIPS, pp. 3104–3112 (2014)
Vaswani, A., Shazeer, N., Parmar, N., et al.: Attention is all you need. In: NIPS, pp. 6000–6010 (2017)
Wang, D., Chen, J., Zhou, H., et al.: Contrastive aligned joint learning for multilingual summarization. In: ACL, pp. 2739–2750 (2021)
Wang, F., Song, K., Zhang, H., et al.: Salience allocation as guidance for abstractive summarization. In: EMNLP, pp. 6094–6106 (2022)
Xian, T., Li, Z., Zhang, C., et al.: Dual global enhanced transformer for image captioning. Neural Netw. 148, 129–141 (2022)
Xie, X., Li, Z., Tang, Z., et al.: Unifying knowledge iterative dissemination and relational reconstruction network for image-text matching. Inf. Process. Manage. 60(1), 103154 (2023)
Xu, S., Zhang, X., Wu, Y., et al.: Sequence level contrastive learning for text summarization. In: AAAI, pp. 11556–11565 (2022)
Xu, S., Li, H., Yuan, P., et al.: Self-attention guided copy mechanism for abstractive summarization. In: ACL, pp. 1355–1362 (2020)
Zaheer, M., Guruganesh, G., Dubey, K.A., et al.: Big Bird: transformers for longer sequences. In: NIPS, pp. 17283–17297 (2020)
Zhang, J., Zhao, Y., Saleh, M., et al.: PEGASUS: pre-training with extracted gap-sentences for abstractive summarization. In: ICML, pp. 11328–11339 (2020)
Zheng, C., Zhang, K., Wang, H.J., et al.: Enhanced Seq2Seq autoencoder via contrastive learning for abstractive text summarization. In: IEEE Big Data, pp. 1764–1771 (2021)
Zhong, M., Liu, P., Chen, Y., et al.: Extractive summarization as text matching. In: ACL, pp. 6197–6208 (2020)
Acknowledgements
This work is supported by National Natural Science Foundation of China (Nos. 62276073, 61966004), Guangxi Natural Science Foundation (No. 2019GXNSFDA245018), Innovation Project of Guangxi Graduate Education (YCSW2023141), Guangxi “Bagui Scholar” Teams for Innovation and Research Project, and Guangxi Collaborative Innovation Center of Multi-source Information Integration and Intelligent Processing.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Huang, Y., Li, Z. (2024). A Semantic Similarity Distance-Aware Contrastive Learning for Abstractive Summarization. In: Liu, F., Sadanandan, A.A., Pham, D.N., Mursanto, P., Lukose, D. (eds) PRICAI 2023: Trends in Artificial Intelligence. PRICAI 2023. Lecture Notes in Computer Science(), vol 14325. Springer, Singapore. https://doi.org/10.1007/978-981-99-7019-3_18
Download citation
DOI: https://doi.org/10.1007/978-981-99-7019-3_18
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-7018-6
Online ISBN: 978-981-99-7019-3
eBook Packages: Computer ScienceComputer Science (R0)