Abstract
Legal Judgment Summarization (LJS) can highly summarize legal judgment documents, improving judicial work efficiency in case retrieval and other occasions. Legal judgment documents are usually lengthy; however, most existing LJS methods are directly based on general text summarization models, which cannot handle long texts effectively. Additionally, due to the complex structural characteristics of legal judgment documents, some information may be lost by applying only one single kind of summarization model. To address these issues, we propose an integrated summarization method which leverages both semantic and structural information to improve the quality of LJS. Specifically, legal judgment documents are firstly segmented into three relatively short parts according to their specific structure. We propose an extractive summarization model named BSLT and an abstractive summarization model named LPGN by adopting Lawformer as the encoder. Lawformer is a new pre-trained language model for long legal documents, which specializes in capturing long-distance dependency and modeling legal semantic features. Then, we adopt different models to summarize the corresponding part regarding its structural characteristics. Finally, the obtained summaries are integrated to generate a high-quality summary involving semantic and structural information. We conduct comparative experiments to evaluate the performance of our model. The results show that our model outperforms the baseline model LEAD-3 by 14.78% on the mean ROUGE score, which demonstrates our method is effective in LJS and is prospected to be applied to assist other tasks in legal artificial intelligence.
Similar content being viewed by others
References
Anand D, Wagh R (2022) Effective deep learning approaches for summarization of legal texts. J King Saud Univ Comput Inf Sci 34(5):2141–2150
Bajaj A, Dangati P, Krishna K, et al (2021) Long document summarization in a low resource setting using pretrained language models. In: Proceedings of the ACL-IJCNLP 2021 student research workshop, pp 71–80
Bhattacharya P, Poddar S, Rudra K, et al (2021) Incorporating domain knowledge for extractive summarization of legal case documents. In: Proceedings of the 18th International conference on artificial intelligence and law, pp 22–31
Devlin J, Chang M, Lee K, Toutanova K (2019) BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American chapter of the association for computational linguistics, pp 4171–4186
Dong L, Yang N, Wang W, et al (2019) Unified language model pre-training for natural language understanding and generation. In: 33rd conference on neural information processing systems, pp 13042–13054
Duan X, Wang B, Wang Z, et al (2019) CJRC: a reliable human-annotated benchmark dataset for Chinese judicial reading comprehension. In: Chinese computational linguistics - 18th China National Conference, pp 439–451
Elaraby M, Litman D (2022) ArgLegalSumm: Improving abstractive summarization of legal documents with argument mining. arXiv preprint arXiv:2209.01650
Erkan G, Radev DR (2004) Lexrank: graph-based lexical centrality as salience in text summarization. J Artif Intell Res 22:457–479
Gao Y, Liu Z, Li J, Guo F, Xiao F (2021) Extractive-abstractive summarization of judgment documents using multiple attention networks. In: Baroni P, et al. (eds) CLAR 2021, LNCS, vol 13040, pp 486–494
Hong Z, Zhou Q, Zhang R, Li W, Mo T (2020) Legal feature enhanced semantic matching network for similar case matching. In: 2020 international joint conference on neural networks, pp 1–8
Hu Z, Li X, Tu C, Liu Z, Sun M (2018) Few-shot charge prediction with discriminative legal attributes. In: Proceedings of the 27th international conference on computational linguistics, pp 487–498
Jain D, Borah MD, Biswas A (2020) Fine-tuning Textrank for legal document summarization: a Bayesian optimization based approach. In: Forum for information retrieval evaluation, pp 41–48
Jain D, Borah MD, Biswas A (2021) Summarization of legal documents: where are we now and the way forward. Comput Sci Rev 40:100388
Kulkarni S, Chammas S, et al (2021) CoMSum and SIBERT: a dataset and neural model for query-based multi-document summarization. In: 16th international conference on document analysis and recognition, pp 84–98
Lebanoff L, Song K, Dernoncourt F, et al (2019) Scoring Sentence Singletons and Pairs for Abstractive Summarization. In: Proceedings of the 57th conference of the association for computational linguistics, pp 2175–2189
Li D, Yang K, Zhang L, Yin D, Peng D (2021) CLASS: a novel method for Chinese legal judgments summarization. In: Proceedings of the 5th international conference on computer science and application engineering, pp 1–5
Li Y, Hu G, Du J, Abbas H, Zhang Y (2020) Multi-task reading for intelligent legal services. Futur Gener Comput Syst 113:218–227
Lin CY (2004) ROUGE: A package for automatic evaluation of summaries. In: Text summarization branches out, pp 74–81
Liu J, Wu J, Luo X (2021) Chinese judicial summarising based on short sentence extraction and GPT-2. In: Qiu H, Zhang C, Fei Z, et al (eds) KSEM 2021, LNCS, vol 12816, pp 376–393
Liu Y (2019) Fine-tune BERT for Extractive Summarization. arXiv preprint arXiv:1903.10318
Louis A (2014) A Bayesian method to incorporate background knowledge during automatic text summarization. In: Proceedings of the 52nd annual meeting of the association for computational linguistics, pp 333–338
Mihalcea R, Tarau P (2004) TextRank: Bringing order into text. In: Proceedings of the 2004 conference on empirical methods in natural language processing, pp 404–411
Nallapati R, Zhou B, et al (2016) Abstractive text summarization using sequence-to-sequence RNNs and beyond. In: Proceedings of the 20th SIGNLL conference on computational natural language learning, pp 280–290
Nguyen D, Nguyen B, et al (2021) Robust deep reinforcement learning for extractive legal summarization. In: Neural information processing - 28th international conference, pp 597–604
Polsley S, Jhunjhunwala P, Huang R (2016) CaseSummarizer: a system for automated summarization of legal texts. In: Proceedings of the 26th international conference on computational linguistics, pp 258–262
Radev DR, Jing H, Sty M, Tam D (2004) Centroid-based summarization of multiple documents. Inf Process Manage 40(6):919–938
Radford A, Wu J, Child R et al (2019) Language models are unsupervised multitask learners. OpenAI Blog 1(8):9
Rush AM, Chopra S, Weston J (2015) A neural attention model for abstractive sentence summarization. In: Proceedings of the 2015 conference on empirical methods in natural language processing, pp 379–389
See A, Liu PJ, Manning CD (2017) Get to the point: summarization with pointer-generator networks. In: Proceedings of the 55th annual meeting of the association for computational linguistics, pp 1073–1083
Tran V, Nguyen ML, Satoh K (2019) Building legal case retrieval systems with lexical matching and summarization using a pre-trained phrase scoring model. In: Proceedings of the seventeenth international conference on artificial intelligence and law, pp 275–282
Vaswani A, Shazeer N, Parmar N, et al (2017) Attention is all you need. In: Annual conference on neural information processing systems, pp 5998–6008
Vicente M, et al (2021) Can Text Summarization Enhance the Headline Stance Detection Task? Benefits and Drawbacks. In: 16th international conference on document analysis and recognition, pp 53–67
Xiao C, Hu X, Liu Z, Tu C, Sun M (2021) Lawformer: a pre-trained language model for chinese legal long documents. AI Open 2:79–84
Xiao C, Zhong H, Guo Z, et al (2019) CAIL2019-SCM: a dataset of similar case matching in legal domain. arXiv preprint arXiv:1911.08962
Yang W, Jia W, Zhou X, Luo Y (2019) Legal judgment prediction via multi-perspective bi-feedback network. In: Proceedings of the 28th international joint conference on artificial intelligence, pp 4085–4091
Yoon J, Muhammad J, Sajid A, Lee J (2022) Abstractive summarization of Korean legal cases using pre-trained language models. In: 16th international conference on ubiquitous information management and communication, pp 1–7
Zhong H, Guo Z, Tu C, et al (2018) Legal judgment prediction via topological learning. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 3540–3549
Zhong H, Xiao C, Tu C, Zhang T, Liu Z, Sun M (2020) How does NLP benefit legal system: a summary of legal artificial intelligence. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 5218–5230
Zhong H, Xiao C, Tu C, Zhang T, Liu Z, Sun M (2020) JEC-QA: a legal-domain question answering dataset. In: 34th AAAI conference on artificial intelligence, pp 9701–9708
Zhong L, Zhong Z, Zhao Z, et al (2019) Automatic summarization of legal decisions using iterative masking of predictive sentences. In: Proceedings of the seventeenth international conference on artificial intelligence and law, pp 163–172
Zhong M, Liu P, Wang D, et al (2019) Searching for effective neural extractive summarization: what works and what's next. In: Proceedings of the 57th conference of the association for computational linguistics, pp 1049–1058
Acknowledgements
This work was supported by Humanities and Social Science Planning Fund [Grant No. 21YJAZH013] from the Ministry of Education, China.
Author information
Authors and Affiliations
Corresponding authors
Ethics declarations
Conflict of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix: The English version of the example in case study
Appendix: The English version of the example in case study
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Dan, J., Hu, W. & Wang, Y. Enhancing legal judgment summarization with integrated semantic and structural information. Artif Intell Law (2023). https://doi.org/10.1007/s10506-023-09381-8
Accepted:
Published:
DOI: https://doi.org/10.1007/s10506-023-09381-8