Skip to main content
Log in

Enhancing legal judgment summarization with integrated semantic and structural information

  • Original Research
  • Published:
Artificial Intelligence and Law Aims and scope Submit manuscript

Abstract

Legal Judgment Summarization (LJS) can highly summarize legal judgment documents, improving judicial work efficiency in case retrieval and other occasions. Legal judgment documents are usually lengthy; however, most existing LJS methods are directly based on general text summarization models, which cannot handle long texts effectively. Additionally, due to the complex structural characteristics of legal judgment documents, some information may be lost by applying only one single kind of summarization model. To address these issues, we propose an integrated summarization method which leverages both semantic and structural information to improve the quality of LJS. Specifically, legal judgment documents are firstly segmented into three relatively short parts according to their specific structure. We propose an extractive summarization model named BSLT and an abstractive summarization model named LPGN by adopting Lawformer as the encoder. Lawformer is a new pre-trained language model for long legal documents, which specializes in capturing long-distance dependency and modeling legal semantic features. Then, we adopt different models to summarize the corresponding part regarding its structural characteristics. Finally, the obtained summaries are integrated to generate a high-quality summary involving semantic and structural information. We conduct comparative experiments to evaluate the performance of our model. The results show that our model outperforms the baseline model LEAD-3 by 14.78% on the mean ROUGE score, which demonstrates our method is effective in LJS and is prospected to be applied to assist other tasks in legal artificial intelligence.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

Notes

  1. http://cail.cipsc.org.cn/.

  2. https://github.com/china-ai-law-challenge/CAIL2020/tree/master/sfzy/baseline.

  3. https://github.com/YunwenTechnology/Unilm

References

  • Anand D, Wagh R (2022) Effective deep learning approaches for summarization of legal texts. J King Saud Univ Comput Inf Sci 34(5):2141–2150

    Google Scholar 

  • Bajaj A, Dangati P, Krishna K, et al (2021) Long document summarization in a low resource setting using pretrained language models. In: Proceedings of the ACL-IJCNLP 2021 student research workshop, pp 71–80

  • Bhattacharya P, Poddar S, Rudra K, et al (2021) Incorporating domain knowledge for extractive summarization of legal case documents. In: Proceedings of the 18th International conference on artificial intelligence and law, pp 22–31

  • Devlin J, Chang M, Lee K, Toutanova K (2019) BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American chapter of the association for computational linguistics, pp 4171–4186

  • Dong L, Yang N, Wang W, et al (2019) Unified language model pre-training for natural language understanding and generation. In: 33rd conference on neural information processing systems, pp 13042–13054

  • Duan X, Wang B, Wang Z, et al (2019) CJRC: a reliable human-annotated benchmark dataset for Chinese judicial reading comprehension. In: Chinese computational linguistics - 18th China National Conference, pp 439–451

  • Elaraby M, Litman D (2022) ArgLegalSumm: Improving abstractive summarization of legal documents with argument mining. arXiv preprint arXiv:2209.01650

  • Erkan G, Radev DR (2004) Lexrank: graph-based lexical centrality as salience in text summarization. J Artif Intell Res 22:457–479

    Article  Google Scholar 

  • Gao Y, Liu Z, Li J, Guo F, Xiao F (2021) Extractive-abstractive summarization of judgment documents using multiple attention networks. In: Baroni P, et al. (eds) CLAR 2021, LNCS, vol 13040, pp 486–494

  • Hong Z, Zhou Q, Zhang R, Li W, Mo T (2020) Legal feature enhanced semantic matching network for similar case matching. In: 2020 international joint conference on neural networks, pp 1–8

  • Hu Z, Li X, Tu C, Liu Z, Sun M (2018) Few-shot charge prediction with discriminative legal attributes. In: Proceedings of the 27th international conference on computational linguistics, pp 487–498

  • Jain D, Borah MD, Biswas A (2020) Fine-tuning Textrank for legal document summarization: a Bayesian optimization based approach. In: Forum for information retrieval evaluation, pp 41–48

  • Jain D, Borah MD, Biswas A (2021) Summarization of legal documents: where are we now and the way forward. Comput Sci Rev 40:100388

    Article  Google Scholar 

  • Kulkarni S, Chammas S, et al (2021) CoMSum and SIBERT: a dataset and neural model for query-based multi-document summarization. In: 16th international conference on document analysis and recognition, pp 84–98

  • Lebanoff L, Song K, Dernoncourt F, et al (2019) Scoring Sentence Singletons and Pairs for Abstractive Summarization. In: Proceedings of the 57th conference of the association for computational linguistics, pp 2175–2189

  • Li D, Yang K, Zhang L, Yin D, Peng D (2021) CLASS: a novel method for Chinese legal judgments summarization. In: Proceedings of the 5th international conference on computer science and application engineering, pp 1–5

  • Li Y, Hu G, Du J, Abbas H, Zhang Y (2020) Multi-task reading for intelligent legal services. Futur Gener Comput Syst 113:218–227

    Article  Google Scholar 

  • Lin CY (2004) ROUGE: A package for automatic evaluation of summaries. In: Text summarization branches out, pp 74–81

  • Liu J, Wu J, Luo X (2021) Chinese judicial summarising based on short sentence extraction and GPT-2. In: Qiu H, Zhang C, Fei Z, et al (eds) KSEM 2021, LNCS, vol 12816, pp 376–393

  • Liu Y (2019) Fine-tune BERT for Extractive Summarization. arXiv preprint arXiv:1903.10318

  • Louis A (2014) A Bayesian method to incorporate background knowledge during automatic text summarization. In: Proceedings of the 52nd annual meeting of the association for computational linguistics, pp 333–338

  • Mihalcea R, Tarau P (2004) TextRank: Bringing order into text. In: Proceedings of the 2004 conference on empirical methods in natural language processing, pp 404–411

  • Nallapati R, Zhou B, et al (2016) Abstractive text summarization using sequence-to-sequence RNNs and beyond. In: Proceedings of the 20th SIGNLL conference on computational natural language learning, pp 280–290

  • Nguyen D, Nguyen B, et al (2021) Robust deep reinforcement learning for extractive legal summarization. In: Neural information processing - 28th international conference, pp 597–604

  • Polsley S, Jhunjhunwala P, Huang R (2016) CaseSummarizer: a system for automated summarization of legal texts. In: Proceedings of the 26th international conference on computational linguistics, pp 258–262

  • Radev DR, Jing H, Sty M, Tam D (2004) Centroid-based summarization of multiple documents. Inf Process Manage 40(6):919–938

    Article  MATH  Google Scholar 

  • Radford A, Wu J, Child R et al (2019) Language models are unsupervised multitask learners. OpenAI Blog 1(8):9

    Google Scholar 

  • Rush AM, Chopra S, Weston J (2015) A neural attention model for abstractive sentence summarization. In: Proceedings of the 2015 conference on empirical methods in natural language processing, pp 379–389

  • See A, Liu PJ, Manning CD (2017) Get to the point: summarization with pointer-generator networks. In: Proceedings of the 55th annual meeting of the association for computational linguistics, pp 1073–1083

  • Tran V, Nguyen ML, Satoh K (2019) Building legal case retrieval systems with lexical matching and summarization using a pre-trained phrase scoring model. In: Proceedings of the seventeenth international conference on artificial intelligence and law, pp 275–282

  • Vaswani A, Shazeer N, Parmar N, et al (2017) Attention is all you need. In: Annual conference on neural information processing systems, pp 5998–6008

  • Vicente M, et al (2021) Can Text Summarization Enhance the Headline Stance Detection Task? Benefits and Drawbacks. In: 16th international conference on document analysis and recognition, pp 53–67

  • Xiao C, Hu X, Liu Z, Tu C, Sun M (2021) Lawformer: a pre-trained language model for chinese legal long documents. AI Open 2:79–84

    Article  Google Scholar 

  • Xiao C, Zhong H, Guo Z, et al (2019) CAIL2019-SCM: a dataset of similar case matching in legal domain. arXiv preprint arXiv:1911.08962

  • Yang W, Jia W, Zhou X, Luo Y (2019) Legal judgment prediction via multi-perspective bi-feedback network. In: Proceedings of the 28th international joint conference on artificial intelligence, pp 4085–4091

  • Yoon J, Muhammad J, Sajid A, Lee J (2022) Abstractive summarization of Korean legal cases using pre-trained language models. In: 16th international conference on ubiquitous information management and communication, pp 1–7

  • Zhong H, Guo Z, Tu C, et al (2018) Legal judgment prediction via topological learning. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 3540–3549

  • Zhong H, Xiao C, Tu C, Zhang T, Liu Z, Sun M (2020) How does NLP benefit legal system: a summary of legal artificial intelligence. In: Proceedings of the 58th annual meeting of the association for computational linguistics, pp 5218–5230

  • Zhong H, Xiao C, Tu C, Zhang T, Liu Z, Sun M (2020) JEC-QA: a legal-domain question answering dataset. In: 34th AAAI conference on artificial intelligence, pp 9701–9708

  • Zhong L, Zhong Z, Zhao Z, et al (2019) Automatic summarization of legal decisions using iterative masking of predictive sentences. In: Proceedings of the seventeenth international conference on artificial intelligence and law, pp 163–172

  • Zhong M, Liu P, Wang D, et al (2019) Searching for effective neural extractive summarization: what works and what's next. In: Proceedings of the 57th conference of the association for computational linguistics, pp 1049–1058

Download references

Acknowledgements

This work was supported by Humanities and Social Science Planning Fund [Grant No. 21YJAZH013] from the Ministry of Education, China.

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Jingpei Dan or Yuming Wang.

Ethics declarations

Conflict of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix: The English version of the example in case study

Appendix: The English version of the example in case study

figure a
figure b

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Dan, J., Hu, W. & Wang, Y. Enhancing legal judgment summarization with integrated semantic and structural information. Artif Intell Law (2023). https://doi.org/10.1007/s10506-023-09381-8

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10506-023-09381-8

Keywords

Navigation