Judicial document intellectual processing using hybrid deep neural networks

WANG Wenguan, CHEN Yunwen, CAI Hua, ZENG Yanneng, YANG Huiyu

Journal of Tsinghua University(Science and Technology) ›› 2019, Vol. 59 ›› Issue (7) : 505-511.

PDF(1250 KB)
PDF(1250 KB)
Journal of Tsinghua University(Science and Technology) ›› 2019, Vol. 59 ›› Issue (7) : 505-511. DOI: 10.16511/j.cnki.qhdxxb.2019.21.015
SPECIAL SECTION: AI AND LAW

Judicial document intellectual processing using hybrid deep neural networks

  • {{article.zuoZhe_EN}}
Author information +
History +

Abstract

This article presents a neural network model for crime prediction, legal article recommendation, and sentence prediction from judicial documents. The model is based on a hybrid attention and CNN model which combines the improved hierarchical attention network (iHAN) and the deep pyramid convolutional neural network (DPCNN) by ResNet. The F1-Scores (mean value of Micro-F1 and Macro-F1) for the crime prediction and related law samples from CAIL-2018 were 85% and 87%. The sentence prediction accuracy is impacted by differences in locations, dates, courts, judges, and defendant attitudes. The model adjusts well to these differences because of its high predictive ability and model generalization. The model prediction outputs for the recommended crime prediction and law items were then added to the model input for the sentence prediction task to further improve the model performance. The model got an excellent result in the sentence prediction task (CAIL-2018) with an F1-Score of over 77%.

Key words

judicial document processing / natural language understanding / verdict prediction / deep neural networks / attention model

Cite this article

Download Citations
WANG Wenguan, CHEN Yunwen, CAI Hua, ZENG Yanneng, YANG Huiyu. Judicial document intellectual processing using hybrid deep neural networks[J]. Journal of Tsinghua University(Science and Technology). 2019, 59(7): 505-511 https://doi.org/10.16511/j.cnki.qhdxxb.2019.21.015

References

[1] KATZ D M. Quantitative legal prediction-or-how I learned to stop worrying and start preparing for the data-driven future of the legal services industry[J]. Emory LJ, 2012, 62:909-966.
[2] ASHLEY K, BRANTING K, MARGOLIS H, et al. Legal reasoning and artificial intelligence:How computers" think" like lawyers[J]. University of Chicago Law School Roundtable, 2001, 8(1):1-28.
[3] SURDEN H. Machine learning and law[J]. Washington Law Review, 2014, 89:87-115.
[4] ALETRAS N, TSARAPATSANIS D, PREOA AT'XU XUIUC-PIETRO D, et al. Predicting judicial decisions of the European Court of Human Rights:A natural language processing perspective[J]. PeerJ Computer Science, 2016, 2:1-19.
[5] WILLIAMS T. Civilisation 2030:The near future for law firms[R]. London, UK:Jomati Consultants, 2014.
[6] NAVIGLI R. Natural language understanding:Instructions for (present and future) use[C]//Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI). California, USA:Morgan Kaufman, 2018:5697-5702.
[7] MIKOLOV T, SUTSKEVER I, CHEN K, et al. Distributed representations of words and phrases and their compositionality[C]//Advances in Neural Information Processing Systems (NIPS). Massachusetts, USA:MIT Press, 2013:3111-3119.
[8] PENNINGTON J, SOCHER R, MANNING C. Glove:Global vectors for word representation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Stroudsburg, PA, USA:ACL,2014:1532-1543.
[9] JOULIN A, GRAVE E, BOJANOWSKI P, et al. Bag of tricks for efficient text classification[C]//Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics. Stroudsburg, PA, USA:ACL, 2017:427-431.
[10] LIU P F, QIU X P, HUANG X J. Recurrent neural network for text classification with multi-task learning[C]//Proceedings of the 25th International Joint Conference on Artificial Intelligence(IJCAI). California, USA:Morgan Kaufman, 2016:2873-2879.
[11] KIM Y. Convolutional neural networks for sentence classification[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Stroudsburg, PA, USA:ACL, 2014:1746-1751
[12] CONNEAU A, SCHWENK H, BARRAULT L, et al. Very deep convolutional networks for text classification[C]//Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics. Stroudsburg, PA, USA:ACL, 2017:1107-1116.
[13] CHO K, VAN MERRIЁNBOER B, GULCEHRE C, et al. Learning phrase representations using RNN encoder-decoder for statistical machine translation[C]//Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Stroudsburg, PA, USA:ACL, 2014:1724-1734.
[14] JOHNSON R, ZHANG T. Deep pyramid convolutional neural networks for text categorization[C]//Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. Stroudsburg, PA, USA:ACL, 2017:562-570.
[15] ZHOU X, WAN X J, XIAO J. Attention-based LSTM network for cross-lingual sentiment classification[C]//Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing (EMNLP). Stroudsburg, PA, USA:ACL, 2016:247-256.
[16] YANG Z, YANG D, DYER C, et al. Hierarchical attention networks for document classification[C]//Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies. Stroudsburg, PA, USA:ACL, 2016:1480-1489.
[17] XIAO C J, ZHONG H X, GUO Z P, et al. Cail2018:A large-scale legal dataset for judgment prediction[J/OL]. (2018-07-04). https://arxiv.org/abs/1807.02478
[18] CHEN X X, XU L, LIU Z Y, et al. Joint learning of character and word embeddings[C]//Proceedings of the 24th International Joint Conference on Artificial Intelligence (IJCAI). California, USA:Morgan Kaufman, 2015:1236-1242.
[19] QIAO C, HUANG B, NIU G C, et al. A new method of region embedding for text classification[C]//International Conference on Learning Representations. Massachusetts, USA:OpenReview, 2018:1-12.
[20] VASWANI A, SHAZEER N, PARMAR N, et al. Attention is all you need[C]//Advances in Neural Information Processing Systems (NIPS). Massachusetts, USA:MIT Press, 2017:5998-6008.
[21] SHEN Y T, HUANG X J. Attention-based convolutional neural network for semantic relation extraction[C]//The 26th International Conference on Computational Linguistics (COLIN). New York, NY, USA:ACM, 2016:2526-2536.
[22] YIN W, SCHVTZE H. Attentive convolution:Equipping CNNs with RNN-style attention mechanisms[J/OL]. (2017-10-02). https://arxiv.org/abs/1710.00519
[23] 顾永忠. 浅析个案间量刑的失衡[J]. 人民司法, 2003, 2:64-66. GU Y Z. Analysis on the unbalance of sentencing between cases[J]. People's Judicature. 2003,2:64-66. (in Chinese)
[24] 龙光伟. 论量刑失衡及其对策[J]. 吉林大学社会科学学报, 2003,2:58-61. LONG G W. On the unbalance of measurement of penalty and the countermeasure[J]. Jilin University Journal Social Sciences Edition. 2003,2:58-61. (in Chinese)
PDF(1250 KB)

Accesses

Citation

Detail

Sections
Recommended

/