TY - GEN
T1 - Context-aware legal citation recommendation using deep learning
AU - Huang, Zihan
AU - Low, Charles
AU - Teng, Mengqiu
AU - Zhang, Hongyi
AU - Ho, Daniel E.
AU - Krass, Mark S.
AU - Grabmair, Matthias
N1 - Publisher Copyright:
© 2021 Owner/Author.
PY - 2021/6/21
Y1 - 2021/6/21
N2 - Lawyers and judges spend a large amount of time researching the proper legal authority to cite while drafting decisions. In this paper, we develop a citation recommendation tool that can help improve efficiency in the process of opinion drafting. We train four types of machine learning models, including a citation-list based method (collaborative filtering) and three context-based methods (text similarity, BiLSTM and RoBERTa classifiers). Our experiments show that leveraging local textual context improves recommendation, and that deep neural models achieve decent performance. We show that non-deep text-based methods benefit from access to structured case metadata, but deep models only benefit from such access when predicting from context of insufficient length. We also find that, even after extensive training, RoBERTa does not outperform a recurrent neural model, despite its benefits of pretraining. Our behavior analysis of the RoBERTa model further shows that predictive performance is stable across time and citation classes.
AB - Lawyers and judges spend a large amount of time researching the proper legal authority to cite while drafting decisions. In this paper, we develop a citation recommendation tool that can help improve efficiency in the process of opinion drafting. We train four types of machine learning models, including a citation-list based method (collaborative filtering) and three context-based methods (text similarity, BiLSTM and RoBERTa classifiers). Our experiments show that leveraging local textual context improves recommendation, and that deep neural models achieve decent performance. We show that non-deep text-based methods benefit from access to structured case metadata, but deep models only benefit from such access when predicting from context of insufficient length. We also find that, even after extensive training, RoBERTa does not outperform a recurrent neural model, despite its benefits of pretraining. Our behavior analysis of the RoBERTa model further shows that predictive performance is stable across time and citation classes.
KW - citation normalization
KW - citation recommendation
KW - legal opinion drafting
KW - legal text
KW - neural natural language processing
UR - http://www.scopus.com/inward/record.url?scp=85112357134&partnerID=8YFLogxK
U2 - 10.1145/3462757.3466066
DO - 10.1145/3462757.3466066
M3 - Conference contribution
AN - SCOPUS:85112357134
T3 - Proceedings of the 18th International Conference on Artificial Intelligence and Law, ICAIL 2021
SP - 79
EP - 88
BT - Proceedings of the 18th International Conference on Artificial Intelligence and Law, ICAIL 2021
PB - Association for Computing Machinery, Inc
T2 - 18th International Conference on Artificial Intelligence and Law, ICAIL 2021
Y2 - 21 June 2021 through 25 June 2021
ER -