CaseEncoder: A Knowledge-enhanced Pre-trained Model for Legal Case Encoding

被引:0
|
作者
Ma, Yixiao [1 ]
Wu, Yueyue [2 ,3 ,4 ]
Su, Weihang [2 ,3 ,4 ]
Ai, Qingyao [2 ,3 ,4 ]
Liu, Yiqun [2 ,3 ,4 ]
机构
[1] Huawei Cloud BU, Shenzhen, Guangdong, Peoples R China
[2] Quan Cheng Lab, Nanjing, Peoples R China
[3] Tsinghua Univ, Inst Internet Judiciary, Beijing, Peoples R China
[4] Tsinghua Univ, DCST, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Legal case retrieval is a critical process for modern legal information systems. While recent studies have utilized pre-trained language models (PLMs) based on the general domain self-supervised pre-training paradigm to build models for legal case retrieval, there are limitations in using general domain PLMs as backbones. Specifically, these models may not fully capture the underlying legal features in legal case documents. To address this issue, we propose CaseEncoder, a legal document encoder that leverages fine-grained legal knowledge in both the data sampling and pre-training phases. In the data sampling phase, we enhance the quality of the training data by utilizing fine-grained law article information to guide the selection of positive and negative examples. In the pre-training phase, we design legal-specific pre-training tasks that align with the judging criteria of relevant legal cases. Based on these tasks, we introduce an innovative loss function called Biased Circle Loss to enhance the model's ability to recognize case relevance in fine grains. Experimental results on multiple benchmarks demonstrate that CaseEncoder significantly outperforms both existing general pre-training models and legal-specific pre-training models in zero-shot legal case retrieval. The source code of CaseEncoder can be found at https://github.com/myx666/CaseEncoder.
引用
收藏
页码:7134 / 7143
页数:10
相关论文
共 50 条
  • [41] Classifying informative tweets using feature enhanced pre-trained language model
    Yandrapati, Prakash Babu
    Eswari, R.
    SOCIAL NETWORK ANALYSIS AND MINING, 2024, 14 (01)
  • [42] Patent classification with pre-trained Bert model
    Kahraman, Selen Yuecesoy
    Durmusoglu, Alptekin
    Dereli, Tuerkay
    JOURNAL OF THE FACULTY OF ENGINEERING AND ARCHITECTURE OF GAZI UNIVERSITY, 2024, 39 (04): : 2485 - 2496
  • [43] Adder Encoder for Pre-trained Language Model
    Ding, Jianbang
    Zhang, Suiyun
    Li, Linlin
    CHINESE COMPUTATIONAL LINGUISTICS, CCL 2023, 2023, 14232 : 339 - 347
  • [44] Knowledge-Enhanced Graph Encoding Method for Metaphor Detection in Text
    Huang H.
    Liu X.
    Liu Q.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2023, 60 (01): : 140 - 152
  • [45] AdaDS: Adaptive data selection for accelerating pre-trained language model knowledge distillation
    Zhou, Qinhong
    Li, Peng
    Liu, Yang
    Guan, Yuyang
    Xing, Qizhou
    Chen, Ming
    Sun, Maosong
    Liu, Yang
    AI OPEN, 2023, 4 : 56 - 63
  • [46] Knowledge graph extension with a pre-trained language model via unified learning method
    Choi, Bonggeun
    Ko, Youngjoong
    KNOWLEDGE-BASED SYSTEMS, 2023, 262
  • [47] Infusing factual knowledge into pre-trained model for finding the contributions from the research articles
    Gupta, Komal
    Ghosal, Tirthankar
    Ekbal, Asif
    JOURNAL OF INFORMATION SCIENCE, 2024,
  • [48] Knowledge-enhanced model with dual-graph interaction for confusing legal charge prediction
    Bi, Sheng
    Ali, Zafar
    Wu, Tianxing
    Qi, Guilin
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 249
  • [49] Semorph: A Morphology Semantic Enhanced Pre-trained Model for Chinese Spam Text Detection
    Lai, Kaiting
    Long, Yinong
    Wu, Bowen
    Li, Ying
    Wang, Baoxun
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 1003 - 1013
  • [50] A Joint Label-Enhanced Representation Based on Pre-trained Model for Charge Prediction
    Dan, Jingpei
    Liao, Xiaoshuang
    Xu, Lanlin
    Hu, Weixuan
    Zhang, Tianyuan
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 694 - 705