{"title":"通过大型和小型模型的协作增强电荷预测","authors":"Bin Wei , Yaoyao Yu , Jiawen Zhang , Yiquan Wu","doi":"10.1016/j.clsr.2025.106168","DOIUrl":null,"url":null,"abstract":"<div><div>Charge prediction is a fundamental task in AI&Law, where the goal is to predict charges based on fact descriptions. Although various methods have been introduced to enhance performance, challenges remain. Specifically, small models (SMs)-based methods such as BERT struggle with hard cases involving low-frequency or confusing charges due to their limited capacity, whereas large language models (LLMs)-based approaches like GPT-4 exhibit difficulties in handling diverse charges owing to insufficient legal knowledge. To overcome these limitations, we propose a hybrid framework that collaborates both large and small models to improve charge prediction performance, based on the idea that combining the strengths of each can overcome their limitations. Initially, SMs provide an initial prediction along with a predicted probability distribution. If the maximum predicted probability falls below a threshold, LLMs step in to reflect and re-predict as needed. Additionally, we construct a confusing charges dictionary and design a two-stage legal inference prompt, which helps LLMs make the secondary prediction for the hard cases. Extensive experiments on two datasets from China and Italy demonstrate the effectiveness of this approach, yielding average F1 improvements of 7.94% and 11.46% respectively. Moreover, a fine-grained analysis demonstrates that our proposed framework is effective in identifying low-frequency and confusing charges.</div></div>","PeriodicalId":51516,"journal":{"name":"Computer Law & Security Review","volume":"58 ","pages":"Article 106168"},"PeriodicalIF":3.2000,"publicationDate":"2025-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Enhancing charge prediction through the collaboration of large and small models\",\"authors\":\"Bin Wei , Yaoyao Yu , Jiawen Zhang , Yiquan Wu\",\"doi\":\"10.1016/j.clsr.2025.106168\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Charge prediction is a fundamental task in AI&Law, where the goal is to predict charges based on fact descriptions. Although various methods have been introduced to enhance performance, challenges remain. Specifically, small models (SMs)-based methods such as BERT struggle with hard cases involving low-frequency or confusing charges due to their limited capacity, whereas large language models (LLMs)-based approaches like GPT-4 exhibit difficulties in handling diverse charges owing to insufficient legal knowledge. To overcome these limitations, we propose a hybrid framework that collaborates both large and small models to improve charge prediction performance, based on the idea that combining the strengths of each can overcome their limitations. Initially, SMs provide an initial prediction along with a predicted probability distribution. If the maximum predicted probability falls below a threshold, LLMs step in to reflect and re-predict as needed. Additionally, we construct a confusing charges dictionary and design a two-stage legal inference prompt, which helps LLMs make the secondary prediction for the hard cases. Extensive experiments on two datasets from China and Italy demonstrate the effectiveness of this approach, yielding average F1 improvements of 7.94% and 11.46% respectively. Moreover, a fine-grained analysis demonstrates that our proposed framework is effective in identifying low-frequency and confusing charges.</div></div>\",\"PeriodicalId\":51516,\"journal\":{\"name\":\"Computer Law & Security Review\",\"volume\":\"58 \",\"pages\":\"Article 106168\"},\"PeriodicalIF\":3.2000,\"publicationDate\":\"2025-07-29\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computer Law & Security Review\",\"FirstCategoryId\":\"90\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2212473X25000410\",\"RegionNum\":3,\"RegionCategory\":\"社会学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"LAW\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer Law & Security Review","FirstCategoryId":"90","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2212473X25000410","RegionNum":3,"RegionCategory":"社会学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"LAW","Score":null,"Total":0}
Enhancing charge prediction through the collaboration of large and small models
Charge prediction is a fundamental task in AI&Law, where the goal is to predict charges based on fact descriptions. Although various methods have been introduced to enhance performance, challenges remain. Specifically, small models (SMs)-based methods such as BERT struggle with hard cases involving low-frequency or confusing charges due to their limited capacity, whereas large language models (LLMs)-based approaches like GPT-4 exhibit difficulties in handling diverse charges owing to insufficient legal knowledge. To overcome these limitations, we propose a hybrid framework that collaborates both large and small models to improve charge prediction performance, based on the idea that combining the strengths of each can overcome their limitations. Initially, SMs provide an initial prediction along with a predicted probability distribution. If the maximum predicted probability falls below a threshold, LLMs step in to reflect and re-predict as needed. Additionally, we construct a confusing charges dictionary and design a two-stage legal inference prompt, which helps LLMs make the secondary prediction for the hard cases. Extensive experiments on two datasets from China and Italy demonstrate the effectiveness of this approach, yielding average F1 improvements of 7.94% and 11.46% respectively. Moreover, a fine-grained analysis demonstrates that our proposed framework is effective in identifying low-frequency and confusing charges.
期刊介绍:
CLSR publishes refereed academic and practitioner papers on topics such as Web 2.0, IT security, Identity management, ID cards, RFID, interference with privacy, Internet law, telecoms regulation, online broadcasting, intellectual property, software law, e-commerce, outsourcing, data protection, EU policy, freedom of information, computer security and many other topics. In addition it provides a regular update on European Union developments, national news from more than 20 jurisdictions in both Europe and the Pacific Rim. It is looking for papers within the subject area that display good quality legal analysis and new lines of legal thought or policy development that go beyond mere description of the subject area, however accurate that may be.