Brian E. Perron, Hui Luan, Bryan G. Victor, Oliver Hiltz-Perron, Joseph Ryan
{"title":"Moving Beyond ChatGPT: Local Large Language Models (LLMs) and the Secure Analysis of Confidential Unstructured Text Data in Social Work Research","authors":"Brian E. Perron, Hui Luan, Bryan G. Victor, Oliver Hiltz-Perron, Joseph Ryan","doi":"10.1177/10497315241280686","DOIUrl":null,"url":null,"abstract":"Purpose: Large language models (LLMs) have demonstrated remarkable abilities in natural language tasks. However, their use in social work research is limited by confidentiality and security concerns when processing sensitive data. This study addresses these challenges by evaluating the performance of local LLMs (LocalLLMs) in classifying and extracting substance-related problems from unstructured child welfare investigation summaries. LocalLLMs allow researchers to analyze data on their own computers without transmitting information to external servers for processing. Methods: Four state-of-the-art LocalLLMs—Mistral-7b, Mixtral-8 × 7b, LLama3-8b, and Llama3-70b—were tested using zero-shot prompting on 2,956 manually coded summaries. Results: The LocalLLMs achieved exceptional results comparable to human experts in classification and extraction, demonstrating their potential to unlock valuable insights from confidential, unstructured child welfare data. Conclusions: This study highlights the feasibility of using LocalLLMs to efficiently analyze large amounts of textual data while addressing the confidentiality issues associated with proprietary LLMs.","PeriodicalId":47993,"journal":{"name":"Research on Social Work Practice","volume":"220 1","pages":""},"PeriodicalIF":1.7000,"publicationDate":"2024-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Research on Social Work Practice","FirstCategoryId":"90","ListUrlMain":"https://doi.org/10.1177/10497315241280686","RegionNum":4,"RegionCategory":"社会学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"SOCIAL WORK","Score":null,"Total":0}
引用次数: 0
Abstract
Purpose: Large language models (LLMs) have demonstrated remarkable abilities in natural language tasks. However, their use in social work research is limited by confidentiality and security concerns when processing sensitive data. This study addresses these challenges by evaluating the performance of local LLMs (LocalLLMs) in classifying and extracting substance-related problems from unstructured child welfare investigation summaries. LocalLLMs allow researchers to analyze data on their own computers without transmitting information to external servers for processing. Methods: Four state-of-the-art LocalLLMs—Mistral-7b, Mixtral-8 × 7b, LLama3-8b, and Llama3-70b—were tested using zero-shot prompting on 2,956 manually coded summaries. Results: The LocalLLMs achieved exceptional results comparable to human experts in classification and extraction, demonstrating their potential to unlock valuable insights from confidential, unstructured child welfare data. Conclusions: This study highlights the feasibility of using LocalLLMs to efficiently analyze large amounts of textual data while addressing the confidentiality issues associated with proprietary LLMs.
期刊介绍:
Research on Social Work Practice, sponsored by the Society for Social Work and Research, is a disciplinary journal devoted to the publication of empirical research concerning the methods and outcomes of social work practice. Social work practice is broadly interpreted to refer to the application of intentionally designed social work intervention programs to problems of societal and/or interpersonal importance, including behavior analysis or psychotherapy involving individuals; case management; practice involving couples, families, and small groups; community practice education; and the development, implementation, and evaluation of social policies.