{"title":"奋起迎接生成式人工智能的挑战","authors":"Inara Scott","doi":"10.1111/jlse.12141","DOIUrl":null,"url":null,"abstract":"<p>ChatGPT launched in November 2022 and quickly became the fastest-growing user application in history, marking one million users in two months—a milestone that took TikTok nine months to achieve and Instagram two and a half years.1 That explosive growth has come with an explosion of concern for the ability of scientists and regulators to understand what it is, how it works,2 and its potential to change life as we know it. Politicians and technology executives alike are calling for more and better regulation to ensure apocalyptic scenarios of artificial intelligence (AI)-aided disasters (everything from AI-created weapons to sentient AI systems) do not come to pass.3 Meanwhile, the practice of law is dealing with the implications of a tool that can pass the Bar exam4 and colleges and universities are grappling with the reality that students can use generative AI to complete just about any assignment they can give.5</p><p>This article is not intended to “solve” the problem of generative AI. Rather, recognizing the astonishing pace of development of generative AI tools and their impact on business law and other higher education classes, it seeks to provide specific, concrete steps that faculty can take to evolve alongside these tools. There is no way to “AI-proof” your classes. However, taking the steps outlined here can help you decide <i>what</i> you want to teach and <i>how</i> you should teach it. It offers a structure for identifying the content you want to keep and what you can let go of and tips for redesigning assignments and syllabi to clarify your approach to students and reduce academic misconduct.</p><p>To understand the profound impact of generative AI and tools like ChatGPT, it is helpful to begin by unpacking some of the language that is used in this field.6 AI, short for “artificial intelligence,” generally refers to the use of machines, particularly computers, to perform tasks that simulate the use of human thinking, reasoning, and learning.7 We encounter AI throughout our day, when using computer applications like a Google search, interacting with a chatbot on a consumer website, or using a virtual assistant like Siri or Alexa. The ubiquity of AI in our daily lives is predicted to increase.8 In the near term, AI's use will likely become more ubiquitous as it becomes embedded in the way we interact with everyday items like our cars, office computers, and coffee machines.9</p><p>The term artificial “intelligence” is controversial because machines do not actually have the capacity to think or learn like humans. Their programming can <i>simulate</i> some aspects of human intelligence, but they do not reason like a human.10 For this reason, scientists distinguish between “strong AI” or “artificial general intelligence” and “weak AI.”11 Weak AI is what we generally have in use today—computers simulating human intelligence while they complete a specific type of task they have been programmed to perform. Strong AI is a theoretical system that thinks, reasons, and has self-awareness like a human (and could evolve to exceed human intelligence—thus prompting doomsday “AI takes over the world” scenarios).</p><p>Machine learning is a type of AI that programs systems to “learn” and improve their ability to perform assigned tasks without explicit programming.12 Machine learning relies on algorithms that allow machines to sort vast quantities of data and use that data to make predictions and find patterns, ultimately teaching themselves through repetition and human intervention, while human programmers tweak the algorithms that guide the learning process.13 The most powerful and complex AI applications today use <i>neural networks</i>—a type of machine learning that simulates the way neurons in the human brain communicate information.14</p><p>Generative AI refers to the use of AI applications to create something new, including images, text, and other forms of media. Large language models (LLMs) use machine learning and highly sophisticated neural networks to process and generate text.15 LLMs are trained on enormous datasets with billions of inputs—imagine starting by dumping all of the Internet into a giant bucket and you get the idea. Somewhat disturbingly, the contents of those databases are not transparent,16 and litigation has alleged that the databases likely contain a significant amount of copyrighted material.17 The LLM uses this dataset to process inputs (i.e., human prompts to the LLM interface) and predict the most helpful, common, and relevant responses.18 In effect, the LLM crowdsources answers to the questions it is asked, by using its skills in pattern identification and prediction to generate the most likely and useful textual responses given the information in its database. The tools we know as ChatGPT, Bard, or Bing are simply interfaces using LLMs to produce responses to prompts from external users.</p><p>Generative AI's flaws are predictable, given how the technology works. Though as humans, we are likely to anthropomorphize LLMs and attribute them with human characteristics,19 they are not thinking beings—they are simply computer models with predictable flaws arising out of an astonishingly complex but ultimately nonhuman system.</p><p>One of the most glaring flaws of LLMs is their potential to produce inaccurate, false, or flawed responses, in which the model puts together strings of text that do not point to accurate information but mimic accurate information in a compelling way. For example, when prompted for a peer-reviewed journal article, an LLM might string together a series of notations that look like a journal article, with a title, volume, issue, and page numbers but which does not point to an existing source. LLMs can create a variety of types of false information, such as false journal articles, false meetings of historical figures, and false medical information.20 Early in the use of LLMs, these responses were dubbed “hallucinations.”21 Notably, even that term is anthropomorphically loaded—it suggests the LLM is operating in a semi-conscious human way, “imagining” events that did not occur, rather than simply acting as a flawed computer model producing inaccurate results.</p><p>Another essential limitation of LLMs is also rooted in their technological structure. LLMs are, by their nature, predicting and generating text that aligns with information already in their database. The information in their database was generally created by humans—flawed and biased humans, producing information that is also flawed and biased, operating in societies with existing structural racism and bias. If faced with a question like “please list the top twenty jazz pianists,” the LLM does not independently assess the work of all living and dead jazz pianists and reach a conclusion about their relative merit. Rather, it crowdsources information, identifying patterns in commonly reported responses to this question and developing a prediction based on that information. LLMs cannot account for historical inequities and biases in reporting on artists, let alone the biases that go into the development of artists and their access to recording contracts or critical acclaim.22 In an area like jazz music, with deep historical roots of inequities in the treatment of men and women,23 one cannot expect an LLM to do anything other than repeat existing biases, which are reflected in popular source materials. As a result, when queried, ChatGPT, crowdsourcing from existing sources, reported no women in the top 20 jazz pianists.24 ChatGPT produces this result, despite the historical significance of figures like Mary Lou Williams, Alice Coltrane, and Shirley Scott, iconic female artists well-known to jazz scholars and aficionados alike.25</p><p>DALL-E, a generative AI tool that creates images based on textual prompts, produces similarly biased responses. When prompted to create a “realistic photograph quality picture of a college professor talking to a large group of students,” DALL-E created the following image, of a white male professor surrounded by white students Figure 1:</p><p>Is this image “incorrect”? Survey data from 2019 suggests over seventy-five percent of professors are white, while fifty-five percent of undergraduates are white.26 Data from 2021 show white men continue to represent the largest percentage of tenure-track faculty.27 Given this landscape, it is entirely consistent with AI's programming to generate images that center white men as professors and to leave women out of a list of top jazz pianists.</p><p>Of course, the answer begets a closer look at the question. If we are asking which jazz musicians have been most widely recognized or what the “average” college professor looks like, the tools are performing as programmed. But there are many other questions embedded in this call and response, and they get at the heart of what it means to use generative AI in education. If the average user of an LLM assumes that this list of jazz pianists <i>accurately</i> represents a list of the <i>most talented</i> jazz pianists ever to have lived, the technology is reinforcing inaccurate information based on existing societal biases. To be clear, the LLM cannot and is not answering that question. The LLM, in this scenario, is also not considering the question of what cultural or legal barriers might prevent or have prevented women from becoming known in the field of jazz music (though it could, if directly asked).28 It is simply reproducing a list of commonly mentioned “top” jazz pianists.</p><p>Similarly, DALL-E is not “incorrect” when it creates a likely image of a professor as white and male. The problem here is that research suggests the images we view shape our ideas of what is possible.29 Thus, if popular media images are dominated by the “most likely” outcomes generated by AI models, stereotypical roles will be harder to reject and move past. Generative AI can emphasize and blind us to the way our cultural biases show up in everything from the medical literature to our notions of beauty because they are, by their programming, designed to reflect a majoritarian view of the world.</p><p>To be clear, DALL-E and ChatGPT are not different in this action than existing search engines or media companies, which are also built on crowdsourcing and majoritarian views. The challenge is that, as these tools become ubiquitous and embedded in our daily lives, we may come to view them as definitive sources of <i>truth</i>, even more so than we currently do with search engines. When a student asks ChatGPT a medical question and receives a conversational response that mimics a medical professional, they may assume the answer they are receiving is accurate for all populations rather than understanding that the information is likely based on a medical study of young, white men and may not be representative of the population as a whole.30</p><p>These are two of the most prominent concerns about generative AI, but there are other flaws worth noting. Generative AI cannot replace the need for professionals in real life to “think on their feet” and react quickly in scenarios that require human empathy and communication skills. It cannot respond in real time to complex scenarios with constantly changing fact patterns and complicated human actors. It bases its responses on existing scenarios and patterns, so it is unlikely to make truly innovative, creative leaps. Because it lacks human intuition and feeling, it also cannot look below the surface of a question to respond to a deeper, more essential question, or even draw out the question that is not asked. As a consultant might say, “AI can only give the client what they ask for—so my job is secure.” Implied: We often do not know what we want, in business or in life, and we rely on the insight from the humans around us to help identify our underlying needs.</p><p>One approach or response to the wave of generative AI would be to plant a flag in the ground and refuse to move or evolve. Rather than changing the content of their courses, in this scenario, faculty might argue that universities and colleges should instead institute more stringent academic integrity policies and punishments and develop better ways to “catch the cheaters.” This position is fundamentally flawed, for two primary reasons: (1) faculty will never be able to “AI-proof” their courses, and (2) learning the responsible use of AI is essential to students’ future success.</p><p>First, while creating an environment that supports academic integrity is important, some assignments and assessments will always be able to evade restrictions on the use of AI. While many faculty have reasonably decided to turn toward more in-class assignments and assessments to reduce cheating, this option is not available to those teaching online. Moreover, not all assignments can be completed during class hours.</p><p>What about developing more and better ways to catch the cheaters? That is, can we not just proctor our way out of this situation? Even keeping in mind that online proctoring has limited application and can only be used for time-limited assessments, the reality is that many universities had begun moving away from AI-enabled proctoring <i>before</i> ChatGPT came on the scene due to concerns about privacy, discrimination, and unreliability.31</p><p>What about AI-text detectors like ZeroGPT and GPTzero?32 Can we not use AI to detect AI? Unfortunately, these tools have been shown to be unreliable33 and biased against non-native English writers.34 In addition, work-arounds are notoriously easy to find, most commonly (and ironically) by using AI paraphrasing tools.35</p><p>Faculty can create assignments that are <i>easier</i> to complete with generative AI and those that are <i>harder</i>. In most cases, however, particularly in large undergraduate survey courses, generative AI will be able to complete most assignments, particularly those completed outside of class. If ChatGPT can pass the Bar exam, it can probably draft a better legal essay than most undergraduate students. It certainly can undertake any multiple-choice questions a faculty member might write.</p><p>This means students will have to decide to complete their assignments knowing that they have the answer key (via ChatGPT) and knowing faculty probably cannot catch them cheating. Students inclined to cheat (for any number of reasons, which might include the struggle of balancing school with full-time work,36 caring for family members,37 or dealing with other challenging life circumstances like food insecurity38 or physical or mental illness39) now have a straightforward way to successfully complete their assignments. Students who do try to complete their assignments without assistance may as a result be at a disadvantage, particularly in classes using curved grading, and faculty would not be able to tell one group from the other. This could lead to grading inequities, not to mention even greater incentives for students who would not normally bend the rules, if only to be able to stand on a level playing field.</p><p>Second, and I believe even more importantly, what faculty teach <i>should have relevance</i> to the way their students will undertake their future work. For those teaching undergraduate business law courses at a business school, students are unlikely to become lawyers. Rather, they will most likely be applying legal content to business situations. They need to learn how to apply the information they are learning to these real-world scenarios. Because their real world will likely include generative AI, learning how to use it safely, responsibly, and effectively to identify legal risks and strategic legal opportunities and how to fill in the known gaps and defects of the technology, particularly as relates to bias and inaccuracies.</p><p>In my experience, faculty want to evolve but are not sure how. Even knowing that students will be likely to use AI tools in their future work, many faculty remain convinced that there is some basic, fundamental information that students need to use the technology responsibly and to avoid the limitations described above. At the same time, faculty are unsure how to integrate this technology into their courses while retaining the essential information students need and focusing on the human-thinking skills that generative AI cannot provide. Finally, many are aware of the limitations of generative AI discussed above and want to avoid using the tool in a way that furthers biases or inhibits creative and critical thinking.</p><p>In this part, I begin to address these challenges. I discuss how to start using generative AI and how to adapt the curriculum to meet the needs of students today. This process will include a look at both content and pedagogy—the <i>what</i> and the <i>how</i>.</p><p>Before moving further, however, I must emphasize that faculty themselves must become conversant with the basics of using generative AI and LLMs. Your subdiscipline's tools may vary, but at a minimum, anyone can start with ChatGPT, Bard, or Bing (all of which have a free version) to get familiar with the technology. Most colleges and universities with a center for teaching and learning have training resources available, including opportunities for coaching and mentoring, and articles about the integration of AI within and beyond their organization,40 and a quick web search for articles about “using ChatGPT in higher education” will yield thousands of results.</p><p>Students can use AI in a variety of ways. It can be used for light text editing (as in Grammarly63 or MS Word), idea generation, drafting entire essays, providing personal tutoring, and brainstorming, among others. Given the variety of ways the AI can be deployed, faculty should communicate clearly to students what uses are allowed in their course and what are not. “Don't use AI” is really not an option, unless you are planning to require that students use a typewriter or pen and paper for all assignments. AI is ubiquitous in web searching, word processing, spreadsheet generation, and presentation software. In the near future, Microsoft will offer an LLM called Co-Pilot embedded in all Office 365 applications, including Word, Excel, and PowerPoint.64 As generative AI becomes integrated into every application, it will be ludicrous to suggest students cannot use it at all. Rather, faculty should specify what they will allow, when, and how.</p><p>To start this process, I offer a tool that I developed for use at the OSU College of Business, which has been adopted generally at OSU: a list of AI icons with associated language that faculty can insert into their syllabi and assignment descriptions.65 This list is intended as a simple guide for faculty to conceive of how AI is likely to be used, both at the course and assignment levels. It does not do the hard work of determining where to allow the use of AI, but it does provide a starting point for faculty feeling overwhelmed or unsure of how the technology might be employed.</p><p>Another key question for faculty to tackle is how they might describe generally in their syllabus how to describe the uses of generative AI that they will allow. I recommend that colleges and universities develop a shared language to facilitate student understanding of permitted uses. In footnote 66, I have linked sample AI statements that faculty can incorporate in their syllabi or in an individual assignment. These statements, which are linked to icons, describe different levels of AI approved for use in an overall course or specific uses that might be approved for an individual assignment.66 This document also includes a sample policy for appropriate citation format for students who want to use language they have imported directly from ChatGPT or another generative AI tool.</p><p>Changes are coming so fast that faculty can lose sight of one key thing: Students want to learn. Students know if they complete a course without gaining any skills, then they are not prepared for future success. Students invest heavily in their education, and they are intent on learning relevant skills they can take into their future. Faculty and students have a shared interest in a relevant and engaging curriculum. Make it a priority to talk to your students about how they are using generative AI, either for school, or work, or in their professional development. Ask them what type of assignments they learn from. Ask them about their interests and find ways to engage them. Talk to them about the importance of your course and how it will apply to their lives. Assume they want to learn and find ways to support that learning journey.</p><p>Generative AI is a tool that humans have created, and its limitations—including its potential to reinforce existing bias and its inability to provide real-time advice in a constantly changing business environment—must be considered. Educators are in a unique position to rethink what we teach and how we teach it in a way that recognizes the coming ubiquity of generative AI but also moves to the forefront of uniquely human skills. Focusing on the essential knowledge that students need, alongside the critical thinking skills they must obtain, is the path forward to a more relevant, engaging, and rewarding pedagogy.</p>","PeriodicalId":42278,"journal":{"name":"Journal of Legal Studies Education","volume":"41 1","pages":"29-37"},"PeriodicalIF":0.5000,"publicationDate":"2024-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/jlse.12141","citationCount":"0","resultStr":"{\"title\":\"Rising to Meet the Challenge of Generative AI\",\"authors\":\"Inara Scott\",\"doi\":\"10.1111/jlse.12141\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>ChatGPT launched in November 2022 and quickly became the fastest-growing user application in history, marking one million users in two months—a milestone that took TikTok nine months to achieve and Instagram two and a half years.1 That explosive growth has come with an explosion of concern for the ability of scientists and regulators to understand what it is, how it works,2 and its potential to change life as we know it. Politicians and technology executives alike are calling for more and better regulation to ensure apocalyptic scenarios of artificial intelligence (AI)-aided disasters (everything from AI-created weapons to sentient AI systems) do not come to pass.3 Meanwhile, the practice of law is dealing with the implications of a tool that can pass the Bar exam4 and colleges and universities are grappling with the reality that students can use generative AI to complete just about any assignment they can give.5</p><p>This article is not intended to “solve” the problem of generative AI. Rather, recognizing the astonishing pace of development of generative AI tools and their impact on business law and other higher education classes, it seeks to provide specific, concrete steps that faculty can take to evolve alongside these tools. There is no way to “AI-proof” your classes. However, taking the steps outlined here can help you decide <i>what</i> you want to teach and <i>how</i> you should teach it. It offers a structure for identifying the content you want to keep and what you can let go of and tips for redesigning assignments and syllabi to clarify your approach to students and reduce academic misconduct.</p><p>To understand the profound impact of generative AI and tools like ChatGPT, it is helpful to begin by unpacking some of the language that is used in this field.6 AI, short for “artificial intelligence,” generally refers to the use of machines, particularly computers, to perform tasks that simulate the use of human thinking, reasoning, and learning.7 We encounter AI throughout our day, when using computer applications like a Google search, interacting with a chatbot on a consumer website, or using a virtual assistant like Siri or Alexa. The ubiquity of AI in our daily lives is predicted to increase.8 In the near term, AI's use will likely become more ubiquitous as it becomes embedded in the way we interact with everyday items like our cars, office computers, and coffee machines.9</p><p>The term artificial “intelligence” is controversial because machines do not actually have the capacity to think or learn like humans. Their programming can <i>simulate</i> some aspects of human intelligence, but they do not reason like a human.10 For this reason, scientists distinguish between “strong AI” or “artificial general intelligence” and “weak AI.”11 Weak AI is what we generally have in use today—computers simulating human intelligence while they complete a specific type of task they have been programmed to perform. Strong AI is a theoretical system that thinks, reasons, and has self-awareness like a human (and could evolve to exceed human intelligence—thus prompting doomsday “AI takes over the world” scenarios).</p><p>Machine learning is a type of AI that programs systems to “learn” and improve their ability to perform assigned tasks without explicit programming.12 Machine learning relies on algorithms that allow machines to sort vast quantities of data and use that data to make predictions and find patterns, ultimately teaching themselves through repetition and human intervention, while human programmers tweak the algorithms that guide the learning process.13 The most powerful and complex AI applications today use <i>neural networks</i>—a type of machine learning that simulates the way neurons in the human brain communicate information.14</p><p>Generative AI refers to the use of AI applications to create something new, including images, text, and other forms of media. Large language models (LLMs) use machine learning and highly sophisticated neural networks to process and generate text.15 LLMs are trained on enormous datasets with billions of inputs—imagine starting by dumping all of the Internet into a giant bucket and you get the idea. Somewhat disturbingly, the contents of those databases are not transparent,16 and litigation has alleged that the databases likely contain a significant amount of copyrighted material.17 The LLM uses this dataset to process inputs (i.e., human prompts to the LLM interface) and predict the most helpful, common, and relevant responses.18 In effect, the LLM crowdsources answers to the questions it is asked, by using its skills in pattern identification and prediction to generate the most likely and useful textual responses given the information in its database. The tools we know as ChatGPT, Bard, or Bing are simply interfaces using LLMs to produce responses to prompts from external users.</p><p>Generative AI's flaws are predictable, given how the technology works. Though as humans, we are likely to anthropomorphize LLMs and attribute them with human characteristics,19 they are not thinking beings—they are simply computer models with predictable flaws arising out of an astonishingly complex but ultimately nonhuman system.</p><p>One of the most glaring flaws of LLMs is their potential to produce inaccurate, false, or flawed responses, in which the model puts together strings of text that do not point to accurate information but mimic accurate information in a compelling way. For example, when prompted for a peer-reviewed journal article, an LLM might string together a series of notations that look like a journal article, with a title, volume, issue, and page numbers but which does not point to an existing source. LLMs can create a variety of types of false information, such as false journal articles, false meetings of historical figures, and false medical information.20 Early in the use of LLMs, these responses were dubbed “hallucinations.”21 Notably, even that term is anthropomorphically loaded—it suggests the LLM is operating in a semi-conscious human way, “imagining” events that did not occur, rather than simply acting as a flawed computer model producing inaccurate results.</p><p>Another essential limitation of LLMs is also rooted in their technological structure. LLMs are, by their nature, predicting and generating text that aligns with information already in their database. The information in their database was generally created by humans—flawed and biased humans, producing information that is also flawed and biased, operating in societies with existing structural racism and bias. If faced with a question like “please list the top twenty jazz pianists,” the LLM does not independently assess the work of all living and dead jazz pianists and reach a conclusion about their relative merit. Rather, it crowdsources information, identifying patterns in commonly reported responses to this question and developing a prediction based on that information. LLMs cannot account for historical inequities and biases in reporting on artists, let alone the biases that go into the development of artists and their access to recording contracts or critical acclaim.22 In an area like jazz music, with deep historical roots of inequities in the treatment of men and women,23 one cannot expect an LLM to do anything other than repeat existing biases, which are reflected in popular source materials. As a result, when queried, ChatGPT, crowdsourcing from existing sources, reported no women in the top 20 jazz pianists.24 ChatGPT produces this result, despite the historical significance of figures like Mary Lou Williams, Alice Coltrane, and Shirley Scott, iconic female artists well-known to jazz scholars and aficionados alike.25</p><p>DALL-E, a generative AI tool that creates images based on textual prompts, produces similarly biased responses. When prompted to create a “realistic photograph quality picture of a college professor talking to a large group of students,” DALL-E created the following image, of a white male professor surrounded by white students Figure 1:</p><p>Is this image “incorrect”? Survey data from 2019 suggests over seventy-five percent of professors are white, while fifty-five percent of undergraduates are white.26 Data from 2021 show white men continue to represent the largest percentage of tenure-track faculty.27 Given this landscape, it is entirely consistent with AI's programming to generate images that center white men as professors and to leave women out of a list of top jazz pianists.</p><p>Of course, the answer begets a closer look at the question. If we are asking which jazz musicians have been most widely recognized or what the “average” college professor looks like, the tools are performing as programmed. But there are many other questions embedded in this call and response, and they get at the heart of what it means to use generative AI in education. If the average user of an LLM assumes that this list of jazz pianists <i>accurately</i> represents a list of the <i>most talented</i> jazz pianists ever to have lived, the technology is reinforcing inaccurate information based on existing societal biases. To be clear, the LLM cannot and is not answering that question. The LLM, in this scenario, is also not considering the question of what cultural or legal barriers might prevent or have prevented women from becoming known in the field of jazz music (though it could, if directly asked).28 It is simply reproducing a list of commonly mentioned “top” jazz pianists.</p><p>Similarly, DALL-E is not “incorrect” when it creates a likely image of a professor as white and male. The problem here is that research suggests the images we view shape our ideas of what is possible.29 Thus, if popular media images are dominated by the “most likely” outcomes generated by AI models, stereotypical roles will be harder to reject and move past. Generative AI can emphasize and blind us to the way our cultural biases show up in everything from the medical literature to our notions of beauty because they are, by their programming, designed to reflect a majoritarian view of the world.</p><p>To be clear, DALL-E and ChatGPT are not different in this action than existing search engines or media companies, which are also built on crowdsourcing and majoritarian views. The challenge is that, as these tools become ubiquitous and embedded in our daily lives, we may come to view them as definitive sources of <i>truth</i>, even more so than we currently do with search engines. When a student asks ChatGPT a medical question and receives a conversational response that mimics a medical professional, they may assume the answer they are receiving is accurate for all populations rather than understanding that the information is likely based on a medical study of young, white men and may not be representative of the population as a whole.30</p><p>These are two of the most prominent concerns about generative AI, but there are other flaws worth noting. Generative AI cannot replace the need for professionals in real life to “think on their feet” and react quickly in scenarios that require human empathy and communication skills. It cannot respond in real time to complex scenarios with constantly changing fact patterns and complicated human actors. It bases its responses on existing scenarios and patterns, so it is unlikely to make truly innovative, creative leaps. Because it lacks human intuition and feeling, it also cannot look below the surface of a question to respond to a deeper, more essential question, or even draw out the question that is not asked. As a consultant might say, “AI can only give the client what they ask for—so my job is secure.” Implied: We often do not know what we want, in business or in life, and we rely on the insight from the humans around us to help identify our underlying needs.</p><p>One approach or response to the wave of generative AI would be to plant a flag in the ground and refuse to move or evolve. Rather than changing the content of their courses, in this scenario, faculty might argue that universities and colleges should instead institute more stringent academic integrity policies and punishments and develop better ways to “catch the cheaters.” This position is fundamentally flawed, for two primary reasons: (1) faculty will never be able to “AI-proof” their courses, and (2) learning the responsible use of AI is essential to students’ future success.</p><p>First, while creating an environment that supports academic integrity is important, some assignments and assessments will always be able to evade restrictions on the use of AI. While many faculty have reasonably decided to turn toward more in-class assignments and assessments to reduce cheating, this option is not available to those teaching online. Moreover, not all assignments can be completed during class hours.</p><p>What about developing more and better ways to catch the cheaters? That is, can we not just proctor our way out of this situation? Even keeping in mind that online proctoring has limited application and can only be used for time-limited assessments, the reality is that many universities had begun moving away from AI-enabled proctoring <i>before</i> ChatGPT came on the scene due to concerns about privacy, discrimination, and unreliability.31</p><p>What about AI-text detectors like ZeroGPT and GPTzero?32 Can we not use AI to detect AI? Unfortunately, these tools have been shown to be unreliable33 and biased against non-native English writers.34 In addition, work-arounds are notoriously easy to find, most commonly (and ironically) by using AI paraphrasing tools.35</p><p>Faculty can create assignments that are <i>easier</i> to complete with generative AI and those that are <i>harder</i>. In most cases, however, particularly in large undergraduate survey courses, generative AI will be able to complete most assignments, particularly those completed outside of class. If ChatGPT can pass the Bar exam, it can probably draft a better legal essay than most undergraduate students. It certainly can undertake any multiple-choice questions a faculty member might write.</p><p>This means students will have to decide to complete their assignments knowing that they have the answer key (via ChatGPT) and knowing faculty probably cannot catch them cheating. Students inclined to cheat (for any number of reasons, which might include the struggle of balancing school with full-time work,36 caring for family members,37 or dealing with other challenging life circumstances like food insecurity38 or physical or mental illness39) now have a straightforward way to successfully complete their assignments. Students who do try to complete their assignments without assistance may as a result be at a disadvantage, particularly in classes using curved grading, and faculty would not be able to tell one group from the other. This could lead to grading inequities, not to mention even greater incentives for students who would not normally bend the rules, if only to be able to stand on a level playing field.</p><p>Second, and I believe even more importantly, what faculty teach <i>should have relevance</i> to the way their students will undertake their future work. For those teaching undergraduate business law courses at a business school, students are unlikely to become lawyers. Rather, they will most likely be applying legal content to business situations. They need to learn how to apply the information they are learning to these real-world scenarios. Because their real world will likely include generative AI, learning how to use it safely, responsibly, and effectively to identify legal risks and strategic legal opportunities and how to fill in the known gaps and defects of the technology, particularly as relates to bias and inaccuracies.</p><p>In my experience, faculty want to evolve but are not sure how. Even knowing that students will be likely to use AI tools in their future work, many faculty remain convinced that there is some basic, fundamental information that students need to use the technology responsibly and to avoid the limitations described above. At the same time, faculty are unsure how to integrate this technology into their courses while retaining the essential information students need and focusing on the human-thinking skills that generative AI cannot provide. Finally, many are aware of the limitations of generative AI discussed above and want to avoid using the tool in a way that furthers biases or inhibits creative and critical thinking.</p><p>In this part, I begin to address these challenges. I discuss how to start using generative AI and how to adapt the curriculum to meet the needs of students today. This process will include a look at both content and pedagogy—the <i>what</i> and the <i>how</i>.</p><p>Before moving further, however, I must emphasize that faculty themselves must become conversant with the basics of using generative AI and LLMs. Your subdiscipline's tools may vary, but at a minimum, anyone can start with ChatGPT, Bard, or Bing (all of which have a free version) to get familiar with the technology. Most colleges and universities with a center for teaching and learning have training resources available, including opportunities for coaching and mentoring, and articles about the integration of AI within and beyond their organization,40 and a quick web search for articles about “using ChatGPT in higher education” will yield thousands of results.</p><p>Students can use AI in a variety of ways. It can be used for light text editing (as in Grammarly63 or MS Word), idea generation, drafting entire essays, providing personal tutoring, and brainstorming, among others. Given the variety of ways the AI can be deployed, faculty should communicate clearly to students what uses are allowed in their course and what are not. “Don't use AI” is really not an option, unless you are planning to require that students use a typewriter or pen and paper for all assignments. AI is ubiquitous in web searching, word processing, spreadsheet generation, and presentation software. In the near future, Microsoft will offer an LLM called Co-Pilot embedded in all Office 365 applications, including Word, Excel, and PowerPoint.64 As generative AI becomes integrated into every application, it will be ludicrous to suggest students cannot use it at all. Rather, faculty should specify what they will allow, when, and how.</p><p>To start this process, I offer a tool that I developed for use at the OSU College of Business, which has been adopted generally at OSU: a list of AI icons with associated language that faculty can insert into their syllabi and assignment descriptions.65 This list is intended as a simple guide for faculty to conceive of how AI is likely to be used, both at the course and assignment levels. It does not do the hard work of determining where to allow the use of AI, but it does provide a starting point for faculty feeling overwhelmed or unsure of how the technology might be employed.</p><p>Another key question for faculty to tackle is how they might describe generally in their syllabus how to describe the uses of generative AI that they will allow. I recommend that colleges and universities develop a shared language to facilitate student understanding of permitted uses. In footnote 66, I have linked sample AI statements that faculty can incorporate in their syllabi or in an individual assignment. These statements, which are linked to icons, describe different levels of AI approved for use in an overall course or specific uses that might be approved for an individual assignment.66 This document also includes a sample policy for appropriate citation format for students who want to use language they have imported directly from ChatGPT or another generative AI tool.</p><p>Changes are coming so fast that faculty can lose sight of one key thing: Students want to learn. Students know if they complete a course without gaining any skills, then they are not prepared for future success. Students invest heavily in their education, and they are intent on learning relevant skills they can take into their future. Faculty and students have a shared interest in a relevant and engaging curriculum. Make it a priority to talk to your students about how they are using generative AI, either for school, or work, or in their professional development. Ask them what type of assignments they learn from. Ask them about their interests and find ways to engage them. Talk to them about the importance of your course and how it will apply to their lives. Assume they want to learn and find ways to support that learning journey.</p><p>Generative AI is a tool that humans have created, and its limitations—including its potential to reinforce existing bias and its inability to provide real-time advice in a constantly changing business environment—must be considered. Educators are in a unique position to rethink what we teach and how we teach it in a way that recognizes the coming ubiquity of generative AI but also moves to the forefront of uniquely human skills. Focusing on the essential knowledge that students need, alongside the critical thinking skills they must obtain, is the path forward to a more relevant, engaging, and rewarding pedagogy.</p>\",\"PeriodicalId\":42278,\"journal\":{\"name\":\"Journal of Legal Studies Education\",\"volume\":\"41 1\",\"pages\":\"29-37\"},\"PeriodicalIF\":0.5000,\"publicationDate\":\"2024-02-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1111/jlse.12141\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Legal Studies Education\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1111/jlse.12141\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"LAW\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Legal Studies Education","FirstCategoryId":"1085","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/jlse.12141","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"LAW","Score":null,"Total":0}
ChatGPT launched in November 2022 and quickly became the fastest-growing user application in history, marking one million users in two months—a milestone that took TikTok nine months to achieve and Instagram two and a half years.1 That explosive growth has come with an explosion of concern for the ability of scientists and regulators to understand what it is, how it works,2 and its potential to change life as we know it. Politicians and technology executives alike are calling for more and better regulation to ensure apocalyptic scenarios of artificial intelligence (AI)-aided disasters (everything from AI-created weapons to sentient AI systems) do not come to pass.3 Meanwhile, the practice of law is dealing with the implications of a tool that can pass the Bar exam4 and colleges and universities are grappling with the reality that students can use generative AI to complete just about any assignment they can give.5
This article is not intended to “solve” the problem of generative AI. Rather, recognizing the astonishing pace of development of generative AI tools and their impact on business law and other higher education classes, it seeks to provide specific, concrete steps that faculty can take to evolve alongside these tools. There is no way to “AI-proof” your classes. However, taking the steps outlined here can help you decide what you want to teach and how you should teach it. It offers a structure for identifying the content you want to keep and what you can let go of and tips for redesigning assignments and syllabi to clarify your approach to students and reduce academic misconduct.
To understand the profound impact of generative AI and tools like ChatGPT, it is helpful to begin by unpacking some of the language that is used in this field.6 AI, short for “artificial intelligence,” generally refers to the use of machines, particularly computers, to perform tasks that simulate the use of human thinking, reasoning, and learning.7 We encounter AI throughout our day, when using computer applications like a Google search, interacting with a chatbot on a consumer website, or using a virtual assistant like Siri or Alexa. The ubiquity of AI in our daily lives is predicted to increase.8 In the near term, AI's use will likely become more ubiquitous as it becomes embedded in the way we interact with everyday items like our cars, office computers, and coffee machines.9
The term artificial “intelligence” is controversial because machines do not actually have the capacity to think or learn like humans. Their programming can simulate some aspects of human intelligence, but they do not reason like a human.10 For this reason, scientists distinguish between “strong AI” or “artificial general intelligence” and “weak AI.”11 Weak AI is what we generally have in use today—computers simulating human intelligence while they complete a specific type of task they have been programmed to perform. Strong AI is a theoretical system that thinks, reasons, and has self-awareness like a human (and could evolve to exceed human intelligence—thus prompting doomsday “AI takes over the world” scenarios).
Machine learning is a type of AI that programs systems to “learn” and improve their ability to perform assigned tasks without explicit programming.12 Machine learning relies on algorithms that allow machines to sort vast quantities of data and use that data to make predictions and find patterns, ultimately teaching themselves through repetition and human intervention, while human programmers tweak the algorithms that guide the learning process.13 The most powerful and complex AI applications today use neural networks—a type of machine learning that simulates the way neurons in the human brain communicate information.14
Generative AI refers to the use of AI applications to create something new, including images, text, and other forms of media. Large language models (LLMs) use machine learning and highly sophisticated neural networks to process and generate text.15 LLMs are trained on enormous datasets with billions of inputs—imagine starting by dumping all of the Internet into a giant bucket and you get the idea. Somewhat disturbingly, the contents of those databases are not transparent,16 and litigation has alleged that the databases likely contain a significant amount of copyrighted material.17 The LLM uses this dataset to process inputs (i.e., human prompts to the LLM interface) and predict the most helpful, common, and relevant responses.18 In effect, the LLM crowdsources answers to the questions it is asked, by using its skills in pattern identification and prediction to generate the most likely and useful textual responses given the information in its database. The tools we know as ChatGPT, Bard, or Bing are simply interfaces using LLMs to produce responses to prompts from external users.
Generative AI's flaws are predictable, given how the technology works. Though as humans, we are likely to anthropomorphize LLMs and attribute them with human characteristics,19 they are not thinking beings—they are simply computer models with predictable flaws arising out of an astonishingly complex but ultimately nonhuman system.
One of the most glaring flaws of LLMs is their potential to produce inaccurate, false, or flawed responses, in which the model puts together strings of text that do not point to accurate information but mimic accurate information in a compelling way. For example, when prompted for a peer-reviewed journal article, an LLM might string together a series of notations that look like a journal article, with a title, volume, issue, and page numbers but which does not point to an existing source. LLMs can create a variety of types of false information, such as false journal articles, false meetings of historical figures, and false medical information.20 Early in the use of LLMs, these responses were dubbed “hallucinations.”21 Notably, even that term is anthropomorphically loaded—it suggests the LLM is operating in a semi-conscious human way, “imagining” events that did not occur, rather than simply acting as a flawed computer model producing inaccurate results.
Another essential limitation of LLMs is also rooted in their technological structure. LLMs are, by their nature, predicting and generating text that aligns with information already in their database. The information in their database was generally created by humans—flawed and biased humans, producing information that is also flawed and biased, operating in societies with existing structural racism and bias. If faced with a question like “please list the top twenty jazz pianists,” the LLM does not independently assess the work of all living and dead jazz pianists and reach a conclusion about their relative merit. Rather, it crowdsources information, identifying patterns in commonly reported responses to this question and developing a prediction based on that information. LLMs cannot account for historical inequities and biases in reporting on artists, let alone the biases that go into the development of artists and their access to recording contracts or critical acclaim.22 In an area like jazz music, with deep historical roots of inequities in the treatment of men and women,23 one cannot expect an LLM to do anything other than repeat existing biases, which are reflected in popular source materials. As a result, when queried, ChatGPT, crowdsourcing from existing sources, reported no women in the top 20 jazz pianists.24 ChatGPT produces this result, despite the historical significance of figures like Mary Lou Williams, Alice Coltrane, and Shirley Scott, iconic female artists well-known to jazz scholars and aficionados alike.25
DALL-E, a generative AI tool that creates images based on textual prompts, produces similarly biased responses. When prompted to create a “realistic photograph quality picture of a college professor talking to a large group of students,” DALL-E created the following image, of a white male professor surrounded by white students Figure 1:
Is this image “incorrect”? Survey data from 2019 suggests over seventy-five percent of professors are white, while fifty-five percent of undergraduates are white.26 Data from 2021 show white men continue to represent the largest percentage of tenure-track faculty.27 Given this landscape, it is entirely consistent with AI's programming to generate images that center white men as professors and to leave women out of a list of top jazz pianists.
Of course, the answer begets a closer look at the question. If we are asking which jazz musicians have been most widely recognized or what the “average” college professor looks like, the tools are performing as programmed. But there are many other questions embedded in this call and response, and they get at the heart of what it means to use generative AI in education. If the average user of an LLM assumes that this list of jazz pianists accurately represents a list of the most talented jazz pianists ever to have lived, the technology is reinforcing inaccurate information based on existing societal biases. To be clear, the LLM cannot and is not answering that question. The LLM, in this scenario, is also not considering the question of what cultural or legal barriers might prevent or have prevented women from becoming known in the field of jazz music (though it could, if directly asked).28 It is simply reproducing a list of commonly mentioned “top” jazz pianists.
Similarly, DALL-E is not “incorrect” when it creates a likely image of a professor as white and male. The problem here is that research suggests the images we view shape our ideas of what is possible.29 Thus, if popular media images are dominated by the “most likely” outcomes generated by AI models, stereotypical roles will be harder to reject and move past. Generative AI can emphasize and blind us to the way our cultural biases show up in everything from the medical literature to our notions of beauty because they are, by their programming, designed to reflect a majoritarian view of the world.
To be clear, DALL-E and ChatGPT are not different in this action than existing search engines or media companies, which are also built on crowdsourcing and majoritarian views. The challenge is that, as these tools become ubiquitous and embedded in our daily lives, we may come to view them as definitive sources of truth, even more so than we currently do with search engines. When a student asks ChatGPT a medical question and receives a conversational response that mimics a medical professional, they may assume the answer they are receiving is accurate for all populations rather than understanding that the information is likely based on a medical study of young, white men and may not be representative of the population as a whole.30
These are two of the most prominent concerns about generative AI, but there are other flaws worth noting. Generative AI cannot replace the need for professionals in real life to “think on their feet” and react quickly in scenarios that require human empathy and communication skills. It cannot respond in real time to complex scenarios with constantly changing fact patterns and complicated human actors. It bases its responses on existing scenarios and patterns, so it is unlikely to make truly innovative, creative leaps. Because it lacks human intuition and feeling, it also cannot look below the surface of a question to respond to a deeper, more essential question, or even draw out the question that is not asked. As a consultant might say, “AI can only give the client what they ask for—so my job is secure.” Implied: We often do not know what we want, in business or in life, and we rely on the insight from the humans around us to help identify our underlying needs.
One approach or response to the wave of generative AI would be to plant a flag in the ground and refuse to move or evolve. Rather than changing the content of their courses, in this scenario, faculty might argue that universities and colleges should instead institute more stringent academic integrity policies and punishments and develop better ways to “catch the cheaters.” This position is fundamentally flawed, for two primary reasons: (1) faculty will never be able to “AI-proof” their courses, and (2) learning the responsible use of AI is essential to students’ future success.
First, while creating an environment that supports academic integrity is important, some assignments and assessments will always be able to evade restrictions on the use of AI. While many faculty have reasonably decided to turn toward more in-class assignments and assessments to reduce cheating, this option is not available to those teaching online. Moreover, not all assignments can be completed during class hours.
What about developing more and better ways to catch the cheaters? That is, can we not just proctor our way out of this situation? Even keeping in mind that online proctoring has limited application and can only be used for time-limited assessments, the reality is that many universities had begun moving away from AI-enabled proctoring before ChatGPT came on the scene due to concerns about privacy, discrimination, and unreliability.31
What about AI-text detectors like ZeroGPT and GPTzero?32 Can we not use AI to detect AI? Unfortunately, these tools have been shown to be unreliable33 and biased against non-native English writers.34 In addition, work-arounds are notoriously easy to find, most commonly (and ironically) by using AI paraphrasing tools.35
Faculty can create assignments that are easier to complete with generative AI and those that are harder. In most cases, however, particularly in large undergraduate survey courses, generative AI will be able to complete most assignments, particularly those completed outside of class. If ChatGPT can pass the Bar exam, it can probably draft a better legal essay than most undergraduate students. It certainly can undertake any multiple-choice questions a faculty member might write.
This means students will have to decide to complete their assignments knowing that they have the answer key (via ChatGPT) and knowing faculty probably cannot catch them cheating. Students inclined to cheat (for any number of reasons, which might include the struggle of balancing school with full-time work,36 caring for family members,37 or dealing with other challenging life circumstances like food insecurity38 or physical or mental illness39) now have a straightforward way to successfully complete their assignments. Students who do try to complete their assignments without assistance may as a result be at a disadvantage, particularly in classes using curved grading, and faculty would not be able to tell one group from the other. This could lead to grading inequities, not to mention even greater incentives for students who would not normally bend the rules, if only to be able to stand on a level playing field.
Second, and I believe even more importantly, what faculty teach should have relevance to the way their students will undertake their future work. For those teaching undergraduate business law courses at a business school, students are unlikely to become lawyers. Rather, they will most likely be applying legal content to business situations. They need to learn how to apply the information they are learning to these real-world scenarios. Because their real world will likely include generative AI, learning how to use it safely, responsibly, and effectively to identify legal risks and strategic legal opportunities and how to fill in the known gaps and defects of the technology, particularly as relates to bias and inaccuracies.
In my experience, faculty want to evolve but are not sure how. Even knowing that students will be likely to use AI tools in their future work, many faculty remain convinced that there is some basic, fundamental information that students need to use the technology responsibly and to avoid the limitations described above. At the same time, faculty are unsure how to integrate this technology into their courses while retaining the essential information students need and focusing on the human-thinking skills that generative AI cannot provide. Finally, many are aware of the limitations of generative AI discussed above and want to avoid using the tool in a way that furthers biases or inhibits creative and critical thinking.
In this part, I begin to address these challenges. I discuss how to start using generative AI and how to adapt the curriculum to meet the needs of students today. This process will include a look at both content and pedagogy—the what and the how.
Before moving further, however, I must emphasize that faculty themselves must become conversant with the basics of using generative AI and LLMs. Your subdiscipline's tools may vary, but at a minimum, anyone can start with ChatGPT, Bard, or Bing (all of which have a free version) to get familiar with the technology. Most colleges and universities with a center for teaching and learning have training resources available, including opportunities for coaching and mentoring, and articles about the integration of AI within and beyond their organization,40 and a quick web search for articles about “using ChatGPT in higher education” will yield thousands of results.
Students can use AI in a variety of ways. It can be used for light text editing (as in Grammarly63 or MS Word), idea generation, drafting entire essays, providing personal tutoring, and brainstorming, among others. Given the variety of ways the AI can be deployed, faculty should communicate clearly to students what uses are allowed in their course and what are not. “Don't use AI” is really not an option, unless you are planning to require that students use a typewriter or pen and paper for all assignments. AI is ubiquitous in web searching, word processing, spreadsheet generation, and presentation software. In the near future, Microsoft will offer an LLM called Co-Pilot embedded in all Office 365 applications, including Word, Excel, and PowerPoint.64 As generative AI becomes integrated into every application, it will be ludicrous to suggest students cannot use it at all. Rather, faculty should specify what they will allow, when, and how.
To start this process, I offer a tool that I developed for use at the OSU College of Business, which has been adopted generally at OSU: a list of AI icons with associated language that faculty can insert into their syllabi and assignment descriptions.65 This list is intended as a simple guide for faculty to conceive of how AI is likely to be used, both at the course and assignment levels. It does not do the hard work of determining where to allow the use of AI, but it does provide a starting point for faculty feeling overwhelmed or unsure of how the technology might be employed.
Another key question for faculty to tackle is how they might describe generally in their syllabus how to describe the uses of generative AI that they will allow. I recommend that colleges and universities develop a shared language to facilitate student understanding of permitted uses. In footnote 66, I have linked sample AI statements that faculty can incorporate in their syllabi or in an individual assignment. These statements, which are linked to icons, describe different levels of AI approved for use in an overall course or specific uses that might be approved for an individual assignment.66 This document also includes a sample policy for appropriate citation format for students who want to use language they have imported directly from ChatGPT or another generative AI tool.
Changes are coming so fast that faculty can lose sight of one key thing: Students want to learn. Students know if they complete a course without gaining any skills, then they are not prepared for future success. Students invest heavily in their education, and they are intent on learning relevant skills they can take into their future. Faculty and students have a shared interest in a relevant and engaging curriculum. Make it a priority to talk to your students about how they are using generative AI, either for school, or work, or in their professional development. Ask them what type of assignments they learn from. Ask them about their interests and find ways to engage them. Talk to them about the importance of your course and how it will apply to their lives. Assume they want to learn and find ways to support that learning journey.
Generative AI is a tool that humans have created, and its limitations—including its potential to reinforce existing bias and its inability to provide real-time advice in a constantly changing business environment—must be considered. Educators are in a unique position to rethink what we teach and how we teach it in a way that recognizes the coming ubiquity of generative AI but also moves to the forefront of uniquely human skills. Focusing on the essential knowledge that students need, alongside the critical thinking skills they must obtain, is the path forward to a more relevant, engaging, and rewarding pedagogy.