...
...
...
...
...
...
...
...
...
...
...
...
An Updated version of this guide is published at the Canisius University AI and Society Initiative Site. This version is not updated after June 5th, 2023 |
---|
Here, COLI is assembling a list of introductory information, tips, and resources concerning artificial intelligence, and it's possible implications for pedagogy and scholarship. This resource is tentative, incomplete, might age rapidly, and probably (at any point in time) a little wrong in places.
...
Google Bard As an AI, how are you classified? "I am classified as a large language model, also known as a conversational AI or chatbot trained to be informative and comprehensive. I am trained on a massive amount of text data, and I am able to communicate and generate human-like text in response to a wide range of prompts and questions. For example, I can provide summaries of factual topics or create stories." Can you describe your training corpus? "Sure. My training corpus is a massive dataset of text and code that I was trained on. It includes text from a variety of sources, such as books, articles, websites, and code. The corpus is designed to teach me about the structure of language and how to use language to communicate effectively. The corpus is also designed to teach me about the world. It includes text about a variety of topics, such as science, history, and current events. This helps me to understand the world around me and to generate text that is relevant and informative. The corpus is constantly being updated with new text. This helps me to stay up-to-date on the latest information and to generate text that is accurate and up-to-date." |
Any successful LLM AI production requires a carefully written prompt from the user. This must properly describe the user's intent, and if the AI's response is not to be dependent on its training corpus, it the user's prompt must supply relevant data as well. Already, education and technology experts speak of "prompt engineering" as a probably probable skill necessary for students in the future, but it remains to be seen whether that is true, as these tools (more or less) evolve.
...
- descriptions of a book whose text or detailed summaries of the same are not in the AI's training data. The AI might develop a plausible but false interpretation or summary based on the book's title, or what information it may have on the book's subject. In COLI, we find that copyrighted works seem particularly troublesome in this respect. It is likely that such works, and text descriptions of those works, have not made it into LLM AI training data.
- scientific or engineering explanations of complex phenomena. One example is rocketry.
- biographies of non-famous individuals. Try asking for a short biography of you and your title, if it is already publicly available on the web. For example, ChatGPT had this to say about COLI Director Dr. Mark Gallimore: "Gallimore holds a Bachelor of Arts degree in English from St. Bonaventure University and a Master of Science degree in Education from Canisius College. He began his career in education as a high school teacher, teaching English and social studies in Buffalo, New York." While he has the highest respect for the aforementioned institutions and high school teachers, this biography of Dr. Gallimore is in every respect false.
We may say that LLM AIs "making stuff up" or "get it wrong," but they are not malfunctioning. They simulate human composition. Ars Technica's Benj Edwards offers a good explanation of this. But the short version was offered by OpenAI CEO Sam Altman, soon after ChatGPT became available in late 2022: "ChatGPT is incredibly limited, but good enough at some things to create a misleading impression of greatness. It's a mistake to be relying on it for anything important right now. It’s a preview of progress; we have lots of work to do on robustness and truthfulness."
...
This tentative guide is concerned with Large Language Model AIs, which are (at the time this is written) the most powerful artificial text generation tools available. Other tools, be it a customer service chatbot, a grammar assistant, or even the non-playable antagonist and their henchmen in video games, can be considered artificial intelligence, although they generally follow much simpler routines than LLM AIs. However, in COLI we expect to see in the next several months and years LLM AI-powered tools and features appear across the internet, in software and on mobile devices. So this distinction may not be relevant over time.
Pedagogy
Each discipline will need to determine the extent to which LLM AIs compel adaptation or alteration of their curriculum. Each faculty member will also likely need to consider the relationship between their teaching style and methods, and LLM AIs. However, here is a list of things that might spur some inspiration.
...
We might use AI to write boilerplate text that professionals use everyday at their work. We might permit students to do the same. We might also ask students to analyze the boilerplate responses generated by AIs, to look for cultural context: in what ways do AIs write a thank you letter, or job application cover letter, that reflects certain social, economic, or cultural status?
Source Analysis
LLM AIs might provide handy source content on which students can practice critical thinking skills. Can students spot certain arguments or descriptions that have political implications, and so are not as "objective" as perhaps the AI's tone might suggest? Can students spot errors or falsehoods? Can students employ web literacy and fact-checking skills to assess veracity or just cultural nuances within an AI's version of a story?
Prompt Engineering
At least in the present, how the user crafts a question or command for an LLM AI, prompt engineering, determines the AI's product. Even subtle re-wording of the same (to humans) questions can produce radically different results from the AI. It seems that, for the time being, prompt engineering might be a useful skill to cultivate in students. What specific practices or procedures produce different kinds of outcomes may depend on different disciplines.
Sources
LLM AIs have learned primarily on open-sourced content. This might be on the internet, or books that are out of copyright. There may be You or your students may use AI to develop sample text for projects or processes. For example, could the AIs write simple scenarios or case studies that students can work through using skills or abilities learned in a course? AIs can produce data in columns or .csv, useful for learning statistics or data processing.
Fictional but Plausible Examples
LLM AIs are designed to simulate human beings, and so may be sources for learning simulations. For example, could ChatGPT write essays in the style of an adolescent student? It might provide teacher education students with examples on which to practice assessment and feedback skills. Can Bing Chat suggest problems for mathematics students to solve?
Source Analysis
LLM AIs might provide handy source content on which students can practice critical thinking skills. Can students spot certain arguments or descriptions that have political implications, and so are not as "objective" as perhaps the AI's tone might suggest? Can students spot errors or falsehoods? Can students employ web literacy and fact-checking skills to assess veracity or just cultural nuances within an AI's version of a story?
Prompt Engineering
At least in the present, how the user crafts a question or command for an LLM AI, prompt engineering, determines the AI's product. Even subtle re-wording of the same (to humans) questions can produce radically different results from the AI. It seems that, for the time being, prompt engineering might be a useful skill to cultivate in students. What specific practices or procedures produce different kinds of outcomes may depend on different disciplines.
Sources
LLM AIs have learned primarily on open-sourced content. This might be on the internet, or books that are out of copyright. There may be exceptions in unpublished training aids. But much of what we assign is copyrighted content, out of necessity, since that is where specialized disciplinary knowledge is found. Writing assignments that ask students to focus on these specialized resources will not be accessible to generative AIs.
...
Some of these are more or less close to the truth. Most suggest that ChatGPT made a reasonable but erroneous guess based on the book title and perhaps some other publicly available descriptions. Since the book has wide circulation and the earliest edition dates from the 1980s, this suggests that considerable quantities of long-form text, and especially copyrighted text, have not been included in the AI's training materials.
This example could be supplied to students (perhaps assigned the Millet, Maslowski, and Feis text) as a warning that the AI production cannot currently, at least, be trusted at face value for veracity.
With Your Students
When we say that AIs are likely a part of the professional future for many of our students, and we need to prepare them to work with or around AIs, what we might mean is simply making students aware of AI's current likely behaviors, and developing in students a habit of remaining aware of trends in AI. For example, if you are teaching economics or anthropology, you might periodically prompt AIs to discuss the day's class subject, content or activity, and then discuss with students how the AI responds.
Companies like OpenAI and Google are sensitive to charges that AIs inherit bias and discrimination present in their human creators or training data. They have taken steps to prevent this, but their products remain controversial. This might be an important conversation to have with students, based on various perspectives presented in sources. However, one should be cautious about in-class or assigned experimentation with, for example, engineering prompts designed to provoke racist replies.
Another concern can be that students must create accounts at OpenAI, Google, and Microsoft to experiment with these AIs. Asking students to provide these companies with personally identifiable information (PII) may be problematic. Students should be encouraged to consult and understand the terms of service, even if it is optional.
But classroom experimentation with AIs might be beneficial, to determine how forms of knowledge and ways of thinking in your discipline interact with AIs. Use a single account, and project it on the big screen in class. Work together as a class to generate or modify prompts. Students might see how AIs stumble with certain questions, or provide simulated but incorrect answers. If an AI cannot perform the kinds of analyses, creativity, or other skills you hope students learn in the course and in the process of assignments, it is good for students to see that for themselves while you are present to answer questions. In engineering prompts and discussing AI outcomes, you also have an opportunity to demonstrate ways of thinking, habits, practices, and procedures that are the substance of your course learning objectivesdates from the 1980s, this suggests that considerable quantities of long-form text, and especially copyrighted text, have not been included in the AI's training materials.
This example could be supplied to students (perhaps assigned the Millet, Maslowski, and Feis text) as a warning that the AI production cannot currently, at least, be trusted at face value for veracity.
With Your Students
When we say that AIs are likely a part of the professional future for many of our students, and we need to prepare them to work with or around AIs, what we might mean is simply making students aware of AI's current likely behaviors, and developing in students a habit of remaining aware of trends in AI. For example, if you are teaching economics or anthropology, you might periodically prompt AIs to discuss the day's class subject, content or activity, and then discuss with students how the AI responds.
Companies like OpenAI and Google are sensitive to charges that AIs inherit bias and discrimination present in their human creators or training data. They have taken steps to prevent this, but their products remain controversial. This might be an important conversation to have with students, based on various perspectives presented in sources. However, one should be cautious about in-class or assigned experimentation with, for example, engineering prompts designed to provoke racist replies.
Another concern can be that students must create accounts at OpenAI, Google, and Microsoft to experiment with these AIs. Asking students to provide these companies with personally identifiable information (PII) may be problematic. Students should be encouraged to consult and understand the terms of service, even if it is optional.
But classroom experimentation with AIs might be beneficial, to determine how forms of knowledge and ways of thinking in your discipline interact with AIs. Use a single account, and project it on the big screen in class. Work together as a class to generate or modify prompts. Students might see how AIs stumble with certain questions, or provide simulated but incorrect answers. If an AI cannot perform the kinds of analyses, creativity, or other skills you hope students learn in the course and in the process of assignments, it is good for students to see that for themselves while you are present to answer questions. In engineering prompts and discussing AI outcomes, you also have an opportunity to demonstrate ways of thinking, habits, practices, and procedures that are the substance of your course learning objectives.
Lastly, it is reasonable to warn students that, apart from the serious moral, ethical, and social implications of academic dishonesty, there is the practical problem of a data footprint. Even if students take steps to anonymize their use of AIs (throwaway email accounts, private browser sessions, and so on) they might generate a digital trail of evidence. Under various scenarios, that evidence may become exposed or fall into the hands of extortionists, which could threaten their future careers and professional lives. This problem exists already with essay mills or contract cheating, and another form of it might appear with AIs, depending on whether AI providers change hands, or equip themselves with sufficient cybersecurity resources.
Ask AIs to Do Your Assignments
...
History
|
Biology
|
Management
|
To generate answers similar to these, here's the prompt: What are some assignments for a undergraduate university (discipline) course that have students practice or demonstrate things LLM AIs cannot do for them?
...
|
To generate answers similar to these, here's the prompt: What are some assignments for a undergraduate university (discipline) course that have students practice or demonstrate things LLM AIs cannot do for them?
Like a lot of responses from LLM AIs, these suggestions are typically vague. And an LLM AI–ChatGPT, for example-- may attempt to simulate the things it tells it can not do very well, if prompted by you (or a student.) Therefore, the faculty member needs strong command of any disciplinary knowledge involved in the assignment, if they are to assess student work for accuracy or integrity. But these can be a good starting line for your process of thinking about assignments that are "AI proof"
Was This Written By AI?
LLM AIs are designed to simulate people's writing, but there are often signs that a text was written by AI:
- AIs do not adhere to standards of accuracy or truth, and so will invent events, people, or other details as are needed to plausibly simulate a person writing about a topic. This can even include sources cited within the text. AIs might misattribute real quotes to the wrong author or speaker. So if something seems (dramatically) untrue, this might be an AI at work.
- If a real author's work is identifiable in another text, it likely differs dramatically from text attributed to an author but in fact composed by an AI. AIs can simulate styles, such as a nine-year-old crafting a book report, or of an American of limited education describing his experiences during the Great Depression, but it likely cannot reproduce the work of an individual who is not a famous author.
- AIs may struggle with understanding a prompt, rather more than most real people. So it may in effect answer the wrong question, in whole or part.
Turnitin
The popular plagiarism prevention and detection service Turnitin purports to have has a toolset for detecting AI-composed writing within student submissions. COLI has experimented with it, and it seems to work in most respects. However, there are reported cases of it failing within the educational technology community, and in any case, Turnitin has not supplied enough information about how this tool works for COLI to recommend it. We Just as Turnitin does, we strongly recommend faculty follow up on any suspected unauthorized AI use among students with further steps. Check citations and quotes. Does the student's submission properly address the assignment prompt? Does it answer the right question or perform the correct procedure? Are quotes and citations properly attributed to real sources? Is there detail or depth of argument enough to satisfy the prompt? Discuss with students the context of the assignment and their submission.
Innovative Pedagogy
If you need to "AI-proof" your course, you have the opportunity to do something more. Can you make your assignments more effective as opportunities for students to practice or demonstrate the skills embodied in your learning objectives and goals?
...