ChatGPT and Education
Sam Altman-the OpenAI CEO-said that it’s best to think of GPT-4 as a reasoning engine. Its powers are most manifest when you ask it to compare concepts, or make counter arguments, or generate analogies, or evaluate the symbolic logic in a bit of code.
On the other hand, in Taming Silicon Valley, Gary Marcus wrote that these systems can produce content that is factually incorrect and even nonsensical. This is because they are based on statistical predictions of words and not a true understanding of the world. The models can generate misinformation, often mixed with truths, and are unable to verify or check facts. These models don’t "understand" the concepts they use or the people they describe, and therefore cannot distinguish between fact and fiction. The models generate new content by piecing together bits of the training data, often without understanding the context. They record the statistics of words, but lack the ability to reason or fact check. This means that while they can produce text that sounds natural, the content is not always correct or truthful. It is important that we are aware of how AI works, what it can and cannot do.
What educators these days must now do is show students not only how to find information, but what information to trust and what not to, and how to tell the difference.
What is it?
It is a generative AI. If we ask Google something, it will direct us to webpages that may contain the answers. We then go look for the answers on those webpages. If we put the same questions to chatGPT, within seconds you’ll get the answers in a slick, well-structured blocks of text several thousand words long on almost any topic it is asked about, from string theory to Shakespeare. Each essay it produces is unique, even when it is given the same question again.
“Questions/requests” technically known as prompts can be anything e.g. write an essay on the significance of MB2 in Endodontics, write a novel about robots in style of Edgar Alan Poe, write a code for webpage containing a picture of a pig etc.
With the ability to produce relatively well structured essays in an instant with virtually zero effort, it looked as if ChatGPT would undermine the way we encourage students to learn and test what they have learned, a cornerstone of education.
The problems
Cheating
The response from educational institutions was swift, but vast ranging from fiercely negative i.e., right out ban, prohibiting the use of this technology in any way shape or form to optimistically positive. Some educators fear that this tool will be used for cheating on exams or assignments. Some schools/universities have blocked access to OpenAI’s website from their networks.
Accuracy
It we ask chatGPT something, the answer most likely will be a mixture of the truth and falsehood. We don’t know which part is written by the oracle and which by hallucinated individual. As far as I have tried, the accuracy is quite impressive, but it is field dependent. It’s been shown to be able to pass a bar exam and the final exam of MBA programs in the US. We just don’t know when it’s going to go delusional. In addition, the answer is spelt out in style full of confidence. The uninformed users may feel that the answers are validated and take it as unquestionable truth.
Pondering
Cheating is not a new problem: schools have survived calculators, Google, Wikipedia, essays-for-pay websites, and more.
Back when calculators first became inexpensive and ubiquitous. There was a debate on how we should incorporate this technology in teaching and learning. Banning it was fashionable for a while, later on though it was allowed on school grounds and even in exam. Math exams since have evolved from simple calculations easily done on calculators to more sophisticated calculations. Justifications, explanations of chosen solutions rather than an answer or final result became a norm in exams.
Googling makes accessing information easy. Memorising facts is no longer as important as the time before Google. Nevertheless we should understand that basic knowledge of facts is just the bedrock on which other forms of learning, such as analysis and evaluation, sit. Perhaps educators should shift the focus from memorising facts to honing students’ analytical and evaluating skills. We should teach what/how to ask Google and what to make of the results. In other words, we need to teach them not only how to find information, but what information to trust and what not to, and how to tell the difference.
If ChatGPT makes it easy to cheat on an assignment, we should throw out the assignment or modify it to fit with higher forms of learning rather than ban the chatbot. This is because the technology makes that particular skill (that was honed or tested) in an assignment redundant. Also banning or blocking it is futile. We just can’t deny the existence of a technology. It’s there. It exists. It’s accessible (might not be at some schools, but it is certainly accessible elsewhere).
As in calculator and Google case, by the same token, we simply cannot reverse the clock to the time when chatGPT didn’t exist, we should instead embrace it, understand it, see how it works, what the limitations, and how to best use it.
This is (potentially) how…
Improve critical thinking skills
Students to use ChatGPT to generate text on a topic and then point out the flaws.
Educators can even lean into ChatGPT’s tendency to falsify, misattribute, and straight-out lie as a way of teaching students about disinformation. Imagine using ChatGPT to pen essays that conceal subtle logical fallacies or propose scientific explanations that are almost, but not quite, correct. Learning to discriminate between these convincing mistakes and the correct answer is the very pinnacle of critical thinking, and this new breed of academic assignment will prepare students for a world fraught with everything from politically correct censorship to deepfakes.
Strengthen argument
ChatGPT can play the role of a debate opponent and generate counterarguments to a student’s positions. By exposing students to an endless supply of opposing viewpoints, chatbots could help them look for weak points in their own thinking.
Students to use ChatGPT to generate an argument and then annotate it according to how effective they think the argument is for a specific audience. Then turn in a rewrite based on their criticism.
Example: write an essay arguing why it’s ok to not searching for MB2 and leave it untreated.
Private tutor
Private tutoring has been shown to lift academic performance of a class up to 2 standard deviation in comparison to a standard teaching class. Students could ask chatGPT to prepare questions to test their understanding of a topic. Assess the answers, study and repeat. All this adds up to a simple but profound fact: anyone with an internet connection now has a personal tutor, without the costs associated with private tutoring. Sure, an easily hoodwinked, slightly delusional tutor, but a tutor nonetheless.
Assignment 2.0
If students want to use ChatGPT in their written assignments, we should assess the prompt as well as—or even rather than—the essay itself. Knowing the words to use in a prompt and then understanding the output that comes back is important.
Textbook 2.0
Future textbooks could be bundled with chatbots trained on their contents. Students would have a conversation with the bot about the book’s contents as well as (or instead of) reading it. The chatbot could generate personalised quizzes to coach students on topics they understand less well.
ChatGPT is a new technology, if we didn’t use it, we wouldn’t know what it capable of. We and our students should experiment with the tool, and find way to guide them through how, when, and where it can be used.
===
Sources
ChatGPT is going to change education, not destroy it
https://www.technologyreview.com/2023/04/06/1071059/chatgpt-change-not-destroy-education-openai/
Banning ChatGPT will do more harm than good
https://www.technologyreview.com/2023/04/14/1071194/chatgpt-ai-high-school-education-first-person/