Should educators worry about ChatGPT?

Ted Underwood
Ted Underwood, Professor

The artificial intelligence chatbot ChatGPT can, among other things, generate essays and write computer code. Since being released to the public for testing late last year, it has raised concerns about students using ChatGPT to complete their homework and led some secondary public schools to ban it and college professors to change their course assignments. Ted Underwood is a professor of English and of information sciences and the associate dean of academic affairs in the School of Information Sciences. He recently commented in Inside Higher Ed on how to view the technology's place in higher education. He talked with News Bureau arts and humanities editor Jodi Heckel.

What is ChatGPT and how is it different from previous versions of chatbots?

Discussion tends to focus on ChatGPT because this product was made widely available for free last fall – and it was a little easier to use than earlier models of language. But ChatGPT is far from unique. 

ChatGPT is based on technology that has been around in one form or another since OpenAI released the first version of generative pre-trained transformers in 2018. The basic idea is that a model is trained to predict the next word in an observed sequence of words. Then when you write a short passage – a "prompt" – the model can predict the next word in the sequence, and then the next word, and so on. To do a really good job, a model needs to recognize high-level patterns and behave as if it understood language. Because models like this grew better at generalizing as researchers increased the size of the model, they are sometimes called "large language models." They're also called "generative AI" because they don't just analyze texts but use what they have learned to create new texts.

ChatGPT improves on earlier versions of this technology by training a model specifically to treat prompts as turns in a conversation and respond, instead of just continuing your statement.

ChatGPT is not unique in this; similar models have been released by Google, Meta and Anthropic. And OpenAI itself has recently released GPT-4, which is better than ChatGPT.

Should educators be worried about students using ChatGPT or other artificial intelligence writing programs to write their research papers, or should they look at how AI applications can be used as educational tools to help students learn?

We should do both. But I would urge us to focus a little less on the short-term fate of our assignments and more on long-term consequences for students.

Some students are using models to help write their papers and do homework, and yes, that is something to worry about. We want students to learn, and if they're just pasting an assignment into a box and hitting return, they're not learning much.

But that's a small part of a bigger issue, which is that the students now entering college are likely to graduate into a world transformed by artificial intelligence. Models like ChatGPT are already being integrated into word processing software and search engines. In 10 years, they will be as familiar as autocomplete is to us now. So, telling students "just say no to AI" is not going to be a sufficient way to prepare them for the 2030s. Students will be using these models, and will need to understand them.

There are definitely some contexts, like a closed-book exam, where it's appropriate to say "don't use AI," just as we currently say "don't look up the answer on the web." But universities also will need to offer courses and assignments that teach students how to understand these tools and use them in appropriate, creative ways.

What are some other uses of AI language models?

Right now, we're approaching AI in the way we often approach a new technology: We're trying to fit it into an existing niche. Large language models are widely understood as writing machines, so we think, "maybe students will use models to write their term papers." Models also seem able to answer questions, so we think, "maybe they'll replace search engines."

A language model isn't a library or a copy of the internet; it's literally just a model of language. People will be disappointed if they expect the language model itself to provide knowledge.

I think we're going to find more interesting ways to use this technology. Instead of asking old questions for which answers already exist, the interesting way to use one of these models is often to hand it new evidence that you want analyzed, while precisely describing the analysis you want to perform.

I like British programmer Simon Willison's way of putting this, which is that a language model is a calculator for words. The model doesn't contain exhaustive knowledge. But it's a flexible little machine that can follow verbal instructions, transform text and think out loud – so to speak – in writing.

You wouldn't ask a calculator to perform a physics experiment or engineer a bridge on its own – and by the same token, we probably shouldn't ask a language model to write important documents on its own. But if we can break a project down into well-defined tasks, a language model may make those tasks easier. A model could, for instance, read through a stack of emails one by one in order to assess their relevance to a question, and then instruct itself to condense the most relevant emails into a summary.

In short, these aren't substitutes for human writing or human knowledge. They're flexible tools for transforming language. We'll need to learn how to use them, and it's even possible that we'll end up using them for analysis more than we do for writing.

Tags:
Updated on
Backto the news archive

Related News

Tibebu joins the School

The iSchool is pleased to announce that Haileleol Tibebu joined the faculty as a teaching assistant professor on January 1, 2025. His research and teaching interests include responsible AI, AI policy and governance, algorithmic fairness, and the intersection of technology and society.

Haileleol Tibebu

Rhinesmith joins the faculty

The iSchool is pleased to announce that Colin Rhinesmith joined the faculty as a visiting associate professor on January 1, 2025. His position will become permanent following approval by the University of Illinois Board of Trustees. He previously served as founder and director of the Digital Equity Research Center at the Metropolitan New York Library Council.

Colin Rhinesmith

SafeRBot to assist community, police in crime reporting

Across the nation, 911 dispatch centers are facing a worker shortage. Unfortunately, this understaffing, plus the nature of the job itself, leads to dispatchers who are often overworked and stressed. Meanwhile, when community members need to report a crime, their options are to contact 911 for an emergency or, in a non-emergency situation, call a non-emergency number or fill out an online form. A new chatbot, SafeRBot, designed and developed by Associate Professor Yun Huang, Informatics PhD student Yiren Liu, and BSIS student Tony An seeks to improve the reporting process for non-emergency situations for both community members and dispatch centers.

Yun Huang

Hoiem receives Schiller Prize for “Education of Things”

Associate Professor Elizabeth Hoiem has won the 2025 Justin G. Schiller Prize from The Bibliographical Society of America for her book, The Education of Things: Mechanical Literacy in British Children's Literature, 1762-1860 (University of Massachusetts Press). The prize, which recognizes the best bibliographical work on pre-1951 children's literature, includes a cash award of $3,000 and a year's membership in the Society. 

Elizabeth Hoiem

Chan authors new book connecting eugenics and Big Tech

Associate Professor Anita Say Chan has authored a new book that identifies how the eugenics movement foreshadows the predatory data tactics used in today's tech industry. Her book, Predatory Data: Eugenics in Big Tech and Our Fight for an Independent Future, was released this month by the University of California Press and featured in the news outlets San Francisco Chronicle and Mother Jones.

Anita Say Chan