How will this tech shape teaching and
learning? What happens to the data that gets
processed through these products? What are the best
ways to use them? Layer on top of that the roles and
responsibilities of people at every level: students,
instructors, department chairs, administrators, and
you can see why so many people are unsure of their
next move.
I was thinking about these questions
as I dove into a case study on the
use of an AI grading tool, written by Rahul Kumar in
the International Journal for Educational
Integrity. An assistant professor in the
department of educational studies at Brock University
in Canada, Kumar developed the story of a
hypothetical adjunct instructor who — like most
adjuncts — is crushed by his workload and fearful of
how that is harming his professional future and his
personal life.
What I appreciate about Kumar’s piece
is that it creates a portrait of someone likely to
turn to AI to alleviate the demands placed on them.
Dr. A.I. Case, as he names his faculty member,
stumbles across a company that uses AI to discreetly
help professors grade papers. Should he use this
tool?
In the “yes” corner, as Kumar
describes it, is the fact that it could free up time
to devote to research and publishing, which could
eventually help him earn tenure. It would restore
work-life balance. It could also be more timely and
consistent than he might be, given that he teaches
twice as many courses — across two universities — as
his tenured colleagues.
In the “no” corner are questions of
cost, privacy, quality, and ethics, among other
things. Professor Case wonders what happens to the
students’ work once it’s uploaded, how good the AI
is, and whether using an AI tool for a core teaching
responsibility is both the right thing to do and a
good thing to do. The paper doesn’t end with his
choice; instead it reminds the reader of the
complexity of the situation.
I spoke to Kumar about what he hopes
people will take from his paper. So much of the
attention around AI, he said, has been focused on
whether students will use these tools inappropriately
or without attribution. But the same questions could
be applied to faculty members. Given how tenuous and
stressful the work environment is for
non-tenure-track faculty, colleges need to be alert
to these possibilities, and faculty members will
inevitably wrestle with them.
“Oftentimes precarity leads to risk
taking,” Kumar noted. And while all stakeholders
should have a say in decisions about the AI tools
that could affect them, that is often not the case.
Many professors are, in fact, operating without
guidelines. As a result, he said, “there’s no real
mechanism to know what people are doing, save for
self disclosure.”
However, there is no simple answer to
the question of whether to use AI tools, Kumar
emphasized. It depends on the type of tool, its
purpose, your goals, and even which discipline you
work in. If you’re teaching someone to write, then
allowing them to use a large language model like
ChatGPT may end up short-circuiting the development
of writing skills. But if you’re teaching a
computer-science class, then maybe the use of an LLM
to help with writing doesn’t matter as much. “We have
to get more sophisticated,” Kumar said, “and say,
Under what conditions, where and when is it OK as
opposed to not OK?”
Give Kumar’s piece a read and let me know
your thoughts on AI in teaching and the questions it
raises around ethics, governance, pedagogy, and
transparency. Your ideas could help inform future
stories and newsletters on the topic. You can reach
me at beth.mcmurtrie@chronicle.com
More
cautionary AI tales
The second story that got me thinking
about the complex nature of our AI-infused world came
from a grading controversy, which you may have seen circulating last
week. A professor at Texas A&M University at
Commerce ran students’ assignments through ChatGPT to
see if they were AI-generated, determined that they
were, and then threatened to fail students,
potentially holding up their diplomas.
The news made the rounds for several
reasons. One is that the professor misunderstood ChatGPT, which is a
large language model capable of turning out
AI-written essays. It is not itself an AI detector. The
other, of course, is that the professor threatened
such a dramatic step. (After the incident made
national news, a university spokesman told The
Washington Post that no student ended up flunking
the course or having their diploma withheld and that
the university is developing policies on AI use or
misuse in the classroom.)
The story highlights the growing
concern faculty members have about whether they can
trust what their students write. Like many of you, I
have read a number of posts on social
media in which faculty members have discovered that
some of their students used AI to produce papers and
complete exams. You may also have seen this opinion piece in The
Review by a graduate instructor who argues that
administrators need to act quickly to determine the
scale of the problem and devise some responses.
Detection tools are on the rise. So
far none of them are highly reliable, according to
academic-integrity experts, but that hasn’t stopped
faculty members from using them. That feeds into the
lack of clarity around proper-AI usage. As the Washington
Post story put it: “protocols on how and when to
use chatbots in classwork are vague and
unenforceable, with any effort to regulate use
risking false accusations.”
This is an ongoing conversation, of
course, and we would like to hear from you. We are
hoping to do some short- and longer-term stories on
how AI is affecting teaching and learning.
Did any of your students try to cheat
with ChatGPT or other AI tools in assignments this
past semester? Are you reworking your courses for the
fall to address AI? Do you have other concerns or
plans around utilizing AI in teaching? Please fill
out this Google form and
let us know what your experiences have been.
AI
insights
- Looking
for resources on how to teach about AI? Check
out this evolving project, called CRAFT, being
developed by Stanford University’s Graduate
School of Education, Institute for
Human-Centered AI, and Stanford Digital
Education.
- This essay, which
appeared in EduResearch Matters, walks
through some issues to consider when it comes to
using AI effectively.
- For a
bleak assessment on what AI means for teaching,
read Ian Bogost’s piece in The
Atlantic.
Thanks for reading Teaching. If you
have suggestions or ideas, please feel free to email
us at beckie.supiano@chronicle.com
or beth.mcmurtrie@chronicle.com.
— Beth
Learn more about our Teaching
newsletter, including how to contact us, at the Teaching newsletter archive
page.
|