Opinion Blog


Rick Hess Straight Up

Education policy maven Rick Hess of the American Enterprise Institute think tank offers straight talk on matters of policy, politics, research, and reform. Read more from this blog.

Classroom Technology Opinion

The Promise and Peril of AI for Education

How will artificial intelligence change students’ relationship to learning?
By Rick Hess — July 09, 2024 8 min read
Image shows a multi-tailed arrow hitting the bullseye of a target.
  • Save to favorites
  • Print

In “Straight Talk with Rick and Jal,” Harvard University’s Jal Mehta and I examine the reforms and enthusiasms that permeate education. In a field full of buzzwords, our goal is simple: Tell the truth, in plain English, about what’s being proposed and what it means for students, teachers, and parents. We may be wrong and we will frequently disagree, but we’ll try to be candid and ensure that you don’t need a Ph.D. in eduspeak to understand us. Today’s topic is artificial intelligence.

Rick

Rick: Let’s talk AI. It’s been a year and a half since ChatGPT was unveiled, which means we’re past the initial fixation on how artificial intelligence will end the world. In education, we’ve now moved onto more traditional pursuits: an admixture of enthusiasm for AI’s “transformative promise” and concerns about cheating, access, and privacy.

I feel contractually obliged to start with my ed-tech mantra: It’s not the technology that matters—it’s what we do with it. And we’ve a pretty lousy track record on that count. Larry Cuban has aptly noted that, from the radio to the computer, ed. tech has long been “oversold and underused.”

But that’s not an iron law. Technology can be transformative. As Bror Saxberg and I observed many years ago, some once-novel technologies—like the book and the chalkboard—did have transformative effects on learning by making it cheaper and more customizable, accessible, and reliable. The book democratized knowledge and expertise. The chalkboard gave teachers new ways to illustrate ideas and explain math problems.

There is obvious potential here, from AI-fueled technologies like tutoring systems to AI-enabled high-quality support for lesson planning and assessment. When it comes to tutoring, for instance, one of the age-old challenges is a dearth of trained, reliable tutors. Well, it doesn’t take a lot of imagination to see how (in theory) 24/7, reliable, personalized AI-powered tutoring could have a profound impact.

The actual benefits, of course, will depend on how AI is used and whether most students can actually learn effectively from a nonhuman mentor, and that’s all very much to be determined. But I want to focus on a seldom-explored dimension here: the ways in which AI seems to be a different kind of tool that’s disorienting and that could erode the relationship between students and knowledge.

Now, this isn’t about whether AI will be good for the economy. After all, some tools are good for productivity but lousy for learning. The development of GPS, for instance, made it far quicker, easier, and more convenient to find our way around. (Anyone who remembers wrestling with crumpled maps in the passenger seat knows what I mean.) There was a cost, of course: GPS ravaged our sense of physical geography. Because geography doesn’t loom especially large in schooling today, the impact of GPS on education has been pretty modest.

When it comes to AI, though, I fear things are much more disconcerting than they are with a GPS. We already have a generation of students who’ve learned that knowledge is gleaned from web searches, social media, and video explainers. Accomplished college graduates have absorbed the lesson that if something doesn’t turn up in a web search, it’s unfindable. And we’ve learned that few people tend to fact-check what they find on the web; if Wikipedia reports that a book said this or a famous person said that, we tend to take it on faith.

Jal: Yep. Especially with young people who are learning to do something for the first time, there is a real risk that AI will replace the hard work of learning how to write, design, or program. Tools like AI are at their best when they complement human intelligence and skill, but those human capacities develop slowly and often painfully, which is why shortcuts are so tempting but ultimately misguided.

Learning environments often take away some of these tools to force students to concentrate on learning the skills they are being taught. For instance, laptops are banned from many Harvard Business School classes because computers provide too many interruptions and professors want students to engage in case discussions without distractions. When I was learning to ski as an adult, my instructor forbade me to use poles until I could get down the mountain without them. When we teach pitching in baseball, we often start the kids on one knee so that they can get the top half of the body right before they worry about the legs. So, there is a long tradition of restricting the learning environment so as to focus on the actual skills that we want to teach.

Teachers can do that in school, but the use of ChatGPT and Microsoft Co-Pilot for homework is, based on what I can see, absolutely rampant. And here I think AI is revealing an old problem rather than creating a new one: There is so much work that just has to be completed, but students see little purpose or meaning in what they are doing. TikTok is awash with videos about HOMEWORK—Half Of My Energy Wasted On Random Knowledge. In many classes, kids do so many worksheets that teachers don’t have time to mark them. If this is what learning is like—particularly if the assignments are mainly asking for summaries of factual knowledge—AI is going to be the solution. But if kids had fewer assignments that connected to a purpose they saw as relevant, and if more of that work were original rather than the same recycled assignments year after year, then perhaps we would see more original thinking and less AI-generated content.

Rick: Agree. Agree. Agree. It strikes me that AI is simultaneously creating powerful ways to bolster instruction, threatening to erode learning, and automating pointless routines. One of the problems is that all of this typically gets scrambled together in the public mind, along with confusion about how AI works—making it tough to think rigorously about risks and benefits.

Given that, I want to flag three points. First, as we’ve noted, AI could prove to be a useful tutor. Experience suggests, though, that few students actually access tutors unless the time is deliberately scheduled and monitored. And there are real practical questions about how well tutoring will work for learners if it’s devoid of human interaction. So, we should be duly measured about the promise of “a tutor in every pocket” until we see more.

Second, the folks working on AI in schools consistently tell me that one of the biggest near-term opportunities is for AI to be a support system and a time-saver for educators. It can help streamline the IEP process so that special educators spend more time educating. It can help with lesson planning, slide decks, and reporting, allowing teachers to focus more on students and less on paperwork. The problem is that AI is only as good as the material it’s been fed. After all, AI doesn’t actually “know” things; rather, it predicts the next word in a sentence or the next pixel in a picture based on the data it’s been trained on. That’s why it can be hugely unreliable—explaining the annotated bibliographies filled with nonexistent articles and the legal briefs citing nonexistent case law. If teachers are relying on ChatGPT to construct lesson plans or instructional materials, they are unintentionally wading into waters rife with huge concerns about accuracy, reliability, and bias.

Third, given these issues with trustworthiness, we should be far more nervous about AI becoming a default resource for learners. For this reason, the most famous works of dystopian fiction may not have been bleak enough about the risks posed to human knowledge and free thought. After all, in 1984 and Fahrenheit 451, the book burners have a tough job. They have to bang on doors and comb through archives in their attempt to stomp out “thoughtcrime.” The emergence of a few giant, poorly trained AI systems could make thought policing easy and breezy: Tweak an algorithm or produce a deepfake and you can rewrite vast swaths of reality or hide inconvenient truths. And it’s all voluntary. No one needs to strip books from shelves. Rather, barely understood AI is obliquely steering students toward right-think.

The legacy of social media is instructive. Twenty years ago, there was a lot of enthusiasm for all the ways in which the democratization of the internet was going to promote healthy discourse, mutual understanding, and knowledge. That mindset was particularly evident in education. Things didn’t turn out quite the way we hoped. I’m hoping we’ll do better this time.

Jal: Great points. A couple of things to pick up on. Will tutors work in the absence of human interaction? It depends on who the learners are and, more importantly, whether they want to learn the subject. David Cohen and I wrote a paper a few years ago that argued that reforms succeeded when they met a need that teachers thought they had. The same is true for students. Khan Academy has been a godsend to high school and college students whose teachers and professors are bad at explaining things. However, if a student is struggling in a compulsory subject and doesn’t really want to learn it, then offering them an AI tutor is not likely to fix things.

Your point about social media is compelling. We’ve gone from the nascent internet promising to democratize knowledge to a much more developed social media ecosystem causing a national panic about what it does to teenagers. Unlike, say, the CD-ROM, AI is not a “thing” that schools are trying to get students to use. It is an all-pervasive technology that will change many aspects of life—and schools, like other institutions, have no choice but to figure out how to respond. Because the technology is rapidly evolving, it’s often not going to be clear what the right answer is. Is it OK for AI to write a Valentine’s Day message? A sermon? It’s not just schools that have a lot to figure out.

Given that, I have one suggestion: Let students in on the conversation. Let them think about when it is OK to use AI—for instance, as a starting point to generate some ideas for writing—and when doing so would impede skill development or create unfairness. Have students draft guidelines about the uses of AI in school, and as they do so, let them read the work of the various task forces and commissions that are pondering the same questions. Not only would this turn AI into a teaching moment, it also promises to have students feel a sense of ownership over AI guidelines. Students are much more likely to follow guidelines that they help develop rather than ones that were imposed upon them by adults.

AI is here to stay. We need to figure out ways to channel it in ways that serve human ambitions, and students should play a part in that process.

The opinions expressed in Rick Hess Straight Up are strictly those of the author(s) and do not reflect the opinions or endorsement of Editorial Projects in Education, or any of its publications.

Events

This content is provided by our sponsor. It is not written by and does not necessarily reflect the views of Education Week's editorial staff.
Sponsor
School & District Management Webinar
Leadership in Education: Building Collaborative Teams and Driving Innovation
Learn strategies to build strong teams, foster innovation, & drive student success.
Content provided by Follett Learning
School & District Management K-12 Essentials Forum Principals, Lead Stronger in the New School Year
Join this free virtual event for a deep dive on the skills and motivation you need to put your best foot forward in the new year.
This content is provided by our sponsor. It is not written by and does not necessarily reflect the views of Education Week's editorial staff.
Sponsor
Privacy & Security Webinar
Navigating Modern Data Protection & Privacy in Education
Explore the modern landscape of data loss prevention in education and learn actionable strategies to protect sensitive data.
Content provided by  Symantec & Carahsoft

EdWeek Top School Jobs

Teacher Jobs
Search over ten thousand teaching jobs nationwide — elementary, middle, high school and more.
View Jobs
Principal Jobs
Find hundreds of jobs for principals, assistant principals, and other school leadership roles.
View Jobs
Administrator Jobs
Over a thousand district-level jobs: superintendents, directors, more.
View Jobs
Support Staff Jobs
Search thousands of jobs, from paraprofessionals to counselors and more.
View Jobs

Read Next

Classroom Technology Q&A Google Executive: What AI Can and Can't Do for Teachers
Jennie Magiera, Google's head of education impact, discusses the role AI should have in K-12 education.
8 min read
Close-up stock photograph showing a touchscreen monitor with a woman’s hand looking at responses being asked by an AI chatbot.
E+
Classroom Technology What Drones Are Doing to Deliver Better Student Engagement
Working with drones can motivate students, as well as teach skills like coding, collaboration, and problem-solving.
2 min read
The view over the shoulder of a high school student while he is holding a drone with the camera image showing on a laptop sitting on a nearby chair.
E+/Getty
Classroom Technology 3 Tips for Using Tech to Meet All Students' Needs
Technology is everywhere in most classrooms, but equitable access to it for all students still isn’t a reality.
2 min read
Photo of elementary school students using laptops in class.
iStock
Classroom Technology Key Questions for Districts to Ask as They Develop an AI Strategy
Here's what educators need to know about creating guidelines and building literacy for the use of artificial intelligence.
4 min read
Photo of computer chip with letter “AI.”
E+