.

How do you solve a problem like … AI? (Opinion)

Screenshot from The Sound of Music
Screenshot from The Sound of Music

Previously, I shared a Vox resource on Generative AI in teaching and learning, with the warning that we risk substituting cool new tech for the desirable difficulties of learning. In the video, they use the metaphor of Google Maps: technology may help us find our way, but it can also prevent us from learning to find our own way, constructing our own mental maps. 

What about Generative AI and artmaking? Does generating imagery/ideating give us a starting point or inhibit our ability to visualize, draft, and experiment? Yes. I’ve learned much from Leon Furze’s post The Myth of the AI First Draft, acknowledging that writing is supposed to be hard work, to a point. We have some limited evidence that, for beginners, exposure to AI samples may marginally improve creative output. What’s challenging about addressing a new development like AI is that implementation has far outpaced the science and education communities’ ability to study it systematically, understand it any depth, and then recommend practices based on data and experience (not just moral panic)

The question of “How do you solve a problem like AI?” may not even be the right one to ask. In a Substack post on teaching writing, Emily Pitts Donahoe embraces a creative approach to working with Generative AI: using it in different ways, at different stages of writing, and reflecting on the experience (What did the Gen AI provide you to work with, how did it help move you forward, and how does it compare to the way you usually write/create?). What’s interesting to me about this approach is that it leaves unanswered the question of whether the AI is good or bad and asks what the experience is like (and the results). 

Teaching in light of generative AI doesn’t necessarily mean adopting a draconian or permissive policy, but it does require us to be clear about what the tool is doing (and what the student is or isn’t doing). We might need to be clearer about the purposes of our assignments, our goals for student learning, and what students might expect in terms of desirable difficulties. We also might need to engage in some purposeful experimentation. This is not an endorsement of AI (I have a fair share of ethical qualms, such as energy use and the need for regulation), but I acknowledge a need for dialogue, further study, and patience. In the meantime, I’m happy to discuss the ideas you have for addressing AI in your courses, and you can always reach me at: asmith@pcad.edu 

Previous Posts

It’s Complicated: The Relationship Between Effort and Learning

Recently, someone shared a New York Times piece with me from Adam Grant (externally linked here), a bestselling author of nonfiction books on applying psychological research to lead more rewarding…

The Role of Joy in Persistence

Quitter’s Day is (unofficially) observed on the second Friday of January (thus, this year it was January 10). It refers to the day when ‘most people’ quit their New Year’s…

Great Expectations and Growth Mindsets

You may have heard this story: a teacher is told that their students have above average IQs, and the teacher engages with their class, with above average results. The teacher…

End of Semester Reflections: What’s in a grade?

We’ve arrived at the end of the fall semester, and students have turned in their final projects. You may be finishing up grades this week, and everyone deserves a well-earned…

The Role of Reflection in Creativity and Learning

In previous posts, I’ve discussed the creative process at different stages: the benefits of limits and the argument for valuing quantity over ‘quality’ in the initial, divergent-thinking stages, and the…

Gratitude for Teachers (and Learners)

In honor of Thanksgiving, this week’s column emphasizes the role of relationships in our paths through higher education. I’m grateful for the many people I’ve met and worked with, and…