.

How do you solve a problem like … AI? (Opinion)

Screenshot from The Sound of Music
Screenshot from The Sound of Music

Previously, I shared a Vox resource on Generative AI in teaching and learning, with the warning that we risk substituting cool new tech for the desirable difficulties of learning. In the video, they use the metaphor of Google Maps: technology may help us find our way, but it can also prevent us from learning to find our own way, constructing our own mental maps. 

What about Generative AI and artmaking? Does generating imagery/ideating give us a starting point or inhibit our ability to visualize, draft, and experiment? Yes. I’ve learned much from Leon Furze’s post The Myth of the AI First Draft, acknowledging that writing is supposed to be hard work, to a point. We have some limited evidence that, for beginners, exposure to AI samples may marginally improve creative output. What’s challenging about addressing a new development like AI is that implementation has far outpaced the science and education communities’ ability to study it systematically, understand it any depth, and then recommend practices based on data and experience (not just moral panic)

The question of “How do you solve a problem like AI?” may not even be the right one to ask. In a Substack post on teaching writing, Emily Pitts Donahoe embraces a creative approach to working with Generative AI: using it in different ways, at different stages of writing, and reflecting on the experience (What did the Gen AI provide you to work with, how did it help move you forward, and how does it compare to the way you usually write/create?). What’s interesting to me about this approach is that it leaves unanswered the question of whether the AI is good or bad and asks what the experience is like (and the results). 

Teaching in light of generative AI doesn’t necessarily mean adopting a draconian or permissive policy, but it does require us to be clear about what the tool is doing (and what the student is or isn’t doing). We might need to be clearer about the purposes of our assignments, our goals for student learning, and what students might expect in terms of desirable difficulties. We also might need to engage in some purposeful experimentation. This is not an endorsement of AI (I have a fair share of ethical qualms, such as energy use and the need for regulation), but I acknowledge a need for dialogue, further study, and patience. In the meantime, I’m happy to discuss the ideas you have for addressing AI in your courses, and you can always reach me at: asmith@pcad.edu 

Previous Posts

Gratitude for Teachers (and Learners)

In honor of Thanksgiving, this week’s column emphasizes the role of relationships in our paths through higher education. I’m grateful for the many people I’ve met and worked with, and…

Critique Benefits the One Giving Critique, Too

In a previous post, I shared the idea (not my own) that peer observation is just as useful for the one doing the observation. It can allow us to get…

Faculty Spotlight – Eric Weeks On the Vulnerabilities of Critique

Talking to Eric Weeks, Associate Professor and Chair of Photography & Video at PCA&D, you get the sense of someone whose ideas about art and teaching have evolved slowly, but…

CT&L Faculty Spotlight: Natasha Warshawsky

As we move the CT&L Corner from the weekly newsletter to a blog hosted on the CT&L webpage, I wanted to start with a new, monthly series of Faculty Teaching…

Peer Observation Benefits the Observer, Too

Yogi Berra is quoted as having said, “You can observe a lot by watching.” The research on peer observation would take the a step further: you can learn a lot…

How to prepare without overpreparing

With classes just two weeks away, I wanted to share two resources that have received positive responses from instructors I’ve interacted with here at PCA&D and in my previous roles.…