A confession: I asked ChatGPT to help me write this.
I鈥檝e been a writer for most of my life, and a teacher of writing for my entire adult life, but I still can鈥檛 stand the sight of a blank page (or screen). I have a hard time focusing my early, often disorganized, thoughts and I feel immense pressure to word everything perfectly the first time, even though my years of training have taught me that a messy draft on paper is better than a perfect one in my head.
Some lessons are easier to teach than to learn.
Enter generative AI: I prompt the tool based on my knowledge of my current writing situation, and I read and analyze the output it gives me before making decisions about my own text. Some of AI鈥檚 ideas or phrasing ring true, and others provide me with a point of departure. Along the way, I get a better sense of my own goals for writing and the kinds of choices I can make to deliver my message effectively.
More often than not, seeing what鈥檚 wrong with the AI output can help me hone in on what I actually want to say and how I want to say it.
It can also help my students: undergraduate and graduate students in the Department of Writing and Rhetoric who might take my Literacies of
Artificial Intelligence orRhetorics of AI and Big Data courses, but also the computer science, advertising and public relations, and health sciences majors who take my Professional Writing class.
They are all aware that AI is drastically changing the workforce they鈥檙e about to enter. But my courses help prepare them for this reality.
In Professional Writing, for instance, we talk about the flood of AI-generated resumes submitted to every job they might apply for (not to mention the AI tools used to screen all those resumes), as well as the potential expectation that they鈥檒l use AI to write (or code) more efficiently or more uniformly once they鈥檙e on the job. We experiment with using AI to do some of that writing, and we have critical conversations about the good, the bad and the uncanny results of those experiments, from the made-up (so-called 鈥渉allucinated鈥) sources ChatGPT confidently provides, to its apparent love of the em dash, to its under-supported examples and surface level analysis.
We talk about what鈥檚 gained and lost when writing tasks are offloaded in part or wholesale to AI tools, and how to decide when we鈥檙e OK with that offloading and when we aren鈥檛. In other words, we talk about what it means to be a writer in the age of AI.
As a teacher, I鈥檓 transparent with my own uses of AI and the things I think these tools do well, like helping me beta test instructions for activities and turning my class notes into slides.
I don鈥檛 outright ban the use of these tools in my classes because doing so both fails to prepare students for workplace writing and robs them of an opportunity to practice a new form of critical digital literacy.
Instead, we foreground discussion of the rhetorical situation (including our audience, genre and purpose) with each writing assignment. We have open conversations about the aspects of the assignment with which an AI tool might be helpful, like comparing and contrasting sources, or acting as a test reader with a specific persona (as well as aspects AI is unlikely to help with much at all, like describing the writer鈥檚 own experiences).
My hope is that as a result of these conversations, students leave my class better able to articulate what they bring to the table as writers in addition to considering where and how AI tools will fit into their writing processes.
More than anything, I want my students to feel like what they know and what they have to say as writers still matter, so maybe they can feel a little bit better when they sit down in front of a blank page.
