Inhuman Rhetoric: Generative AI and Crisis Communication

Erick Piller Nicholls State University

Abstract

This article considers the rhetorical risks of using generative AI to compose organizational communication during crises or in the aftermath of tragedies. It focuses on a case study in which representatives of Vanderbilt University’s Peabody College of Education and Human Development disclosed their use of ChatGPT to write a response to a school shooting at another university. The author argues that although generative AI can often be useful in technical and professional communication, it can also undermine perceptions of “rhetorical humanity” if its use is disclosed or discovered, making it rhetorically risky in certain contexts. Thus, knowing when not to utilize AI is an important aspect of AI literacy for practitioners.

Journal
Journal of Business and Technical Communication
Published
2025-01-01
DOI
10.1177/10506519241280594
Open Access
Closed
Topics

Citation Context

Cited by in this index (0)

No articles in this index cite this work.

Cites in this index (1)

  1. Journal of Business and Technical Communication
Also cites 6 works outside this index ↓
  1. 10.1177/23294906221074311
  2. 10.1080/03637751.2010.502536
  3. 10.1111/j.1460-2466.2008.01410.x
  4. 10.1111/j.1083-6101.2006.00019.x
  5. 10.1108/JPBM-02-2022-3864
  6. 10.1177/1077699014538827
CrossRef global citation count: 8 View in citation network →