[Air-L] CFP: Special issue on (un)Stable Diffusions
Fenwick Mckelvey
mckelveyf at gmail.com
Wed Jun 7 09:14:35 PDT 2023
Please circulate widely and consider applying.
CFP: (un)Stable Diffusions: General-purpose artificial intelligence’s
publicities, publics, and publicizations
Edited by Fenwick McKelvey, Joanna Redden, Jonathan Roberge, and Luke Stark
(names in alphabetical order)
To be published in the open access Journal of Digital Social Research
<https://www.jdsr.io/>. Please submit your abstracts here:
https://forms.gle/hWQxvuRTWoFhVs3W6
The recent release of so-called “general-purpose artificial intelligence”
(GPAI) systems have prompted a public panic that AI-generated fiction is
now indistinguishable from fact. General-purpose AI refers to systems
“intended by the provider to perform generally applicable functions such as
image and speech recognition, audio and video generation, pattern
detection, question answering, translation and others” (Artificial
Intelligence Act, Council of the European Union, Procedure File:
2021/0106(COD), Title 1, Article 3(1b)). Contemporary technologies referred
to as GPAI include OpenAI’s ChatGPT as well as numerous text-to-image
generation software such as DALL-E and Stable Diffusion. GPAI systems are
already being deployed to support and entrench existing asymmetries of
power and wealth. For instance, the online news outlet CNET recently
disclosed that it has been publishing stories written by an AI and edited
by humans for months (Main, 2023).
The current concern over ChatGPT is an important moment in AI’s publicity
and publicization (Hansen, 2021), or what Noortje Marres refers to as
“material things” acting as “crucial tools or props for the performance of
public involvement in an issue” (Marres, 2010, p. 179). Amidst countless
opinion pieces and hot takes discussing GPAI, this special issue details
how scandal, silence, and hype operate to promote and publicize AI. We seek
interventions that question AI’s publicity and promotion as well as new
strategies of engagement with AI’s powerful social and political influence.
Concern over Generative AI, we argue, is limited by publicity around these
systems that has been framed by hype, silence, or scandal (Brennen, 2018,
2018; Sun et al., 2020). Publicity refers to the relations between affected
peoples and matters of shared concerns (Barney, 2014; Marres, 2015, 2018).
Historically, these relations have been mediated by the press (Schudson,
2008), but GPAI’s coincides with uncertainty about journalism’s status and
a rise of direct, one-step flow like effects of either citizen-to-citizen
or, in the case of ChatGPT, a direct link (Bennett & Manheim, 2006).
Scholars have observed that publicity around AI follows several distinct
patterns:
1. Hype (Ananny & Finn, 2020; Broussard et al., 2019; Mosco, 2004). This
discourse is prominent. Concepts like the fourth industrial revolution and
disruption function as self-fulfilling, with the consequence that
technology arrives always as good news. The launch of ChatGPT is a hallmark
of this hyped mode, fitting into a well-worn mode of a “normative framework
of publicity [that is] drained of its critical value, and convert[ed] from
a democratic asset to a democratic liability” (Barney, 2008, p. 92).
Opening AI to the public, no matter the consequences for society, as in the
case of ChatGPT, is portrayed as a good even as this publicization shifts
focus to acceptance and inevitability.
2. Silence. If not positive, AI coverage is marked by gaps and aporias,
or in effect, closures when aspects of AI remain too uncontroversial to
report as well as larger work into the logics of AI imaginaries (Bareis &
Katzenbach, 2021). The result is that AI-related issues seldom make the
political information cycle, as in the case of ChatGPT’s potential
violations of privacy and copyright law, discussed below (Chadwick, 2013).
3. Scandals – or what we refer to as proofs of social transgressions –
are a pronounced feature of contemporary technological coverage and
governance (Bosetta, 2020; Lull & Hinerman, 1998; McKelvey et al., 2018;
Trottier, 2017). Scandals, which we stress do not necessity involve
opportunities for public engagement nor democratic praxis, result from a
mutually reinforcing relationship between newsrooms looking for easy,
high-engagement stories (Blanchett et al., 2022; Cohen, 2015; Dodds et al.,
2023) and the affordances of social media, largely functioning as a
distraction from other tasks (Hall et al., 2018).
4. Inevitability. AI discourses are dominated by technology firms,
government representatives, AI investors, global management consultancies,
and think tanks. These voices present a faith in data driven systems to
address social problems while also increasing efficiency and productivity
(Bourne 2019, Beer 2019). Such discourses are used to reinforce the idea
that the increasing use of AI applications across all spheres of life is
inevitable, while sidelining or ignoring meaningful engagement with the
ways these applications cause harm (Walker 2022).
Our special issue seeks interventions focused on:
1. Critical and comparative studies of AI’s publicities with regards to
the launch and hype of AI. We particularly welcome papers that focus on
cases outside the Global North;
2. Ethnographic, discursive, or engaged research with AI’s public such
as AutoGPT, HustleGPT, or other publics forming around the use, misuse or
resistance to GPAI;
3. Interventions or reflections on critical practices, such as community
engagement and mobilization, futures literacy, or capacity building, for
better publicizations of AI de-centering the strategic futuring employed by
large technology firms.
Please submit an extended abstract (1000 words) by 1 August 2023.
Accepted full papers due 1 December 2023.
Planned publication Spring 2024.
https://forms.gle/hWQxvuRTWoFhVs3W6
More information about the Air-L
mailing list