[Air-L] Discussion of ChatGPT and other AIs and their uses

Charles Melvin Ess c.m.ess at media.uio.no
Sat Mar 7 05:08:54 PST 2026


Thanks to Peter for bringing this up - obviously critical discussion 
that has exploded in the literatures I'm still (somewhat) familiar with, 
especially over the last 5 years or so.

I've responded to Peter privately, partly because I wanted to send him a 
copy of a paper that is about to be published.  But here some of the 
additional response may be of more shared use or interest.

I and a close relative have been active users of ChatGPT over the past 
couple of years - primarily in the domains of scripting and elementary 
programming. At one point, for example, I was tinkering with C programs 
running under 2.11 BSD on a PiDP-11 emulator / replica I built as part 
of my exploring all of this. My relative comes to all of this with 40+ 
years of professional experience, including decades as a systems 
administrator / security expert.

The short story is that the results for us have been decidedly mixed. 
As a start: as Peter rightly notes, everything depends on asking the 
right question / the question properly. Failure to do so can lead down 
some long, very frustrating, and ultimately useless, if not destructive 
rabbit holes. (Though the same can also often result from the right 
questions ...)

It is also clear after a point what the underlying nature of these 
devices is - namely, sophisticated inference engines, totally devoid of 
any sort of semantic understanding, etc., that generate what I often 
call SWAGs - scientific (sorta) wild-ass guesses.  (Less kindly, they 
are "stochastic parrots" that only can regurgitate the material they 
have been "trained" upon - so Bender et al, 2022.)

It further becomes clear that we are often its guinea pigs / trainers: 
our responses to its suggestions - most especially then they go wrong as 
they so often do - are then taken up as part of its training data and, 
hopefully, helps improve it in some way or another for use further down 
the road.

Last but not least - and no surprise: it is "trained" along many of the 
sophisticated lines of the various algorithms and related techniques to 
keep our "attention" that are familiar from SoMe.  The most obvious 
examples are the occasional choices offered between two different 
"personalities." What my relative aptly calls the butt-kissing algorithm 
is also on strong display, along with other simple tricks - "you're so 
close, now," etc.
(In other words: follow the money ...)

I continue to use it - as informed and guided by this sort of awareness 
of its strengths and limits: most pointedly, not as a replacement for my 
own thinking / muddling through complex projects - including research 
and writing - but as a very limited augmentation.

This is, as those who know me might suspect, further grounded in a 
fairly extensive philosophical framework and (relational / feminist / 
critical posthumanist ...) anthropology, one rooted (as a start - only 
as a start) in the Socratic dialogues as well as Aristotle: specifically 
Plato's The Phaedrus and the warnings there towards the end of the 
dialogue regarding the risks posed by the then-new media technology of 
writing.
Contra a tendency in recent media and communication studies (from ca. 
2012 forward, based on what I've managed to find) to interpret the story 
of Theuth and Ammon's warnings as a "media panic" - I rather show in a 
forthcoming paper via an ancient interpretive framework (ring 
composition / inclusio narratives) that this story serves first of all 
as a pedagogical device that helps bring home several of the lessons 
that unfold for the young Phaedrus cross the course of the dialogue.

As a start: that "bare letters" alone only provide us with a _semblance_ 
of wisdom and knowledge, and thereby present the risk of what Shannon 
Vallor has aptly characterized as "deskilling," our failing to acquire 
the practices and abilities ("virtues") necessary for acquiring such 
knowledge and wisdom on our own - most centrally, the capacity of 
phronesis  as a form of self-correcting, specifically ethical judgment.
What the moral panic reading tends to further miss is the subsequent 
discussion of "the gardens of letters" - the positive, proper uses of 
writing (including the central art of writing well vs. writing 
disgracefully which is thematic to the dialogue from the outset) as a 
complementary media technology requisite to the philosophical pursuit of 
knowledge and the capacity for good judgment as essential elements to 
good lives of flourishing.

Such a complementary approach, last but not least, can be traced through 
early critiques of classic AI, including Joseph Weizenbaum and 
especially Hubert Dreyfus (1973) - who identifies Plato's Euthyphro as 
the beginning of the story of AI as it starts out with a request for a 
rule-based approach to ethics, which turns out to be futile: phronesis, 
as not reducible to rules, is required instead - including by AI 
programmers themselves, as Dreyfus further illustrates.

In more technical terms: there arises here what becomes called the frame 
problem. The upshot is that such judgment is not computationally 
tractable - hence Weizenbaum's foundational distinction between judgment 
and calculation (1976).

<Aside>: those familiar with the origins and development of the AoIR 
ethics guidelines for internet research may recognize that phronesis, 
contra rule-based approaches such as utilitarianism and deontology, has 
been central to our development and suggestions for use of these 
guidelines from the start.
Another discussion entirely, but I have the very strong impression, from 
several sources and experiences over the past 2+ decades, that by 
grounding our guidelines on the already solid and fruitful phronesis / 
judgments of both new and more experienced researchers as a primary 
beginning point, coupled with the aim of building and using the 
guidelines in such a way as to foster our own phronetic abilities - 
rather than tick-boxing or, ho ho, asking ChatGPT ... - are primary 
contributors to their extensive and successful use. </aside>

Still more recently, work by, e.g., Virginia Dignum (2019, 2021) and 
Katherina Zweig (2019) (a computer scientist and bioinformaticist, 
respectively, show similar limits to generative AI approaches - 
reiterating the point that AI / ML / LLMs offer only semblances of 
knowledge, and hence represent similar risks of deskilling should we 
turn more and more of our cognitive / writing / ethical, etc. loads over 
to the machines.
All of this captures a central point from the Phaedrus - "know your 
tools," i.e., their strengths and limits, in order to then know how to 
best make appropriate use of them towards these larger ends.

If any of this is of further interest - including those of us who have 
discussed the Phaedrus over the years here - I'm happy to send along the 
paper detailing all of this (forthcoming in the Danish Yearbook of 
Philosophy: Open Source). I conclude there with some examples of such a 
complimentary approach to using LLMs in contemporary research and 
teaching that might be helpful or suggestive in particular.

Of course, critical comments and suggestions always welcome.

Best of luck with it all and thanks again to Peter for raising these issues.

- charles ess
Professor Emeritus, University of Oslo

On 07/03/2026 08:17, Peter Timusk via Air-L wrote:
> Hello I haven't really posted to this list for a decade or more I think. I
> work in statistical computing these days and rarely spend time on my
> employer's ( Statistics Canada) Internet Use surveys though I did write one
> short study in 2009.
> 
> At work we do a lot of statistical computing and even analysts are
> expected to be able to calculate statistics for their own papers. We are
> moving off the statistical computing language SAS and going open source
> with R and Python. We are also expected to start exploring AI for use in
> making things more efficient and saving money.but also to do so with high
> ethical standards and of course working within our culture of citizen and
> business data privacy.
> 
> I just wanted to write this email to see if there is any discussion
> possible here via email.
> 
> I have used AI for the past few months to do some volunteer work. Here I am
> making an online collection of PDFs into a searchable set of blog posts on
> a Word Press site. I have learned that one can ask long search questions to
> ChatGPT and it will produce code examples and workflows, say to show me how
> to use R and the PDFTools an R package to extract text from these PDFs. I
> generally have to know what I am looking for to ask the questions but
> am happy so far with the results of this AI use. I also have to read the
> code and understand the loops and logic going on which allows my
> programming experience to validate the results.
> 
> I am about to use Python suggested by ChatGPT to make an offline archive of
> all my Facebook posts from 2007 to now and also publish these in some neat
> fashion to a WordPress blog.
> 
> anyways hoping to join the discussions on AI bias and ethics that have been
> going on for awhile.
> _______________________________________________
> The Air-L at listserv.aoir.org mailing list
> is provided by the Association of Internet Researchers http://aoir.org
> Subscribe, change options or unsubscribe at: http://listserv.aoir.org/listinfo.cgi/air-l-aoir.org
> 
> Join the Association of Internet Researchers:
> http://www.aoir.org/



More information about the Air-L mailing list