Humanist Discussion Group

Humanist Archives: Jan. 3, 2025, 7:39 a.m. Humanist 38.298 - AI, poetry and readers (or inaugurating the new year)

				
              Humanist Discussion Group, Vol. 38, No. 298.
        Department of Digital Humanities, University of Cologne
                      Hosted by DH-Cologne
                       www.dhhumanist.org
                Submit to: humanist@dhhumanist.org




        Date: 2025-01-02 18:28:42+00:00
        From: Tim Smithers <tim.smithers@cantab.net>
        Subject: Re: [Humanist] 38.285: AI, poetry and readers

Dear Jim,

I took a post December Solstice break, hence the delayed, but
not lessened, thank you for your further response and new
proposal.

Your concise summary nicely outlines, I think, where we agree.
I am, however, still attached to the romantic idea that it
matters, and makes a difference, that human produced text
comes from writing words formed in and by minds.  Even if the
mind in question is no longer with us, or never known to us,
and not accessible to us, any attempt to interpret some text
does, I would say, tacitly assume the text is the marks left
by the writing of some words which did, at some point,
originate and reside in a human mind.  If we don't assume
this, reading and interpreting and understanding text becomes
like seeing "faces" in the clouds: there are no faces there to
see, not really, it just looks like it.  (I'm here ignoring
any belief that The Gods draw pictures in the clouds for us to
see, recognise, and marvel at.  Which is close to what I think
is going on when people claim to see words and meanings and
more in today's automatically generated text.)

So, I agree we don't need to know anything about the mind that
originated some words that were then written down, and that we
are left with the text of, to be able to form rational
interpretations of this text.  Our interpretation, to be a
fair interpretation, needs no knowledge of what the original
words meant to the mind that created them, and doesn't require
any attempt to surmise this original meaning, nor to be
anything like this original meaning.  What we make of the
text, after fair interpretation, is enough.  But, I do think
our supposing, or perhaps knowing, that the text we interpret
in our own way was the result of some human mind forming and
writing words, and not some process that just generates text
we can read as if it came from some mind originated words
written down, is what makes forming an interpretation a
sensible thing to do.  If we didn't know that, and how, text
results from writing words formed to say things with, I don't
think we'd spend time trying to interpret any text.  If that
makes me a romantic, then I'm a romantic.

Your new proposal presents new difficulties for me.  This is
based upon the idea, as you state it, that

    "...  AI generated text is modelled on human minds and
     human word-output."

For me, this is a strong claim, and strong claims need strong
reasons and/or evidence to support them.  This is because I
hold to a strict definition on what makes something a model of
something: just calling, or describing something as, a model
does not make that something a model of anything, not
necessarily.  The term LLM (Large Language Model) is an
example of this kind of empty assertion.  In what way is an
LLM a model of language, and is that language in general, or
some language in particular?  Are we saying here that because,
given an input sequence of text tokens, the LLM will out put
the text token with the highest estimated probability of being
the next text token in the input sequence, that it is
therefore a model of language?  If the answer is yes, it
models what text token comes next in human written text, then
I'm off back to Mars to talk with the Martians.  If the answer
is yes, but it models something else about language, I've no
idea what, and nor have any computational linguists I've asked
this question of been able to tell me.  What I've got back are
vague explanations along the lines, text is to do with words,
words are to do with language, therefore an LLM is a model of
language.  Alice would never allow such lose kind of thinking
through the looking glass.

To have a model of something we must show, well enough, that
is satisfies the Modelling Relation across a sufficient range
of conditions for it to serve as the model we say it is, and
need it to be.  What we say is the model must be shown to
stand in a sufficient equivalence relationship with what is
modelled and how this modelled subject is situated in, and
interacts with, its natural or typical situation or context.
This requires making well judged simplifications and
idealisations of what we take to be, or judge to be, important
aspects and behaviours of the subject, and it requires well
defined mappings of chosen [observable] surface features of
the subject to surface features designed to be part of our
model, and either the same, or perhaps different, designed
surface features of the model to be mapped back to what are
chosen to be corresponding surface features of the subject.
And then, we need to verify, validate, calibrate, and test our
model, and present all this for inspection by others, before
we can properly talk of having a model.  [This idea of the
Modelling Relation is based upon the work of Robert Rosen,
1985: Anticipatory Systems — Philosophical, Mathematical &
Methodological Foundations, Pergamon Press, and is what I
teach PhDers in a course I do on model making and model using
in research.]

So, we need to ask, I'd say, does a machine built by
[automating the] digging out, from massive amounts of human
made text, huge numbers of detailed statistical relationships
between the mostly unreadable text tokens all the text is
first broken into, model well enough the mental goings on when
a person forms words to say something with, and then writes
these words down?  I would say no, it definitely doesn't.  The
statistical properties of large amounts of human made text
have no necessary relationship with the mental goings on that
forms and writes words that are meaningful to at least the
author.  Alphabets and how we use them to write down words, to
be left with text, are human made artefacts.  Certainly useful
artefacts, but they display only the shared ingenuity of human
intelligence.  They do not capture, in some understandable
way, the inner workings of human minds, and nor dos any text
written using one of these alphabets.  I know of no cognitive
science that claims that a deep statistical analysis of human
made text affords a path into the mental goings on that forms
things to say and words to say them with in the heads of
people who language.  As you have consistently pointed out,
there is no needed connection between some text and the
thoughts and goings on in the mind that first formed the words
written down that results in some text, and such a connection
is not needed to make fair interpretation of that text.
Written words do not, you and I insist, deliver into the mind
of readers the meaning the author intends their words to have.
It takes a human reader to have words again from some text --
no human reader, no words -- and it is the mind of the reader
that forms words from the text, and forms, from those words,
what they mean, or might mean, to the reader.  And, this
reader made understanding need have nothing to do with what
the words meant to the original author.  Using the dug out
statistical patterns of human made text to generate more text
does not, I would say, make this text any kind of model of
human minds, directly nor indirectly, just like the clouds
forming "faces" isn't a model of the cognitive goings on when
a person draws another person's face.

Those "sonnets recently posted" are only sonnets to you who is
able to read the generated text as being the text of written
sonnets.  There is nothing in and of the text that makes the
texts sonnets.  To have sonnets, and not just the marks on the
page we call text, needs you, someone who knows plenty about
poems and sonnets, and is thus able to read the text as a
sonnet and judge how good is the sonnet you get from your
reading of the text.  (It's not the same for me.  I am
unqualified to judge how good, or not, this text is as the
text of sonnets.)  The machine that automatically generated
this sonnet-looking text is not a sonnet writer, it's just a
text generator that uses statistical patterns found in human
made text to form new text which have the appearance of
sonnets to human readers who know what sonnets are.  The text
is 'plastic poems', and, like plastic flowers, no amount of
looking like sonnets makes them real sonnets.  Real sonnets
start with some word forming in a human mind, I romantically
want to insist.  A not unimportant detail left out of all this
talk about text is the typographical design and formatting of
the text.  There are conventional forms used to present the
text of poems that are designed to aid the reading of the text
as poems, and this too, I think matters.  But all this detail
is ripped off the texts used to "train" LLMs when it gets
converted into text tokens.  It is then put back in to the
output text by some [secrete] internal workings of ChatGPT. It
is not something generated by the LLM, which only generates
text tokens.

On a more detailed point, you say

   "...  AI can't write anything that I'm aware of that
    doesn't have some pre-existing pattern.  ..."

First, these automated text generators don't write, they
generate text, but because of the way the so called LLM (Large
Language Model) inside things like ChatGPT is used, new and
never seen before, text patterns are possible; it can, and
therefore might, compose text token patterns into forms that
are not present anywhere in the so called training data.  This
is typical of generative mechanisms, statistical or otherwise.
However, it needs you, a poet and human reader of text, to see
that this might be a new sonnet form.  ChatGPT, and other
automate text generators don't know what sonnets are, nor what
forms they may take, they only know about text tokens and
loads of statistical relationship between these text tokens
that can be found in loads of human made text.

Only humans write words, machine only generate text.  And,
only humans can read text and, by doing so, form words and
thus meanings, in their heads.  Machines don't [yet] do this.
ChatGPT does no reading, and does no word forming with which
to say things it has decide to say.


Before I stop, I would like to add here my thanks to,
agreement with, and appreciation of, Willard's call for and
(long time) support of open discussion here, and Miran
Hladnik's and Maurizio Lana's recent public support of, and
arguments for why we need open discussion.  Thank you all!

-- Tim




> On 19 Dec 2024, at 10:24, Humanist <humanist@dhhumanist.org> wrote:
>
>
>              Humanist Discussion Group, Vol. 38, No. 285.
>        Department of Digital Humanities, University of Cologne
>                      Hosted by DH-Cologne
>                       www.dhhumanist.org
>                Submit to: humanist@dhhumanist.org
>
>
>
>
>        Date: 2024-12-18 21:28:37+00:00
>        From: James Rovira <jamesrovira@gmail.com>
>        Subject: Re: [Humanist] 38.280: AI, poetry and readers
>
> Thanks for the response, Tim. I think I have an idea about how to advance
> our conversation. It may involve a different way of thinking about AI
> generated text.
>
> But first, I was referring to Derrida's 1958 Introduction to Husserl's
> Origins of Geometry. Reading it made me want to write something about the
> importance of the triangle in western philosophy, which goes back to
> Socrates. I can't find the full text online. I might have a .pdf somewhere
> I can send you.
>
> Now, here's the crux of the matter, as I understand it, for you:
>
> "But, [with AI] there was no mind involved in the generation of this text;
> there were no words written down; there was no Shakespeare forming the
> words and writing them down for us to read, and interpret, long after
> Shakespeare's mind is gone."
>
> My previous response to that was that the origin doesn't matter because we
> don't have the mind present. We both agree that a mind originates humanly
> written words, while in the case of AI there is no mind present for that
> specific arrangement of words, but my response was that at the interpretive
> end, mind is equally absent in both cases. That answer wasn't satisfactory
> to you, so you reasserted the difference that an originary, human,
> intentional mind makes for the meaning of words. That is a romantic notion
> to me. In practical terms, the work of interpretation is the same, because
> it is word-based.
>
> But, here's where I'd like to suggest a different idea: words themselves
> are the product of human minds. Patterns of words are the product of human
> minds. AI generates text that follows statistically probable patterns of
> words *in response to a human prompt, and the source text for that
> statistically probable response consists of words** already produced by
> human minds*. AI generated text is modeled on human minds and human
> word-output.
>
> So while the AI itself doesn't have a mind, the AI arrangement of text
> is *indirectly
> *the product of human minds. AI can't write anything that I'm aware of that
> doesn't have some pre-existing pattern. I'm curious what would happen if we
> asked it to invent a new form of poetry? I've innovated a couple short
> poetic forms myself.
>
> Those sonnets recently posted to this list were rather good and could have
> plausibly been written by a human. But AI didn't come up with them out of
> nowhere. I might think the human being who wrote them was a kind of tool,
> but then human beings reduce themselves to tools all of the time. And they
> were "rather good." They weren't great. They weren't self-reflective,
> self-critical, advancing the conversation. They didn't force us to
> reconsider the values being expressed. They were following conventions.
> People do that all of the time. The world is full of minor poets. AI is at
> least a competent poet.
>
> I don't believe I could have this conversation with an AI. It would need
> prompts from me, not responses. So I'm not saying AI and human beings are
> interchangable. I'm only talking about short, discreet, literary products
> such as poems.
>
> Jim R



_______________________________________________
Unsubscribe at: http://dhhumanist.org/Restricted
List posts to: humanist@dhhumanist.org
List info and archives at at: http://dhhumanist.org
Listmember interface at: http://dhhumanist.org/Restricted/
Subscribe at: http://dhhumanist.org/membership_form.php