Humanist Discussion Group

Humanist Archives: Dec. 10, 2024, 6:12 a.m. Humanist 38.271 - AI, poetry and readers

				
              Humanist Discussion Group, Vol. 38, No. 271.
        Department of Digital Humanities, University of Cologne
                      Hosted by DH-Cologne
                       www.dhhumanist.org
                Submit to: humanist@dhhumanist.org




        Date: 2024-12-09 10:43:54+00:00
        From: Tim Smithers <tim.smithers@cantab.net>
        Subject: Re: [Humanist] 38.265: AI, poetry and readers

Dear Jim,

Thank you, again, for engaging thoughtfully and constructively
with my wonderings and wanderings.  Your clarity helps, I
think, to talk about "generative" poems.

First.  I must admit, I do feel out on a plank here, and one
extended from a ship I'm hardly qualified to be a passenger
on, let alone knowledgable enough to crew.

I'll start walking along this plank with something you say
which I agree with, mostly, from later on in your message.

    "A poem is nothing other than a bunch of words arranged
     into lines.  If it looks like a poem, it is a poem.  It
     doesn't matter how it was produced.  ..."

Yes, I agree, a poem is "a bunch of words," but not a bunch of
text.  For me this difference is crucial.  And, it matters
what, or better said, who does the looking here, and it
matters that there is a looker doing the looking, and one that
can see that they are looking at marks on a page that stand
for words.  This is not what today's automated text generators
do; these just generate text.  Not words.  I agree that it
doesn't matter how the words of a poem are produced, but, to
me, it must be words that were produced, not just text.  So
any text produce by an automatic text generator that looks
like the text of words that form a poem which has been
written, is not a poem; it is, I would say, a de-generative
poem.

I'll try to explain why and how, in what is a [very] long
post!

The first difference I see, from the end of the plank, shows
up here.  You say ...

    "I would never argue with Tim that some kind of human
     intent is always behind the production of text.  ..."

But I would not say this.  I don't want to say this.  My
difficulty is with the word "text."  It should be, and, I
think, it needs to be, "words."  I'd agree with saying "...
some kind of human intent is always behind the production of
words."  All this so called Generative AI stuff is, I think,
clearly showing us that no intent, human or otherwise, is
needed to generate text, and text that is readable into words
by us.  But, as you know, Jim, this is not the first time
people have made text generating systems to make words from.
A good example would be David Bowie’s and Ty Roberts’
Verbasizer program of 1995, I think: try it here
<https://verbasizer.com/>.  What's different now is just a
matter scale.  And, of course, that Bowie intended to use
[some of] the generated text as their own words in a creative
way, and was, I think, particularly good at doing this.  Most
people who use automated text generators today don't do this.
They just read the generated text as if it is the result of
some words made with some intent, and either don't care that
there was no real intent, or presume that the Generative AI
systems they use actually have intentional capacities.  Which
they don't!

Next, I'll dive in here, with your claim:

    "So the meaning of the poem resides in the words on the
     page, not in the poet who produced them.  As a result,
     there's no reason why an AI generated poem can't be
     subject to the same kind of analysis as a humanly written
     poem."

I agree with your first words, "So the meaning of the poem
resides in the words ...," and, I agree with your last words
of this first sentence, "...  not in the poet who produced
them."  But, I don't agree with what you put between these
words: "on the page."  I'll get back to your second sentence
here after dealing with this first difference.

The words of the poem are not, I want to say, on the page.
They are in the heads of the human readers reading the text of
the poem which _is_ on the page, and they were, at some time,
in the head of the poet who formed them and then wrote them
down, as text on a page.  I know this is an unusual
distinction, and one we don't usually make.  But it is, I
think, a distinction we need when it comes to understanding
today's so called Generative AI systems, and, in particular,
automated text generation machines.

Words don't exist on the page, nor in the compression waves of
the air that allow us to hear the words we or others speak.  I
want to say words, and thus whatever the words may mean to
people, only exist in the heads of human writers and readers,
and speakers and listeners.  [Words perhaps also exist in the
heads of some other animals, but I'll leave this question
behind here.]  Speaking and listening, and writing and
reading, and signing and seeing, and touching and feeling, (as
in braille), are thus ways we have of attempting to have the
words existing in one [human] head, also having an existence
in another [human] head, or lots of other heads.  I call this
languaging.  In the richness we experience this languaging, it
is a unique human quality.  And it needs, I think, humans for
it to truly happen, so far, at least.

Seen this way, I think, means that to have any words you need
something with a capacity to form words and combinations of
words, to think with, and to say things with, perhaps just to
itself, but more usually to other things that also have word
forming and word using capacities, together with sufficient
hearing, seeing, feeling, and thereby reading and
understanding capacities.  And, as you've probably
anticipated, I don't think Generative AI systems like Claude
3.5 Sonnet and ChatGPT have any capacity to form words which
they then write out as text for us to then read, so as to have
in our heads the words they first formed in order to say
something they intended to say.  These generative systems
don't have any capacities for hearing, seeing, feeling,
reading, and understanding words, nor any intentions.  They
only generate the marks of text, which, yes, we, with all our
languaging capacities, can read and understand things from.
But, no words have existed anywhere in the text generation
processes used by these systems.  Text generating Generative
AI systems work only with very large numerical vector
representations of text-tokens [marks that make text] and
pre-built estimated probabilities for a great many [very]
detailed statistical text-token patterns found in the texts
left over from enormous amounts of human writing.

If we treat the text generated by things like Claude 3.5
Sonnet and ChatGPT as words, then, it seems to me, we are
saying, at least tacitly (!), that these systems _do_ have
word forming and using capacities, and thus that these
systems, like us humans, traffic in meanings, using all the
different languaging capacities this entails.  And, some
people, I think, will want to say they do have word forming
and word using capacities like we humans have.  The only way
of defending a position that says these machines do not have
anything like human languaging intelligence, and the
capacities that give us this intelligence, is to hold that
there is a distinction between words and text, albeit a
distinction we don't usually use, nor needed much before.

Here I'll return to the second sentence of your words I
started with above:

    "As a result, there's no reason why an AI generated poem
     can't be subject to the same kind of analysis as a
     humanly written poem."

First, I want to insist that whatever it might look like to us
languaging beings, what we get from these AI text generators
is not what I am prepared to call a poem, not a real one, at
least.  It's a plastic poem, as in plastic flowers.  It may
look like the text of a poem but it isn't the text of a real
one, it's just something "made of plastic" to look like a poem
...  because the system that generated this looks-like-a-poem
text has plenty of statistics about what the text of written
down poems look like as marks on a page, but not any notion of
what words are, what meaning is, nor of what a poem, as we
humans understand poems to be, nor any knowledge or
understanding of what languaging is.  These Generative AI
systems have no real languaging capacities, just impressively
large matrix arithmetic capacities for crunching loads of very
large numerical vector representations of text-tokens and the
estimated probabilities of many detailed relationships between
text tokens.

But, yes, we language capacitated beings can indeed do the
same kind of analysis on the text generated by these machines
as we do on the text of humanly written poems -- that is,
those of us qualified to do this can.  But, when we do this,
we first need to construct words from the text we read, and,
from these words, form meanings, which are needed in our
analysis efforts.  But, if we do this, we should, I think,
acknowledge that the text we read, and form words from, and
thus meanings from, was not the result of any writing of words
by a poet: no words pre-existed the text we read.  So, all the
word and meaning forming we do in our analysis are the first
words, and meanings, formed.  This means, I would say, that
this kind of analysis is different from the analysis of a real
poem written with intent by a poet.  In this analysis of some
text as if it is text from the forming and writing of a poem,
necessarily involves, perhaps tacitly, the forming of a poem
to analyse.  It is you who do this analysis of automatically
generated text, that makes the poem you analyse, and you have
to do this to have any poem to analyse; to have an words, and
thus meanings, to work with.  Real poems formed and written by
a poet come with words.  Automatically generated text does not
come with words, only text.  You have to supply the words,
which makes you the poet here, I would say.

I think this difference matters because the machine that
generated the text we take to be the text of the written words
of a poem, has none of the capacities needed to do even a poor
version of this analysis.  But that won't stop it generating
text that looks like the outcome of some analysis, if we
prompt it to generate text in this way.  These Generative AI
systems have a seemingly unlimited capacity for a kind of
artificial pretending.  They will generate no end of text that
looks a lot like text that resulted from someone writing
words.  But no words were ever made.

So, suppose we do such an analysis of the text generated by
some machine, and thus treat this generate text as if it is
the text of a poem that some one formed and wrote down.  What
will we get?  Well, the analysis of a poem, of course.  But it
is only the analysis of a poem because we, who did the
analysis, made the generated text into a poem, not because the
text we chose to analyse is the text of a poem that was made
and written down by someone.  There was no poem making and
writing down in the generating of the text we analysed.  We
just chose to treat the text as if there was.  Any poem that
results is made by us who do this analysis.  There is no poet
here.  This is, I would say, analogous to someone doing a
botanical analysis of an artificial flower, just as they might
do when studying a real flower.  Nothing stops the botanist
from doing this, except, perhaps, honesty, but no amount of
botanical analysis will turn the artificial flower into a real
flower; there wasn't a flower there to start with.  The same
goes for the automatically generated text that, to us, looks
like the text of a written down poem; no amount of analysis of
this text will turn it into a poem a poet wrote, no matter
what this analysis might uncover in the words we form from a
reading of the text, words, I repeat, which were not there
until we did the reading, word forming, and analysis.

What we get, and can only get, from automatic text generators
is text that has never been words in anybody's head.  And
that, I think, makes a big difference, an important
difference, a difference we should always acknowledge, and not
ignore, or forget, of seek to hide, or pretend doesn't matter.
Not unless, that is, we can confidently show how all that goes
on in our heads when we language is something rather like
matrix arithmetic using very large numerical vectors shown to
reliably represent, and only represent, text-tokens, and the
estimated probabilities of how they relate, and, of course,
that we can also explain how all this statistical matrix
arithmetic came to be in our heads in the first place.  [All
of which just happens to be something the Connectionists
don't' appear to feel a need to explain.  But why would they
when you just need to accept, and then run with, the
Connectionist dogma: build lots of connected things you call
"neurones" then it must be intelligent, and intelligent like
us.]

Before stopping I want to say more about this "strange
distinction" I make here: it's text we have on the page, not
words; words only exist in the heads of the author and the
reader(s): the author who wrote the words of their poem, and
the reader who forms the same words from their reading of the
text, but who does not necessarily have the same meanings from
these words as the author had when they wrote them, or, even
have any meaning at all, perhaps.

This distinction is not really all that strange, I'd say.
Take a musical score, for example, for the last movement of
Beethoven's 7th symphony, say.  We don't usually say the score
is the music, that the music is on the page.  It's marks that
are on a page, not the music.  If it was the music on the
page, we'd just need to give people printed copies of this
score and they'd have the music.  But they don't.  Performance
is needed; performance by suitably qualified, skilled, and
practiced, musicians.  And, just like text can be read into
words that mean different things to the different people who
read the same text, so a musical score can be, and just about
always is, performed differently, and this difference matters.
These different performances, interpretations as we call them,
are taken to be a kind of exploration of what the written
score can be used to make.  And, mostly, we take it that the
composer who first wrote the score intended it to be used in
this way, albeit with some indications and occasionally
further instructions, on how they want their work to be
performed, but not interpreted: composer can't instruct this,
and don't try to, of course.

Similarly, a Laban notation score -- using the notion invented
by Rudolf von Laban to record and thereby analyse human
movements and motions, and, in particular, dance movements and
motions -- is not the dance, it's marks on a page.  Once
again, performance is needed, and interpretation is needed, to
have the real [notated] dance.

Semiotics.  This is all semiotics.  Which depends upon a clear
distinction between the sign and the signed or signified.  The
art of semiotics lies in accurately identifying the signified
from the sign, and thus what a particular sign is intended to
mean, or can be reasonably understood to mean, and, more
interestingly perhaps, what some combination of signs may be
reasonably taken to mean.  This art includes the need to
identify when we have a empty sign, something we have been
told, perhaps, or believe, to be a sign, but which isn't a
sign; it doesn't signify anything, but just looks like it
does.  Empty signs are what automatic text generators
generate, but they make these empty signs to look so like
content-full signs that we nearly automatically read them as
such, and make words from our reading of these signs, so that
the combinations of the words we form say things to us, but
all these words and what they say to us are in our heads, and
only in our heads, and are all made by us; they do not come
from the page.  Only text is on the page, only signs, and, in
this case, only empty text is there.  It takes us, and our
languaging capacities and skills and imagination to make words
in our head from this empty text, or what we might call
word-less text.  And word-less text results in de-generative
poems, I think.  Not real poems.

I've used the phrase "in the head" lots here, to say words
only exist in the heads of people.  This is, of course, a
short-hand for all the cognitive goings on when we humans
language, and about which I don't want to say much here.  We
know some about all this, but not everything, and, perhaps not
as much as we like to think we know.  These cognitive goings
on are the proper realm of a wide variety of kinds of
researchers, in the Cognitive Sciences, Neurosciences,
Psychology, Anthropology, Linguistics, Philosophy of Mind and
Language, including Epistemology, AI [the research discipline]
and, of course, Literature and Literary Studies in all its
forms, to name the most obvious ones, but probably not all.
I've treated all this as being wrapped up in a Blackbox I call
the head of a person.

Still, there is one thing I want to say something about, of
what goes on inside these Blackboxes; about the nature of
words as they exist in our heads.  They are not simple things;
they are not mere simple symbols, or signs.  They are, it
seems to me, and must be, complex things; things which have
important properties and qualities of their own, and which
have internal state and structure, in order to work as what we
call words, to think with and to say things with, and make
poems from, as we do.

For one thing, words, and thus what we take them to mean, or
use them to mean, often have a fogginess, or mistiness, that
text does not, and must not, have.  Text, to work properly,
needs to be reliably readable; we need to be able to
accurately read the text so we can re-construct the words that
were written and which therefore left the marks on the page we
read them from.  But this fogginess or mistiness is important;
we need what we want our words to mean or what we take the
words we reconstruct from text to mean, to be a little
undefined, mouldable, flexible, slippery, suggestive, ...  It
is, I think, this lack of definitiveness of words that makes
languaging possible.  If we had to be definitive in every word
we used, and in every combination of words we form and use,
languaging would fail so often we wouldn't use it much; errors
in speaking or writing the definitive words, and errors in
(re)constructing the definitive words from what we hear or
read, and thus errors in the definitive meanings of the words,
would occur too often for our word based interactions, ie,
languaging, to be viable.

This is why programming computers, which looks rather like a
kind of languaging, and which we often talk of this way, is so
different from human natural languaging.  The "words" and
constructions of programming languages, and programs, do need
to have definitive denotations for us to be able to use them
to build programs computers can execute.  And, as a result, we
often make mistakes in the forming of these "words" and
correct combinations of them.  We call these mistakes "bugs,"
a word which has lots of useful slipperiness and flexibility,
as demonstrated in this use here!

How is languaging between us viable is also a question I don't
want to get into here; it's another long story, I think.  But,
the way I see it, and have approached it -- in designing and
implementing an artificial language for people to use to
interact with a robot; a "spoken" language built from sounded
musical notes which the robot used to say [simple and limited]
things to people, and which it could hear and understand
itself say, and which people could [learn to] use to say
[simple, and limited] things to the robot -- is that
languaging is necessarily a situated activity which requires
enough shared common situation and conditions, which therefore
makes it better treated as a distributed cognition phenomenon,
and not just a "in side the head" phenomenon, as I have
indicated here with my Blackbox treatment.

The nature and structure of words as they exist in our heads,
and as we use them in our thinking and languaging, are, I
think, less studied than they need to be for a proper
understanding of how languaging works.  Words, in our heads,
are not just simple symbols for meanings which we select and
combine to capture and express what we want to say and to mean
-- though this is an impression we may get from looking at
what linguists say.  I don't see it as being nearly as simple
as this.  And, I think this from the work I, and many others
in Symbol Processing AI, have done on building computational
systems that do know, understand, and reason about certain
things.  Yes, this does involve processing symbols, but these
symbols often have to be more than simply tokens that signify
well defined, and fixed, things.  The symbols that get
processed by the computational mechanisms we have to design
and implement to get any actual knowing, understanding, and
reasoning to happen, have to have complex internal modifiable
structures and variables, which are, themselves processed,
manipulated, and changed, by the computations of the the
knowing, understanding, and reasoning, we build our systems to
do.  This is a reality of so called Symbol Processing AI often
missed, including by people who do this kind of AI work, but
who don't reflect much on what they have built and why it does
what it does, if, indeed, it actually does what they say it
does.  John Searle, and The Chinese Room Argument they
presented in 1980, suffers from this failure to understand the
real nature of symbols in AI systems, and that they are not,
and can't just be, simple signs for things to have some
intelligent behaviour.

But, notice, John Searle was right about the processing of
[simple] symbols not being enough for real understanding, but
this is all we have in automated text generation systems;
loads of processing of numerically represented text-tokens, a
kind of large but nonetheless simple symbol, but often dressed
up to look like it has lots of clothes on.  There are no
clothes, I think.

If we accept automatically generated text to be the same as
the intentional forming and writing of words we will let these
so called AI systems win, and win without even being
intelligent.  It is, as always, up to us humans to avoid
dehumanising ourselves.

There's more to say, but I'll stop here, and thank you, Jim,
and anybody else who has made it down to here, for reading all
this text, and for all your efforts of making words in your
heads from it all.  I'll be interested to see if the words you
make from this text mean anything to any of you all?

-- Tim


PS: Willard.  I hope I've not broken the record for the most
amount of text ever posted to the Humanist list?




> On 2 Dec 2024, at 08:36, Humanist <humanist@dhhumanist.org> wrote:
>
>
>              Humanist Discussion Group, Vol. 38, No. 265.
>        Department of Digital Humanities, University of Cologne
>                      Hosted by DH-Cologne
>                       www.dhhumanist.org
>                Submit to: humanist@dhhumanist.org
>
>
>
>
>        Date: 2024-12-01 19:46:47+00:00
>        From: James Rovira <jamesrovira@gmail.com>
>        Subject: Re: [Humanist] 38.263: AI, poetry and readers
>
> Once again, Tim prompts me to respond. My own background includes work in
> poetry: half of my dissertation was about a poet, William Blake, and I've
> taught both formal and free verse at the undergraduate and graduate levels,
> and I've supervised creative theses that were collections of poetry,
> besides having published a number of poems myself. That's my background.
>
> I would never argue with Tim that some kind of human intent is always
> behind the production of text. Most literary scholarship disregards intent
> as inaccessible or irrelevant, however. The fact that a poem, even one
> generated by a prompt fed to a computer, is an intentionally made product
> usually doesn't help us understand it. If we do enough biographical and
> historical research, and if we even know the author and the time and
> conditions of the writing of the poem, that will help us understand the
> author at the moment of writing the poem, but not what the poem in any
> absolute sense *means.* T.S. Eliot's essay "Tradition and the Individual
> Talent" is a good reflection on the relationship of poem to author from the
> author's point of view:
> <https://www.poetryfoundation.org/articles/69400/tradition-and-the-individual-
> talent>.
>
> So the meaning of the poem resides in the words on the page, not in the
> poet who produced them. As a result, there's no reason why an AI generated
> poem can't be subject to the same kind of analysis as a humanly written
> poem.
>
> Before I move on, I'd like to define a couple of terms: Formal verse
> follows predetermined rhyme and metrical patterns (at least one or the
> other, often both) while free verse does not. Free verse may follow such
> patterns, but they are invented by the poet for the poem.
>
> What I'm going to say is about formal verse first: I love the invocation of
> Lewis Carroll at the end of Tim's post, who was a mathematician. I believe
> that background in mathematics contributed to the excellence of his verse.
> I frequently compare a line of poetry to a bar of music (and I'm hardly
> original in doing so), and the mathematical qualities of a bar of music is
> hardly contested. Neither should the mathematical qualities of a line of
> formal verse. Many kinds of formal verse forms consist of a fixed pattern
> of stressed and unstressed syllables often combined with a fixed rhyme
> scheme.
>
> As a result, the tendency, or danger, is for the poetic form to turn the
> poet into a computer rather than for the poet to own the form as meaningful
> self-expression. As a result, there's no reason that a computer couldn't
> produce something that at least fits the pattern, even a conceptual pattern
> that requires a certain kind of rhetorical gesture around, say, line 9 of a
> sonnet, or the last two lines of an Elizabethan sonnet. As a result,
> there's no reason why the computer generated poem couldn't be subject to
> analysis *as if* a human being had written it. It is still made up of human
> words.
>
> A poem written by a computer as an intentionally devised means of authentic
> self expression would probably be in binary, or in electrical impulses, or
> some kind of representation of them. If computers really wrote their own
> poetry, only a programmer could read it, or an engineer with an amprobe and
> a volt meter.
>
> Just to be clear, I do not believe this will ever happen.
>
> HOWEVER, once you begin writing *parodies*, you turn even a free verse
> source poem *into a formal verse form*. It suddenly has fixed parameters,
> even for specific conceptual moves and rhetorical devices. And above all
> else, a good parody is an inherently bad poem imitating a good one, and the
> better it is at imitation *while still being a bad poem*, the better a
> parody it is. So I think AI is especially suited for writing parody poems.
> I only find it sad that it can't laugh along with us.
>
> What we're circling around without discussing, however, is what a poem
> fundamentally *is*.
>
> I think that many people believe poetry is a deeply meaningful and symbolic
> means of human self-expression.
>
> Every bad human poet who has ever lived thought just that. Douglas Adams
> thought just that in a very funny way. Poets who focus on their own
> feelings and ascribe inherent worth to them while they are writing their
> poems massively suck, both as human beings and as poets.
>
> Whenever poetry rises to meet that criteria, it is only as a by-product.
>
> A poem is nothing other than a bunch of words arranged into lines. If it
> looks like a poem, it is a poem. It doesn't matter how it was produced.
> That is the hard and cruel truth. Eliot faced that truth; Milton faced it.
> The poem doesn't care about the poet. The poem does not express the poet.
> The poem expresses itself. The poem *leaves* the poet and then does
> whatever it will, and if it's a really good poem, it means a lot more than
> the poet ever hoped to express. So good poets really only care about the
> words on the page. "Care" is an emotion, and there is probably another
> emotion behind that care, but the focus of their attention is on the artful
> arrangement of words on a page or screen. If the focus is anywhere else,
> the poem is bad.
>
> A poem is not soft and meaningful and of deep feeling. A poem is hard and
> cruel and meets you only on its own terms. It does not know you or care
> about you. It does not feel: it may provoke feelings, but it *doesn't care
> about your feelings*. It simply is. If a poem became sentient, it'd be a
> real jerk. So good poets very carefully *construct*, but not *write*,
> poems. A poem is a kind of emotional/conceptual program that, once arranged
> by a human poet, works outside of its own parameters.
>
> So I disagree with this claim: "But, looking-like a poem is not a
> sufficient, nor even a necessary, condition of being a poem."
>
> I think that is the only condition of being a poem. I think I would agree
> with Tim's claim if we were describing a *great* poem. I don't think AI can
> do that, only because it would need to understand nuance incorporating too
> much outside the poem, beyond syntactical patterns.
>
> But most human beings who write poetry actually write bad poetry. Some
> write decent poetry, even good poetry. Most will never write great poetry
> ever, at all, their entire lives. But if they're lucky, they might write a
> few great lines here and there. That makes all the rest of it worth it.
>
> Jim R
>
> On Sun, Dec 1, 2024 at 3:01 AM Humanist <humanist@dhhumanist.org> wrote:
>
>>
>> To be clear.  It takes a person to decide to present some text
>> as a poem, whether they wrote the words, and are thus left
>> with the text marks of their words, or got the text generated
>> by some gee whiz machine.  I bet if the generated text didn't
>> look somehow like a "poem" to this person, they wouldn't
>> present the text as a "poem."  But, looking-like a poem is not
>> a sufficient, nor even a necessary, condition of being a poem.
>> Trying to make "poems" using text generators thus easily
>> results in what I call 'Artificial Flower' "poems" -- texts
>> which may look like poems to many people, but ain't poems, not
>> real ones -- and, cannot result in things I would call
>> 'Artificial Light' poems -- marks that are artificially made
>> somehow to be used to recreate words well chosen to form real
>> poems, and which may not look like any poems we've seen
>> before, but, which, upon genuine analysis and interpretation,
>> can be shown to be, and thus appreciated as, real poems,
>> perhaps in new and exciting ways.
>>
>> -- Tim



_______________________________________________
Unsubscribe at: http://dhhumanist.org/Restricted
List posts to: humanist@dhhumanist.org
List info and archives at at: http://dhhumanist.org
Listmember interface at: http://dhhumanist.org/Restricted/
Subscribe at: http://dhhumanist.org/membership_form.php