Discussion about this post

User's avatar
Terry Underwood, PhD's avatar

My coursework in linguistics happened during the early 1970s just as transformational grammar was sprouting wings. During my doctoral work in language and literacy, most of what I found useful then and still today emerged from Michael Halliday and functional grammar. More recently, James Hudson’s word grammar rooted in default inheritance has helped me think about bot speak (his 2008 book). Charles Fillmores early work on frame theory is useful as well. LLMs operate via syntactic parsing to some degree but the real magic I think comes from the training methods. Picture a bot scanning 500 different streams of text within a functional genre (say, the discipline of history) simultaneously for recurrent linguistic patterns . For 24 hours the bot trains on millions of texts running in parallel streams and finds that historians signal levels of confidence in information depending on whether an analysis is done using primary documents vs secondary documents. Different language patterns show up in historical writing which extends beyond one sentence. The bot writes one sentence at a time, but a word selected for a slot in medias res has affordances that will impact word choices in sentences to appear later in text. It’s not really generating sentence after sentence as separate entities but as parts of larger text structure. Cognitive verbs function differently depending on subject matter. For example, analysis as a mental protocol in tge context of poetry is associated with much different words, phrases, sentences, and text structure than in, say, conducting an autopsy. Of course bots produce meaningful texts. How do I know? I can make sense of it. The difference is the bot can’t. That doesn’t make it meaningless. Here is where Charles Fillmore comes in handy. I also have found tremendous help from Otto Jespersens book from the 1920s on the philosophy of grammar. Words do have fixed meanings available to all of us and to bots. Words also have private lives inside unique utterances (Bakhtin) with centripetal force. The longer a chat goes on in terms of conversational turns, the movement in bot output becomes more attuned to the specific idiolect of the user. That’s why bots can take a muddy half formed idea and figure out what you might be trying to say.

Expand full comment
1 more comment...

No posts