Template:M intro work Large Learning Model: Difference between revisions

Line 72: Line 72:
{{quote|“What an astonishing thing a book is. It’s a flat object made from a tree with flexible parts on which are imprinted lots of funny dark squiggles. But one glance at it, and you’re inside the mind of another person, maybe somebody dead for thousands of years. Across the millennia, an author is speaking clearly and silently inside your head, directly to you.  
{{quote|“What an astonishing thing a book is. It’s a flat object made from a tree with flexible parts on which are imprinted lots of funny dark squiggles. But one glance at it, and you’re inside the mind of another person, maybe somebody dead for thousands of years. Across the millennia, an author is speaking clearly and silently inside your head, directly to you.  


Writing is perhaps the greatest of human inventions, binding together people who never knew each other, citizens of distant epochs. Books break the shackles of time. A book is proof that humans are capable of working magic.”
''Writing'' is perhaps the greatest of human inventions, binding together people who never knew each other, citizens of distant epochs. Books break the shackles of time. A book is proof that humans are capable of working magic.”
:— Carl Sagan, ''Cosmos''
:— Carl Sagan, ''Cosmos''
“I think you underestimate the power of ''reading'', Professor Sagan.”
“I think you underestimate the power of ''reading'', Professor Sagan.”
:—[[Jolly Contrarian|JC]]}}
:—[[Jolly Contrarian|JC]]}}
Fittingly, the first [[chatbot]] was a designed as a parlour trick. In 1966 Joseph Weizenbaum, a computer scientist at MIT created “[[ELIZA]]” to explore communication between humans and machines. [[ELIZA]] used pattern matching and substitution techniques to generate realistic conversations. By today’s standards, [[ELIZA]] was rudimentary, simply regurgitating whatever was typed into it, reformatted as an open-ended statement or question, thereby inviting further input. As a session continued, the user’s answers became more specific and elaborate, allowing [[ELIZA]] to seem ever more perceptive in its responses.  
=====Theres’ a hole in my model, ELIZA=====
Fittingly, the first [[chatbot]] was a designed as a parlour trick. In 1966 Joseph Weizenbaum, a computer scientist at MIT created the programme “[[ELIZA]]” to explore communication between humans and machines. [[ELIZA]] used pattern matching and substitution techniques to generate realistic conversations. You can try it out [https://web.njit.edu/~ronkowit/eliza.html here].


Even though [[ELIZA]] was a basic “keepy uppy” machine, it proved surprisingly addictive — even to those who knew how it worked. Weizenbaum was famously shocked how easily people, including his own secretary, were prepared to believe [[ELIZA]] “understood” them and contributed meaningfully to the interaction.  
By today’s standards, [[ELIZA]] was rudimentary, simply regurgitating whatever was typed into it, reformatted as an open-ended statement or question, thereby inviting further input.
 
Even though [[ELIZA]] was a basic “keepy uppy” machine, it proved surprisingly addictive. Weizenbaum was famously shocked how easily people were prepared to believe [[ELIZA]] “understood” them and contributed meaningfully to the interaction. When trying it out, Weizenbaum’s own secretary, who had watched him build the programme for months and knew how it worked, shooed him out of the room so she could have privacy with the machine.  


This is, of course, how all “mind-reading” works: with the right kind of clever questions, the conjurer extracts from the subject herself all the information she needs to create the illusion.  
This is, of course, how all “mind-reading” works: with the right kind of clever questions, the conjurer extracts from the subject herself all the information she needs to create the illusion.  
=====LLMs as conjuring tricks=====
[[LLM]]s work the same way. Like all good conjuring tricks, [[generative AI]] relies on misdirection: its singular genius is that it lets us misdirect ''ourselves'', into wilfully suspending disbelief, never noticing who is doing the creative heavy lifting needed to turn machine-made screed into magic: ''we are''. We are neuro-linguistically programming ''ourselves'' to be wowed by LLMs when the clever part is really happening inside our own heads.


[[LLM]]s work the same way. Like all good conjuring tricks, [[generative AI]] relies on misdirection: its singular genius is that it lets us misdirect ''ourselves'', into wilfully suspending disbelief, never noticing who is doing the creative heavy lifting needed to turn machine-made screed into magic: ''we are''. We are neuro-linguistically programming ''ourselves'' to be wowed by LLMs.
By writing prompts, we create our own expectation of what we will see. When the pattern-matching machine produces something roughly like that, we use our own imaginations to backfill, frame, filter, correct, boost, render, sharpen and polish the output into what we ''wanted'' to see. We ''construe'' that output as commensurate we can with our original instructions.  
 
Yet again, we subordinate ourselves to suit the convenience of the machines. We really need to have some self-respect and kick this habit.


By writing prompts, we create our own expectation of what we will see. When the pattern-matching machine produces something roughly like it, we then use our own imaginations to backfill, frame, filter, correct, boost, sharpen and polish the output into what we ''want'' to see. We render that output as commensurate we can with our original instructions.  
When we say, “fetch me a tennis racquet”, and the machine comes back with something more like a lacrosse stick, we are far more impressed than we would be had a human done the same thing: we would think such a human were a bit dim. But with [[generative AI]] we don’t, at first, even notice we are not getting what we asked for. We might think, “oh, nice try!” or “well, that will do,” or perhaps, “ok, computer: try again, but make the basket bigger, the handle shorter, and tighten up the net.” We can iterate this way until we have what we want — though note all the intelligence is coming from ''us'' — or we could just Google for a conventional photo of a tennis racquet. As the novelty wears off, that is what more and more people will do.


When we say, “fetch me a tennis racquet”, and the machine comes back with something more like a lacrosse stick, we are far more impressed than we would be had a human done the same thing. We would think the human a bit dim. But with [[generative AI]] we don’t, at first, even notice we are not getting what we asked for. We might think, “oh, that will do,” or perhaps, “ok, computer: try again, but make the basket bigger, the handle shorter, and tighten up the net.” We can iterate this way until we have what we want — or we could just use a conventional photo of a tennis racquet.
First impressions can be stunning, but the second look reveals an absurdist symphony. AI image generation famously struggles with hands, eyes and logical three-dimensional architecture. It is just as true of text prompts: on closer inspection we see the countless minute logical ''cul-de-sacs'', bad guesses and ''non sequiturs'' from which the clever story miraculously [[Emergence|emerge]]s. (To be sure, many human authors write in logical ''cul-de-sacs'', bad guesses and ''non sequiturs'', but that is another story.) Again, where is the magic in this process? ''Inside the [[meatware]]''.


AI-generated image generation famously struggles with hands, eyes and logical three-dimensional architecture. First impressions can be stunning, but the second look reveals an absurdist symphony. It is just as true of text prompts: on close inspection we can see the countless minute logical ''cul-de-sacs'' and two-footed hacks from which the clever story miraculously [[Emergence|emerge]]s. (To be sure, many human authors write in logical ''cul-de-sacs'' and two-footed hacks, but that is another story.)
Either way, the novelty soon palls and, as we persevere, we begin to see more and more of the magician’s wires. The first couple of passes are great: they get from zero to 0.5 of what we wanted. But the marginal improvement in each following round diminishes, as the machine reaches asymptotically towards an upper capability in “producing what you had in mind”, which we estimate, unscientifically, as about 75% of it.  


Either way, the novelty soon palls and, as we persevere, we begin to see the magician’s wires. We get a sense of how the model goes about what it does: its familiar tropes and tics and persistent ways of doing things which are never quite what you have in mind. The piquant surprise at what it produces dampens at each go-round, eventually settling into an [[Entropy|entropic]] and vaguely dissatisfying quotidian.  
As we go, we get a sense of how the model works: its familiar tropes, tics and consistent ways of doing things which are never quite what you have in mind. The piquant surprise at what it produces dampens at each go-round, eventually settling into an [[Entropy|entropic]] and vaguely dissatisfying quotidian.  


In this way, the appeal of iterating a targeted work product with a random pattern-matcher soon loses its lustre. The first couple of passes are great: they get from zero to 0.5. But the marginal improvement in each following round diminishes, as the machine reaches asymptotically towards its upper capability in producing what you had in mind, which we estimate unscientifically as about 75% of it.  
Now, as [[generative AI]] improves — assuming it ''does'' improve: there are some indications it may not; see below — the threshold of its “expectation-meeting capability” may move towards 100% but will never quite get there. And, again, don’t underestimate how important the meatware is in that refining process: many of its improvements will be down to how we learn to better to frame our queries. “Prompt-engineering” becomes the real skill, rather than the dumb, parallel pattern-matching process that responds to it.  


Now, as [[generative AI]] improves towards 100 — assuming it does improve: there are some indications it may not; see below — that threshold may move but it will never get to 100. In the mean time, as each successive round takes more time and bears less fruit, mortal enthusiasm and patience with the LLM will have long-since waned: well before the [[Singularity]] arrives.
Ours is the skill going in, and ours is the skill construing the output. The machine just does the boring bit. But this is what machines have ''always'' done: the bits that require strength, speed, reliability, and economy. ''Not ingenuity''. This is the basic proposition of mechanising. Humans have been mechanising things since someone invented the wheel.  


And many improvements we will see will largely be in the [[meatware]]: as we refine and elaborate our queries; we learn how better to frame our queries, and “prompt-engineering” becomes the skill, rather than the dumb, parallel pattern-matching process that responds to it. Ours is the skill going in, and ours is the skill construing the output. The machine just does the boring bit. But this is all machines have ''ever'' done. This is the basic proposition of mechanising.
=====LLMs and literary theory=====
{{quote|
''POLONIUS'': What do you read, my lord? <br>
''HAMLET'': Words, words, words. <br>
''POLONIUS'':  What is the matter, my lord? <br>
''HAMLET'':  Between who? <br>
''POLONIUS'':  I mean the matter that you read, my lord. <br>
:—''Hamlet'', II, ii
}}
Now, in all kinds of literature ''bar one'', “reading” is where the real magic happens. ''Construal''. It is the [[Emergent|emergent]] creative act and community consensus that renders ''Hamlet'' a timeless cultural leviathan and {{br|Dracula: The Undead}} forgettable pap.<ref>Maybe not ''that'' forgettable, come to think of it: it has stayed with me 15 years, after all.</ref> A literary work may start with the text, but it stays there barely a moment. The “meaning” of literature is necessarily personal to the reader: it lives between our ears, and within the cultural milieu that interconnects the reading population over the generations.


So in all kinds of literature ''bar one'', “construal” is where the real magic happens. It is the [[Emergent|emergent]] creative act and community consensus that renders ''King Lear'' a timeless cultural leviathan and {{br|Dracula: The Undead}} forgettable pap.<ref>Maybe not ''that'' forgettable, come to think of it: it has stayed with me 15 years, after all.</ref> A literary work may start with the text, but it stays there barely a moment. The “meaning” of literature is necessarily personal to the reader: it lives between our ears, and within the cultural milieu that interconnects the reading population.<ref>Call me [[post-modern]] — go on, do — but I don’t hold with [[Carl Sagan]]’s idea that a book teleports its author “inside our heads”. That would be to equate reading with symbol-processing. It absolutely isn’t, and that metaphor gravely underestimates the human brain when in construction mode. </ref>
Call him [[post-modern]] — go on, do — but the [[JC]] doesn’t hold with [[Carl Sagan]]’s idea that a book teleports its author “inside our heads”. That would be to equate reading with symbol-processing. It absolutely isn’t, and that metaphor gravely underestimates the human brain when in construction mode. Nor, by the way, is this in any way to diminish Shakespeare’s towering genius, but rather to observe that, however impossibly brilliant it is, it is swamped by the flood of exposition, analysis, interpretation, re-rendition and performance, that has gone on since he published it, not to mention reader’s own role in “construction”.


“Construal” and “construction” are interchangeable in this sense: over time that cultural milieu takes the received corpus of literature and, literally, ''constructs'' it into edifices its authors can have scarce have imagined. ''Hamlet'' speaks, still, to the social and human dilemmas of the twenty-first century in ways Shakespeare cannot possibly have contemplated.<ref>A bit ironic that Microsoft should call its chatbot “Bard”, of all things.</ref> to be clear: the reader of literature is no symbol processor, decrypting text to reveal a one-to-one scale assembly of the content of the author’s head. Literature is not an instruction manual.
“Construal” and “construction” are interchangeable in this sense: over time that cultural milieu takes the received corpus of literature and, literally, ''constructs'' it into edifices its authors can scarce have imagined. ''Hamlet'' speaks, still, to the social and human dilemmas of the twenty-first century in ways Shakespeare cannot have contemplated.<ref>A bit ironic that Microsoft should call its chatbot “Bard”, of all things.</ref> to be clear: a reader is no simple “symbol processor”, decrypting text to reveal a one-to-one assembly of the author’s intention in her own head. Literature is no instruction manual, recipe nor a computer programme.


Now there is one kind of “literature” where the last thing the writer wants is for the reader use her imagination to fill in holes in the meaning. Where clarity of authorial intention is paramount; where communicating and understanding ''purpose'' is the sole priority: ''legal'' literature.  
Now there is one kind of “literature” that is like a computer programme: where the ''last'' thing the writer wants is for the reader use her imagination, to construct a meaning: where clarity of authorial intention is paramount; where communicating and understanding ''purpose'' is the sole priority: ''legal'' literature.  


The ''last'' thing a legal drafter wants is to cede interpretative control to the reader. Rather, she seeks to squash all opportunities for improvisation presented by creative ambiguity. This is why, just as there are no atheists in foxholes, [[there are no metaphors in a trust deed|there are no metaphors in a Trust Deed]].  
Rather than ceding interpretative control to the reader, a legal drafter seeks to squash all opportunities for improvisation and stomp out all ambiguity. Just as there are no atheists in foxholes, [[there are no metaphors in a trust deed|there are no metaphors in a Trust Deed]].  


Legal drafting seeks to be as [[finite]] as it can be. It strives do to readers what code does to hardware: to reduce them to mere symbol-processing machines, extracting the author’s single incontrovertible meaning. To leaves as little room as possible for interpretation. That there is such a living to be made conducting commercial litigation demonstrates how hard this is.
Legal drafting seeks to be as [[finite]] as it can be. It strives do to readers what code does to hardware: to reduce them to mere symbol-processing machines, extracting the author’s single incontrovertible meaning. But, in a natural language that is constructed out of dead metaphors, this is very, very hard to do. That there is such a living to be made conducting commercial litigation shows that.


This is one reason why [[legalese]] is so laboured. It is designed to chase down all blind alleys, previsualise all phantoms and prescribe outcomes for all logical possibilities. To do so, it removes all ambiguity and renders the text as mechanical, precise and reliable as it can be. ''[[There are no metaphors in a trust deed]]''. Where normal literature favours possibility over certainty, legal language bestows [[certainty]] at the cost of [[possibility]], and to hell with literary style and elegance.
It is one reason why [[legalese]] is so laboured. It is compelled to chase down all blind alleys, previsualise all phantoms and prescribe outcomes for all logical possibilities. To remove all possible ambiguity and render the text as mechanical, precise and reliable as it can be. ''[[There are no metaphors in a trust deed]]''. Where normal literature favours possibility over certainty, legal language bestows [[certainty]] at the cost of [[possibility]], and to hell with literary style and elegance.


Legal language is, in [[James Carse]]’s sense, ''finite''. Literature is ''[[Finite and Infinite Games|infinite]]''.
Where literary language is, in [[James Carse]]’s sense, ''[[Finite and Infinite Games|infinite]]'', legal language is ''finite''.


Now: the punchline. Given how integral the reader and her cultural baggage are to the creative act in ''normal'' literature, we can see how, in that domain, a [[large learning model]], which spits out text ripe with possibilities, begging for someone to “construct” it, is a feasible model: to move from a model where ''most'' of the creative work is done by the reader to one where ''all'' of it is, is no great step.  
Now: the punchline. Given how integral the reader and her cultural baggage are to the creative act in ''normal'' literature, we can see how, in that domain, a [[large learning model]], which spits out text ripe with interpretative possibilities, begging for someone to “construct” it, is a feasible model for that kind of language: to move from a model where ''most'' of the creative work is done by the reader to one where ''all'' of it is, is no great step.  


There is enough bad human literature like that out there now, that is is no great stretch to imagine doing without the human altogether. In that case, what does it matter what the text says, as long as it is coherent enough for an enterprising reader to make something out of it?  
There is enough bad human literature in existence already, that is is no great stretch to imagine doing without the human altogether. Who cares what the text is ''meant'' to say, as long as it is coherent enough for an enterprising reader to make something out of it?  


''But that does not work at all at all for legal language''. Legal language is code: it must say exactly what the parties require: nothing more or less, and it must do it in a way that leaves nothing open to a later creative act of interpretation. We should regard legal drafting as closer to computer code than literature: a form of symbol processing where the meaning resides wholly within and is fully limited by the text.  
''But that does not work at all at all for legal language''. Legal language is code: it must say exactly what the parties require: nothing more or less, and it must do it in a way that leaves nothing open to a later creative act of interpretation. Legal drafting is as close to computer code as natural language gets: a form of symbol processing where the meaning resides wholly within and is fully limited by the text.  


But unlike computer code, the operating system it is written for is not a closed logical system, and even the best-laid code can still run amok. You can’t run it in a sandbox to see if it works.
But unlike computer code, the operating system it is written for is not a closed logical system, and even the best-laid code can still run amok. You can’t run it in a sandbox to see if it works.