Template:M intro work Large Learning Model: Difference between revisions

no edit summary
No edit summary
 
(22 intermediate revisions by the same user not shown)
Line 1: Line 1:
{{d|LLM|/ɛl ɛl ɛm/ (''also “[[large language model]]”'')|n|}}
{{d|LLM|/ɛl ɛl ɛm/ (''also “[[large language model]]”'')|n|}}{{drop|O|nce upon a}} time, an [[LLM]] was a “Master of Laws”: the postgraduate mark of the ''sensei'' in the society of legal service providers — either of that, or of the indolence of one not prepared to strike out and put what she has learned into practice — but still: it spoke to perseverance, depth, comprehension and mastery, however pigeon-hearted its motivation.
Once upon a time, an [[LLM]] was a “Master of Laws”: the postgraduate mark of the ''sensei'' in the society of legal service providers — either of that, or of the indolence of one not prepared to strike out and put what she has learned into practice — but still: it spoke to perseverance, depth, comprehension and mastery, however pigeon-hearted its motivation.


If the [[Thought leader|thoughtleaderati]] are to be believed, now all one needs for that kind of expertise is a different kind of “LLM”: a “[[large language model]]”. [[Artificial intelligence]] rendered by a pattern-recognising, parallel-processing [[chatbot]].  
If the [[Thought leader|thoughtleaderati]] are to be believed, now all one needs for that kind of expertise is a different kind of “LLM”: a “[[large language model]]”. [[Artificial intelligence]] rendered by a pattern-recognising, parallel-processing [[chatbot]].  
Line 13: Line 12:


====Cui bono?====
====Cui bono?====
''Who benefits'', primarily, from this emergent technology? Experience should tell us that the first  — and often the ''last'' — to benefit from legal productivity tools are the ''lawyers''. Should we expect [[this time is different|this time be different]]?
{{drop|W|ho benefits, primarily}}, from this emergent technology? Experience should tell us that the first  — and often the ''last'' — to benefit from legal productivity tools are the ''lawyers''. Should we expect [[this time is different|this time be different]]?


Now, it is a truism that she who has a tool uses it, firstly, to improve her own lot. A commercial lawyer’s “lot” is predicated on two things: (1) ''[[time and attendance|time]]'' taken, and (2) ''[[Ineffable|ineffability]]'': the sense that what she does “passeth all [[muggle]] understanding”.
Now, it is a truism that she who has a tool uses it, firstly, to improve her own lot. A commercial lawyer’s “lot” is predicated on two things: (1) ''[[time and attendance|time]]'' taken, and (2) ''[[Ineffable|ineffability]]'': the sense that what she does “passeth all [[muggle]] understanding”.
Line 23: Line 22:
This is, in itself, a neat “[[simplification|simplification defeat device]]”: if you make a contract template sufficiently convoluted, the one-off cost of simplifying it so vastly outweighs the cost of just “tweaking” and living with it that few clients will ever take that first step to simplify. Even though the the ongoing costs of ''not'' rationalising dwarf the one-off costs of doing so, the long-term savings are always over that hump.
This is, in itself, a neat “[[simplification|simplification defeat device]]”: if you make a contract template sufficiently convoluted, the one-off cost of simplifying it so vastly outweighs the cost of just “tweaking” and living with it that few clients will ever take that first step to simplify. Even though the the ongoing costs of ''not'' rationalising dwarf the one-off costs of doing so, the long-term savings are always over that hump.


And bear in mind it will be the lawyers who deploy [[LLM]]s as a tool, not their clients. Why? ''Because of that [[ineffability]]''. An [[LLM]] is a pattern-matching device. It understands nothing. It cannot provide unmediated legal advice. It can only ever be a “back-breaker”: the “last mile” needs a human who knows what she is doing, understands the context and complicated human psychology at play in the cauldron of 1 [[negotiation]]. An [[LLM]] can draw pretty, impressive-at-a-distance doodles, but it cannot do that. Nor can it write [[legal opinion]]s — well, not meaningful ones — and nor, unmediated by a law firm, does it have the insurance policy or deep, suable pockets for which a client is paying when it seeks legal advice in the first place.
And bear in mind it will be the lawyers who deploy [[LLM]]s as a tool, not their clients. Why? ''Because of that [[ineffability]]''. An [[LLM]] is a pattern-matching device. It understands nothing. It cannot provide unmediated legal advice. It can only ever be a “back-breaker”: the “last mile” needs a human who knows what she is doing, understands the context and complicated human psychology at play in the cauldron of [[negotiation]]. An [[LLM]] can draw pretty, impressive-at-a-distance doodles, but it cannot do that. Nor can it write [[legal opinion]]s — well, not ''meaningful'' ones — and nor, unmediated by a law firm, does it have the insurance policy or deep, suable pockets for which a client is paying when it seeks legal advice in the first place.


An [[LLM]] can only be deployed, that is to say, by someone with skin in the game; who is prepared to put ''herself'' in jeopardy by accepting the assignment, which jeopardy she defends by the simple expedient of ''knowing what she is doing'' and checking her [[LLM]]’s output.
An [[LLM]] can only be deployed, that is to say, by someone with skin in the game; who is prepared to put ''herself'' in jeopardy by accepting the assignment, which jeopardy she defends by the simple expedient of ''knowing what she is doing'' and checking her [[LLM]]’s output.
Line 48: Line 47:


====Who’s client? Oh, right: she’s a lawyer, too.====
====Who’s client? Oh, right: she’s a lawyer, too.====
“But, [[JC]], come on. Be realistic. It is dog-eat-dog out there. Any lawyer keeps the bounty of the [[LLM]] from her clients will soon have her lunch eaten by others who won’t. You cannot fight the invisible hand. We are in a race to the bottom.”
{{drop|“B|ut, [[JC]], come}} on. Be realistic. It is dog-eat-dog out there. Any lawyer keeps the bounty of the [[LLM]] from her clients will soon have her lunch eaten by others who won’t. You cannot fight the invisible hand. We are in a race to the bottom.”


But are we?  
But are we?  
Line 59: Line 58:


====Conservative motivation====
====Conservative motivation====
Nor should we underestimate the overwhelming power of the lawyer’s intuition that ''what has gone before is sacrosanct''.
{{drop|N|or should we}} underestimate the overwhelming power of the lawyer’s intuition that ''what has gone before is sacrosanct''.


Lawyers are the last great [[positivist]]s: they understand instinctively that what has been already laid down by someone else — “posited” — is ''safer'' and than anything new that they might themselves contribute. The [[common law]] with its [[doctrine of precedent]], after all, is to all intents a divine commandment: ''in times of doubt, to do what has been done before''.  
Lawyers are the last great [[positivist]]s: they understand instinctively that what has been already laid down by someone else — “posited” — is ''safer'' and than anything new that they might themselves contribute. The [[common law]] with its [[doctrine of precedent]], after all, is to all intents a divine commandment: ''in times of doubt, to do what has been done before''.  
Line 70: Line 69:


====Literary theory, legal construction and LLMs====
====Literary theory, legal construction and LLMs====
Fittingly, the first chatbot was a designed as a parlour trick. In 1966 Joseph Weizenbaum, a computer scientist at MIT created “[[ELIZA]]” to explore communication between humans and machines. [[ELIZA]] used pattern matching and substitution techniques to generate realistic conversations. By today’s standards, [[ELIZA]] was rudimentary, simply regurgitating whatever was typed into it, reformatted as an open-ended statement or question, thereby inviting further input. As a session continued, the user’s answers became more specific and elaborate, allowing [[ELIZA]] to seem ever more perceptive in its responses.  
{{quote|“What an astonishing thing a book is. It’s a flat object made from a tree with flexible parts on which are imprinted lots of funny dark squiggles. But one glance at it, and you’re inside the mind of another person, maybe somebody dead for thousands of years. Across the millennia, an author is speaking clearly and silently inside your head, directly to you.  


Even though [[ELIZA]] was a basic “keepy uppy” machine, it proved surprisingly addictive — even to those who knew how it worked. Weizenbaum was famously shocked how easily people, including his own secretary, were prepared to believe [[ELIZA]] “understood” them and contributed meaningfully to the interaction.  
Writing is perhaps the greatest of human inventions, binding together people who never knew each other, citizens of distant epochs. Books break the shackles of time. A book is proof that humans are capable of working magic.”
:— Carl Sagan, ''Cosmos''
“I think you underestimate the power of ''reading'', Professor Sagan.”
:—[[Jolly Contrarian|JC]]}}
=====Theres’ a hole in my model, ELIZA=====
{{drop|F|ittingly, the first}} [[chatbot]] was a designed as a parlour trick. In 1966 Joseph Weizenbaum, a computer scientist at MIT created the programme “[[ELIZA]]” to explore communication between humans and machines. [[ELIZA]] used pattern matching and substitution techniques to generate realistic conversations. You can try it out [https://web.njit.edu/~ronkowit/eliza.html here].


This is, of course, how all “mind-reading” works: with the right kind of clever questions, the conjurer extracts from the subject herself all the information she needs to create the illusion.  
By today’s standards, [[ELIZA]] was rudimentary, simply regurgitating whatever was typed into it, reformatted as an open-ended statement or question, thereby inviting further input.  


[[LLM]]s work the same way. Like all good conjuring tricks, [[generative AI]] relies on misdirection: its singular genius is that it lets us misdirect ''ourselves'', into wilfully suspending disbelief, never noticing who is doing the creative heavy lifting needed to turn machine-made screed into magic: ''we are''. We are neuro-linguistically programming ''ourselves'' to be wowed by LLMs.
Even though [[ELIZA]] was little more than a basic “keepy uppy” machine, it proved surprisingly addictive. Weizenbaum was famously shocked how easily people were prepared to believe [[ELIZA]] “understood” them and contributed meaningfully to the interaction. When trying it out, Weizenbaum’s own secretary, who had watched him build the programme for months and knew how it worked, shooed him out of the room so she could have privacy with the machine.  


Yet again, we are subordinating ourselves to easy convenience of the machines. We need to have some self-respect and kick this habit.
This is, of course, how all “mind-reading” works: by asking the right kinds of question, the conjurer extracts from the subject all the information needed to create the illusion of telepathy.


By writing prompts, we create our own expectation of what we will see. When the pattern-matching machine produces something roughly like it, we use our own imaginations to frame, filter, boost, sharpen and polish the output into what we want to see. We render that output as commensurate we can with our original instructions.  
Here is the thing: ''[[LLM]]s work the same way''.  


When we say, “fetch me a tennis racquet”, and the machine comes back with something more like a lacrosse stick, we are far more impressed than we would be had a human done the same thing. We would think the human a bit dim. But with [[generative AI]] we don’t, at first, even notice we are not getting what we asked for. We might think, “oh, that will do,” or perhaps, “ok, computer: try again, but make the basket bigger, the handle shorter, and tighten up the net.” We can iterate this way until we have what we want — or we could just use a conventional photo of a tennis racquet.
=====LLMs as conjuring tricks=====
{{quote|
''POLONIUS'': What do you read, my lord? <br>
''HAMLET'':  Words, words, words. <br>
''POLONIUS'':  What is the matter, my lord? <br>
''HAMLET'':  Between who? <br>
''POLONIUS'':  I mean the matter that you read, my lord. <br>
:—''Hamlet'', II, ii
}}
{{drop|L|ike all good}} conjuring tricks, [[generative AI]] relies on misdirection: its singular genius is that it lets us misdirect ''ourselves''. We wilfully suspend disbelief, never noticing who is creatively re-rendering machine-made screed as magic. ''We are''. We are neuro-linguistically programming ''ourselves'' to be wowed by a {{plainlink|https://en.wikipedia.org/wiki/Digesting_Duck|mechanical duck}}, when the clever part is really happening inside our own heads.


AI-generated image generation famously struggles with hands, eyes and logical three-dimensional architecture. First impressions can be stunning, but the second look reveals an absurdist symphony.  It is just as true of text prompts: on close inspection we can see the countless minute logical ''cul-de-sacs'' and two-footed hacks. (Many humans write in logical ''cul-de-sacs'' and two-footed hacks, but that is another story.)
By writing “prompts”, we create an expectation of what we will see. When the pattern-matching machine produces something roughly like that, we use our own imaginations to backfill, frame, filter, correct, boost, render, sharpen and polish the output into what we ''wanted'' to see. We ''construe'' the output to conform to our original instructions.  


Either way, the novelty soon palls: as we persevere we begin to see the magician’s wires.We get a sense of how the model goes about what it does. It has its familiar tropes and tics and persistent ways of doing things which aren’t quite what you have in mind. The piquant surprise at what it produces dampens at each go-round, eventually settling into an [[Entropy|entropic]] and vaguely dissatisfying quotidian.  
When we say, “fetch me a tennis racquet”, and the machine comes back with something more like a lacrosse stick, we are far more impressed than we would be of a human who did the same thing: we would think such a human a bit ''dim''. But with [[generative AI]] we don’t, at first, even notice we are not getting what we asked for. We might think, “oh, nice try!” or “well, that will do,” or perhaps, “ok, computer: try again, but make the basket bigger, the handle shorter, and tighten up the net.” We can iterate this way until we have what we want — though note all the refining intelligence is coming from ''us'' — or we ''could'' just Google for a conventional photo of a tennis racquet.


In this way the appeal of iterating a targeted work product with a random pattern-matcher loses its lustre. The first couple of passes are great: they get from zero to 0.5. But the marginal improvement in each following round diminishes, as the machine reaches asymptotically towards its upper capability in producing what you had in mind, which we estimate unscientifically as about 75% of it.
Yet an alarming number of people seem to want to believe that human ingenuity pales against a magic box. This involves quite a lot of motivated forgetfulness. For example, it was claimed of Google glass that it could read facial expressions to detect underlying emotional states:


Now, as [[generative AI]] improves  towards 100 — assuming it does improve: there are some indications it may not; see below — that threshold may move but it will never get to 100. In the mean time, as each successive round takes more time and bears less fruit, mortal enthusiasm and patience with the LLM will have long-since waned: well before the [[Singularity]] arrives.


And many improvements we will see will largely be in the [[meatware]]: as we refine and elaborate our queries; we learn how better to frame our queries, and “prompt-engineering” becomes the skill, rather than the dumb, parallel pattern-matching process that responds to it. Ours is the skill going in, and ours is the skill construing the output. What the machine does is the boring bit.  
{{quote|The app detects emotion in the faces it sees using an algorithm generated by a machine learning system. This AI system was trained on large datasets of faces to decode the emotions from facial expressions.<ref>[https://www.cbc.ca/radio/quirks/march-30-erasing-memories-biggest-t-rex-and-the-smell-of-parkinson-s-and-more-1.5075050/google-glasses-could-help-kids-with-autism-read-emotional-cues-in-people-s-faces-1.5075055 As reported on CNBC]</ref>}}


In all kinds of literature ''bar one'', construal is where the real magic happens: it is the [[Emergent|emergent]] creative act that renders ''King Lear'' an timeless cultural leviathan and {{br|Dracula: The Undead}} forgettable pap<ref>Maybe not ''that'' forgettable, come to think of it: it has stayed with me 15 years, after all.</ref>. A literary work may start with the text, but it barely stays there for a moment. The “meaning” of literature is personal: it lives between our ears, and within the cultural milieu that interconnects the reading population.  
Sounds amazing, huh? But ''humans don’t need AI goggles to read each others’ emotions''. They do it naturally. They are really good at it — far better than any [[algorithm]] — already. Even ''dogs'' can read facial expressions.<ref>[https://www.rspca.org.uk/-/blog_how_dogs_know_what_were_feeling#:~:text=In%20recent%20times%2C%20research%20has,human%20emotions%20by%20smell%20alone. Dogs know what we are feeling], RSPCA.</ref>


“Construal” and “construction” are interchangeable in this sense: over time that cultural milieu takes the received corpus of literature and, literally, ''constructs'' it into edifices its authors can have scarce have imagined. ''Hamlet'' speaks, still, to the social and human dilemmas of the twenty-first century in ways Shakespeare cannot possibly have contemplated.<ref>A bit ironic that Microsoft should call its chatbot “Bard”, of all things.</ref>
Why are we aspiring to have machines do badly things we already do effortlessly? Who benefits from that?


Now there is one kind of “literature” where the last thing the writer wants is for the reader use her imagination to fill in holes in the meaning. Where clarity of authorial intention is paramount; where communicating and understanding ''purpose'' is the sole priority: ''legal'' literature.  
So it is with [[generative AI]]. AI image generation struggles with hands, eyes and logical three-dimensional architecture. It is just as true of text prompts: on close inspection we see the countless logical ''cul-de-sacs'', bad guesses and ''non sequiturs'' from which a clever story miraculously [[Emergence|emerge]]s. (To be sure, many human authors write in logical ''cul-de-sacs'', bad guesses and ''non sequiturs'', but that is another story.) Again, where is the magic in this process? ''Inside the [[meatware]]''.


The ''last'' thing a legal drafter wants is to cede interpretative control to the reader. Rather, she seeks to squash all opportunities presented by creative ambiguity. Just as there are no atheists in foxholes, [[there are no metaphors in a trust deed|there are no metaphors in a Trust Deed]].  
Either way, as we persevere, we see more and more of the magician’s wires. We sense how the model works: its familiar tropes, tics and consistent ways of doing things, which are never quite what you had in mind. The first couple of passes are great: they get from zero to 50% of what we wanted. But the marginal improvement in each following round diminishes. The piquant surprise at what the machine can produce dampens at each go-round, reaching asymptotically towards an upper capability, well short of 100%, in “producing what you had in mind”.  


Legal drafting seeks to do to readers what code does to computer hardware: it reduces the reader to a machine, a mere symbol processor. It leaves as little room as possible for interpretation.  
Eventually we settle for an [[Entropy|entropic]], vaguely dissatisfying quotidian. “Okay, it isn’t quite what I had in mind. but in the interests of time it will do.” 


This is one reason why [[legalese]] tends to be so laboured. It is designed to chase down and prescribe outcomes for all logical possibilities, remove all ambiguity and render the text mechanical, precise and reliable. Where normal literature favours possibility over certainty, legal language bestows [[certainty]] at the cost of [[possibility]], and to hell with literary style and elegance.
Again, we subordinate our own interests to the machine’s. We accept a mediocre job we would not take from a human. Should we not be more exacting? Should we not hold an LLM at least to the a standard of a human?


Legal language is ''finite''. Literature is ''[[Finite and Infinite Games|infinite]]''.
Ours is the skill going in, and ours is the skill construing the output. The machine does the boring bit:  the bit that machines have ''always'' done: applies uninspired strength, speed, reliability, and economy.  


Now: the punchline. Given how important the reader and her cultural baggage are to the creative act in normal literature, we can see how a large learning model is a feasible model in that domain: to move from a model where ''most'' of the creative work is done by the reader to one where ''all'' of it is is no great step. Indeed, there is enough mediocre literature out there that meets this description now, only written by humans, but to formula and slavishly aping hackneyed archetypes that is is no great stretch to do without the writer altogether. In that case, what does it matter what the text says, as long as it is coherent enough for an enterprising reader to make something out of it?
''Not ingenuity''.
=====LLMs and literary theory=====
{{quote|
''HORATIO'': O day and night, but this is wondrous strange. <br>
''HAMLET'': And therefore as a stranger give it welcome.<br>
There are more things in heaven and earth, Horatio, <br>
Than are dreamt of in your philosophy. <br>
:— ''Hamlet'', I, v}}


''But that does not work at all at all for legal language''. The language must say exactly what the parties require: nothing more or less, and it must do it in a way that leaves nothing open to a later creative act of interpretation. We should regard legal drafting as closer to computer code than literature: a form of symbol processing where the meaning resides wholly within and is fully limited by the text. But unlike computer code, you can’t run it in a sandbox to see if it works.
{{drop|N|ow, in all}} kinds of literature, ''bar one'', “reading” is where the real magic happens. ''Construal''. It is the [[Emergent|emergent]] creative act and community consensus that renders ''Hamlet'' a timeless cultural leviathan and {{br|Dracula: The Undead}} forgettable pap.<ref>Maybe not ''that'' forgettable, come to think of it: it has stayed with me 15 years, after all.</ref>
 
A literary work may start with the text, but it stays there barely a moment. The “meaning” of a literary work is ''necessarily personal'' to the reader: it lives between the reader’s ears, and within the cultural milieu that interconnects the reading population over the generations — ''all as construed by the apparatus between the reader’s ears''.
 
Call him [[post-modern]] — go on, do — but the [[JC]] doesn’t hold with [[Carl Sagan]]’s idea, above, that a book teleports its author “inside our heads”. That is to equate ''construal'' with ''symbol-processing''. It absolutely isn’t, and that [[metaphor]] — that the brain is a glorified Turing machine — gravely underestimates the human brain when in construction mode.
 
This, by the way, is not in any way to diminish Shakespeare’s towering genius, but rather to observe that, however impossibly brilliant it is, it is swamped by the flood of exposition, analysis, interpretation, re-rendition and performance that has gone on since he published it, all as narratised and framed by the apparatus between the reader’s ears. The quote above — as Hamlet deals with a ghost — is apposite because — well, because ''thinking makes it so''.<ref>''Hamlet'', II, ii.</ref>
 
“Construal” — interpreting and assigning meaning to — and “construction” — building — have the same etymology. They are interchangeable in this sense: over time that cultural milieu takes the received corpus of literature and, literally, ''constructs'' it, into edifices its authors can scarce have imagined.
 
It is because a reader is ''not'' a simple “symbol processor”, that she does more than merely decrypt text to reveal a one-to-one assembly of the author’s intention in her own head, that ''Hamlet'' can speak, still, to the human dilemmas of the twenty-first century.
 
=====On the lack of metaphors in foxholes=====
{{Drop|T|here is one}} kind of “literature” that ''is'' like a computer programme: where the ''last'' thing the writer wants is for the reader use her imagination, , or bring her cultural baggage in from the hall and use it to “construct” a meaning. In this singular domain, clarity of the writer’s intention is paramount: the only priority is divining the what those who wrote the text meant by it.
 
This is, of course, ''legal'' literature.
 
Rather than ceding interpretative control to the reader, a legal drafter seeks to squash any chance for improvisation, to pre-emptively stomp on all ambiguity. Just as there are no atheists in foxholes, ''[[there are no metaphors in a trust deed|there are no metaphors in a Trust Deed]]''.
 
Legal drafting seeks to be as “[[finite]]” as it can be. It strives as far as it can to readers to mere symbol-processing machines; to allow them only the job of extracting the author’s incontrovertible meaning. But, in a natural language constructed out of dead [[metaphor|metaphors]], this is very, very hard to do: that such a comfortable living is to be made conducting commercial litigation shows that.
 
It is one reason why [[legalese]] is so laboured. To render text as mechanical, precise and reliable as it can be, lawyers are compelled to chase down all blind alleys, previsualise all phantoms and prescribe outcomes for all logical possibilities. Where the best “literary” language prefers possibility to certainty, ''legal'' language sanctifies [[certainty]] above all else, at the cost of [[possibility]]. To hell with literary style and elegance.
 
Where literary language is, in [[James Carse]]’s sense, ''[[Finite and Infinite Games|infinite]]'', legal language is ''finite''.
=====LLMs okay for literature: bad for legal drafting=====
{{drop|N|ow: the punchline}}. Given how integral the reader and her cultural baggage are to the creative act in normal literature, we can see how, in that domain, a [[large language model]], freely spitting out text ripe with interpretative possibilities — text that positively ''begs'' for someone to “construct” it, just to make sense of it — is a feasible model.
 
The interpretative task is paramount: to move from a model where most of the creative work is done by the reader to one where all of it is, is no great step. We can imagine doing without a human writer altogether. The Writers Guild of America can certainly imagine that. Who cares what the text is meant to say, as long as it is coherent enough for an enterprising reader to make something out of it?
 
''But that does not work for legal language''. Legal language is code: it must say exactly what the parties require: nothing more or less, and it must do it in a way that leaves nothing open to later creative interpretation. Legal drafting is a form of symbol processing, where the meaning resides wholly within and is fully limited by the text.
 
A random-word generator that creates plausible sounding legal text, but whose singular meaning is incoherent, or not exactly reflective of the parties’ intent, is not much use. If skilled lawyers must sit with it while it iterates, the question arises: why persist with it at all? Just have the lawyers write the stuff in the first place.
 
We have seen it suggested that one might invert the process instead, letting humans do the first cut, then pressing the machine into action to critique the human drafts, to find potential errors and omissions.
 
But this is to get the [[division of labour]] exactly backwards, using expensive, context-sensitive [[meatware]] to do the legwork and a dumb machine to provide the “magic”. And besides, there is a design flaw in any legal process which supposes that the risk in a legal contract is distributed evenly throughout its content, and that therefore the legal proposition is one of handling volume.
 
Boilerplate is [[boilerplate]] for a reason. It is pinned down; tried, tested and done: it takes those things that should go without saying out of the equation. There is nothing to be gained from having a [[large language model]] drafting boilerplate from scratch each time — especially if, randomly, it ''changes'' things: ''[[The quotidian is a utility, not an asset|boilerplate is a utility, not an asset]]''. It is the part of the deal the lawyers already spend the least time on. The parts that matter — that they do spend time on — are typically discrete, manageable parts of drafting. It is hard to see how an LLM really assists in the process.
 
We come back to that question of basic human self-respect. We ''need to stop subordinating ourselves to machines''. We should not compromise just to optimise for machine processing, to make it easier for machines to manage what we do.


====Meet the new boss —====
====Meet the new boss —====
We don’t doubt that LLM is coming, nor that the legal industry will find a use for it: just that there is a ''useful'', sustained use for it. It feels more like a parlour trick: surprising at first, diverting after a while, but then the novelty wears off, and the appeal of persevering with what is basically a gabby but unfocussed child wears pales.
{{drop|J|C does not}} doubt that LLM is coming, nor that the legal industry will find a use for it: just that there is a ''useful'', sustained use for it. It feels more like a parlour trick: surprising at first, diverting after a while, but then the novelty wears off, and the appeal of persevering with what is basically a gabby but unfocussed child wears pales.


The traditional legal model faces existential challenges for sure, but they are not presented, and will not be addressed by random word generators.
The traditional legal model faces existential challenges for sure, but they are not presented, and will not be addressed by random word generators.


==== Coda: is ChatGPT getting worse? ====
==== False memory syndrome ====
{{drop|L|arge learning models}} already can execute this “research assistant” mode. It seems less flashy but — if we can trust it — this is a compelling use case that will remain once our inflated expectations has collapsed into the trough of despair. We could ask Bing AI to track down a long-lost childhood story.{{Quote|1=Bing, do you know a children’s story involving a fox or a wolf or something being chased by a pack of hunting dogs and it hides under a rock. The hiding place is too small, and it has to poke one part of its body out. It has an argument with each of its body parts to see which is the most deserving of being saved and concludes its tail is the least deserving, so it sticks its tail out. The hunting dogs arrive and grab it by the tail and catch it. Moral of the story is something like you are only as good as your weakest link.}}At first, Bing shrugged its shoulders and said, “sorry”. When we persevere with a few more details, success!{{Quote|1=I found the story you were looking for! It’s called “The Wolf and His Tail”. In this story, a wolf is being chased by a pack of hunting dogs and hides under a rock, The hiding place is too small, and it has to poke one part of its body out. It has an argument with each of its body parts to see which is the most deserving of being saved and concludes that its tail is the least deserving so it sticks its tail out. The hunting dogs arrive and grab it by the tail and catch it. The moral of the story is that you are only as good as your weakest link.}}Magnificent! We can ask who wrote the story:{{Quote|“The Wolf and His Tail” is a children’s story that was written by Barbara Juster Esbensen. It was first published in 1972.}}There is only one problem: ''this is not true''. Barbara Juster Esbensen was indeed a children’s author, but she wrote no such book. It is as if Bing has been badgered into false memory syndrome.
 
Perhaps we should see what Google’s Bard has to say. We can input the same prompt. This is the result:{{Quote|The story you’re describing sounds like the Aesop fable “The Fox and the Hunter.” In this fable, a fox is being chased by a pack of hunters. The fox finds a narrow crevice in a rock and hides inside. However, the crevice is too small, and the fox’s tail is sticking out. The hunters arrive and see the fox’s tail, and they grab it and pull the fox out of the crevice. The fox is then killed by the hunters.
 
The moral of the story is that a chain is only as strong as its weakest link. This means that if any part of a system is weak, the entire system can fail. The fox’s tail was its weakest link, and as a result, the fox was caught by the hunters.}}We can be a bit embarrassed to have forgotten about Aesop, until we go looking for the Aesop’s fable and discover that ''it doesn’t exist either''. Bard just made this up.
 
==== Whatever the use caseCoda: is ChatGPT getting worse? ====


In other news, scientists are concerned that [[ChatGPT]] might be getting worse. Studies indicate that its accuracy at tasks requiring computational accuracy, like playing noughts and crosses or calculating prime numbers, is rapidly diminishing.
{{drop|I|n other news}}, scientists are concerned that [[ChatGPT]] might be getting worse. Studies indicate that its accuracy at tasks requiring computational accuracy, like playing noughts and crosses or calculating prime numbers, is rapidly diminishing.


Perhaps [[ChatGPT]] is getting bored, or might it have something to do with the corpus increasingly comprising nonsense text generated on the hoof by some random using ChatGPT?
Perhaps [[ChatGPT]] is getting bored, or might it have something to do with the corpus increasingly comprising nonsense text generated on the hoof by some random using ChatGPT?