Natural language processing: Difference between revisions

Jump to navigation Jump to search
no edit summary
No edit summary
No edit summary
Line 1: Line 1:
{{a|tech|}}One of the [[Holy Grails of reg tech]] is [[natural language processing]], a handful of varieties of the same thing: a machine that reads {{t|contract}}s for you.  
{{a|tech|}}A great hope of [[reg tech]] is [[natural language processing]], which presents itself in a handful of varieties of the same thing: a machine that reads {{t|contract}}s for you.  


===Examples===
===Examples===
Line 11: Line 11:


===Legal agreement review===
===Legal agreement review===
:“''[[AI]] can only follow instructions.The [[meatware]] can make a call that the instructions are stupid.''”
There is a well-known and widely feted [[natural language processing]] application<ref>Which shall remain nameless, though you don’t have to be a total ''nerd'' to know who we have in mind.</ref> which purports to save resources and reduce risk by performing a preliminary review of, say, [[confidentiality agreement]]s against a preconfigured [[playbook]].  
There is a well-known and widely feted [[natural language processing]] application<ref>Which shall remain nameless, though you don’t have to be a total ''nerd'' to know who we have in mind.</ref> which purports to save resources and reduce risk by performing a preliminary review of, say, [[confidentiality agreement]]s against a preconfigured [[playbook]].  


The idea is [[triage]]. The application scans the agreement and, using its [[natural language processing]], will pick up the policy points, compare them with the playbook and highlight them so the poor benighted lawyer can quickly deal with the points and respond to the negotiation. The software [[vendor]] proudly points to a comparison of their software against human equivalents in picking up policy points in a sample of agreements. The software got 94% of the points. The meatware only got 67%. the Software was quicker. And it needed less coffee. But in itself this may highlight a shortfall, not a feature, in the application. Not everything in the playbook really is a point. Much of it is rubbish, or [[nice-to-have]]s, or the paranoid ramblings of a [[chicken licken]]y credit officer somewhere. The very value a lawyer brings is to go "yeah, that's fine, jog on, nothing to see here. That is the one thing a [[natural language processing]] AI can’t do. The AI forces you to negotiate points regardless of how stupid they are.
The idea is [[triage]]. The application scans the agreement and, using its [[natural language processing]], will pick up the policy points, compare them with the [[playbook]] and highlight them so the poor benighted lawyer can quickly deal with the points and respond to the negotiation. The software [[vendor]] proudly points to a comparison of their software against human equivalents in picking up policy points in a sample of agreements. The software got 94% of the points. The [[meatware]] only got 67%. The Software was quicker. And — chuckle — it needed less coffee. Headline: ''dumb machine beats skilled human''.
 
But this may highlight a shortfall, not a feature, in the application. The day a [[palaver]] of [[risk controller]]s set their [[playbook]] parameters at their ''exact'' hard walkaway point is the day [[Good luck, Mr. Gorsky|Mr. Gorsky]] gets to the moon. So, not everything in the [[playbook]] ''says'' is a problem really ''is'' a problem. Much of a playback will be filled with [[nice-to-have]]s and other paranoid ramblings of a [[chicken licken]]y somewhere in a controller group. The very value a lawyer brings is to see a point and say, "yeah, that's fine, jog on, nothing to see here. That is the one thing a [[natural language processing]] [[AI]] can’t do. The AI ''forces'' you to negotiate ''all'' playbook points, regardless of how stupid they are. True: this isn’t the AI’s fault, but it ''is'' inevitable, and it ''is'' the AI’s limitation.
 
[[AI]] can only follow instructions.The [[meatware]] can make a call that the instructions are stupid.
 
And besides, having the [[AI]] spot the issues and asking the [[meatware]] to fix the drafting gets the [[triage]] backwards. Picking up the points — and recognising the stupid parts of the [[playbook]] is the “high value work”. That is what the [[meatware]] should be doing. Fixing the drafting is the dreary detail. That is where you want the AI to kick in. But contextually amending human language — you know, ''real'' “natural language processing” — is ''hard''. No {{t|AI}} that we have seen just yet can do it.  


But it get the [[triage]] backwards. Rather than having the lawyer pick up the major points (the high value work) and then employing the [[AI]] to process and finalize the detail, it is the [[AI]] which picks up the major points and tasks the lawyer with completing the clerical work. For the process to be productive the lawyer must rely on the AI to have identified '''all''' salient points. Otherwise, the lawyer must read the agreement in full as a sense check. In practice, natural language processing is not sophisticated enough to allow this level of comfort, nonetheless lawyers are encouraged to trust it. Hence a buried risk.
And how comfortable can we really be that the AI ''has'' spotted everything? If we assume — colour me cynical — the “natural language processing” isn’t quite as sophisticated as its marketers would have you believe<ref>That is is a glorified key-word search, in other words.</ref> then it is a bit [[reckless]] to put your faith in the [[reg tech]]. Is there no human wordsmith who could fool the [[AI]]?<ref>I bet I could. It is hardly challenging to insert an [[indemnity]] which does not use the words “[[indemnity]]”, “[[hold harmless]]” or “[[reimbursement|reimburse]]”.</ref> what if there is an odious clause not anticipated by the [[playbook]]?<ref>Given how fantastically paranoid a gathering of [[risk controller]]s can be this seems a remote risk, I grant you, but risks are [[fractal]], remember. And [[emergent]] in unexpectable ways. The [[collective noun]] for a group of [[risk controller]]s is a [[Palaver]], by the way.</ref> If the meatware can’t wholly trust the AI to have identified '''all''' salient points the lawyer must ''still'' read the whole agreement to check. Ergo, no time or cost saving.


Furthermore the reality is that many of the policy points in the [[playbook]] will be non-essential "perfect world" recommendations ("[[nice to have]]s") which an experienced negotiator will quickly be able to wave through in most circumstances.  
Furthermore the reality is that many of the policy points in the [[playbook]] will be non-essential "perfect world" recommendations ([[nice-to-have]]s”) which an experienced negotiator will quickly be able to wave through in most circumstances.  


But this software is designed to facilitate "right-sourcing" the negotiation to cheaper (ergo less experienced) negotiators who will rely on the playbook as guidance, will not have the experience to make a commercial judgement unaided and will therefore be obliged either to [[escalate]], or to engage on a slew of [[nice to have]] but bottom-line unnecessary negotiation points with the counterparty. Neither are good outcomes. Again, an example of [[reg tech]] creating [[waste]] in a process where investment in experienced human personnel would avoid it.  
But this software is designed to facilitate "right-sourcing" the negotiation to cheaper (ergo less experienced) negotiators who will rely on the playbook as guidance, will not have the experience to make a commercial judgement unaided and will therefore be obliged either to [[escalate]], or to engage on a slew of [[nice-to-have]] but bottom-line unnecessary negotiation points with the counterparty. Neither are good outcomes. Again, an example of [[reg tech]] creating [[waste]] in a process where investment in experienced human personnel would avoid it.  


The basic insight here is that if a process is sufficiently low in value that experienced personnel are not justified, it should be fully automated rather than partially automated and populated by inexperienced personnel
The basic insight here is that if a process is sufficiently low in value that experienced personnel are not justified, it should be fully automated rather than partially automated and populated by inexperienced personnel


{{ref}}
{{ref}}

Navigation menu