Skip to content

Inversion

An important question in developing LLM-compatible software in legal (and elsewhere) is whether the control model is "direct" or "inverted." In short: Do I ask my DMS to summarize, and it asks ChatGPT (direct), or do I ask ChatGPT and it pulls the document from the DMS (indirect)?

From what I've seen, most LegalTech products are doing the former. I think there are a few reasons why, but I'm not sure they always hold up.

๐—ฃ๐—ฟ๐—ผ๐—บ๐—ฝ๐˜ ๐—˜๐—ป๐—ด๐—ถ๐—ป๐—ฒ๐—ฒ๐—ฟ๐—ถ๐—ป๐—ด ๐—œ๐˜€ ๐—›๐—ฎ๐—ฟ๐—ฑ: By using direct control, the LegalTech product takes control of the prompting, including not just a single prompt but the ability to construct chains of prompt-response-prompt-response (so-called "agentic" workflows). In the early days of LLMs that made sense. There was an art to prompt-engineering, and "agentic" workflows required additional programming. That's either entirely not true today or vanishing. Users have gotten more experienced at prompting for one thing. But more importantly, modern LLMs are very good at understanding any prompt you give them, and they all now have the ability to utilize the data they can access to construct agentic workflows that involve going back and forth with the data. More than that, I'd argue that in many situationsโ€”especially given the trillions of dollars being spent on these modelsโ€”AI companies are going to be BETTER at these things than a LegalTech company.

๐—ฆ๐—ฒ๐—ฐ๐˜‚๐—ฟ๐—ถ๐˜๐˜†: The argument here is "don't give an LLM access to your data." But look, that's true in both scenarios. No one has a competitive "closed loop" LLM; everyone is sending data to first-party LLMs; and everyone, regardless of the control model, has to deal with the security implications of that, which are real. In both cases you are going to be relying on your LegalTech vendor to have constructed the proper security guardrails. In the direct control scenario, they have to protect against prompt injection and similar AI attacks. In the inverted control scenario, they have to ensure proper authorization and scope. And here too, you could make a good argument that AI companies are BETTER at dealing with AI-specific security concerns, whereas LegalTech companies are better at protecting access to data.

๐—œ๐˜'๐˜€ ๐— ๐˜† ๐——๐—ฎ๐˜๐—ฎ: I think a lot of companies don't want to allow LLMs to connect to their software, not because of security but because they want to "leverage" the customer information they have to sell an LLM product. I don't want ChatGPT to access the documents I store; I'd rather sell you my AI summarization service. I don't want Claude to have access to my caselaw database; I want to sell you my AI research product. OK, but the question legal CONSUMERS should care about is "is that better"โ€”is your AI "wrapper" better than what would happen if you let ChatGPT connect, or are you just bundling.

To be clear, I'm not arguing the inverted control model is always right. My point is itโ€™s a valid model and one that I think shouldโ€”and willโ€”gain more currency in LegalTech.