One Question: How do we Build Legal AI that can be Trusted?

Inside Practice • December 29, 2025

Knowledge discipline and architectural control now matter more than models.

For most law firms, the AI conversation has already moved on from curiosity.

The early questions of can this draft, summarize, speed things up?, have largely been answered. What firms are grappling with now is something more difficult and more consequential: what does it mean to trust AI inside a legal operating environment?


Not “trust” in the abstract. Trust in the very practical sense that partners, risk teams, and clients care about:


  • Can we rely on the output?
  • Can we explain how it was produced?
  • Can we defend the system that produced it?


Across recent Inside Practice discussions with firm leaders, innovation teams, and technologists, a consistent pattern is emerging. Legal AI rarely fails because the model underperforms. It fails because the surrounding system was never designed for reliability.


Two pressure points show up again and again: how firms handle knowledge, and how they control access as AI systems become more autonomous. Each exposes a different weakness. Together, they explain why so many pilots stall when firms try to scale.


Why “adding retrieval” hasn’t delivered reliability

Retrieval-Augmented Generation (RAG) is often positioned as a corrective - a way to ground AI output in firm knowledge and reduce hallucinations. In practice, many firms discover that retrieval solves one problem while revealing several others.


Outputs may be technically linked to internal documents, yet still fall short of legal standards. Answers sound plausible but miss critical nuance. Authorities are incomplete. Context is flattened. The result is work that feels risky rather than reliable.


The issue is rarely the retrieval technique itself. It is the condition of the knowledge being retrieved.


Most law firm knowledge environments were not built with machine consumption in mind. Information is spread across systems, structured inconsistently, maintained unevenly, and constrained by governance requirements that matter enormously in legal practice, ethical walls, matter boundaries, jurisdictional sensitivities.


When that reality is ignored, RAG becomes a thin layer over unresolved complexity.


What firms are learning (often the hard way) is that reliability begins upstream. Before retrieval ever takes place, knowledge must be prepared: structured so meaning is preserved, terminology aligned across practices, outdated material addressed, and context narrowed deliberately so systems are not forced to guess.


Governance here is not a policy exercise. It is an architectural one. If context is too broad, models improvise. If it is too narrow, they underperform. Designing that balance is work firms can no longer avoid.


These realities are the focus of From Retrieval to Reliability: Making RAG Work in Law Firms, a live Inside Practice discussion on January 23, 2026, led by Illitch Real and Zsolt Apponyi of Rubiklab. The session is grounded not in theory, but in the operational realities firms encounter once they try to deploy RAG beyond controlled demos.


When AI stops answering questions and starts taking action

As firms move beyond retrieval and drafting, a second challenge emerges, one that raises the stakes considerably.


Agent-based systems change the nature of the risk. These systems do not simply respond to prompts. They interact with tools, touch data, and initiate actions. At that point, familiar governance questions become more urgent:


  • What can the system access?
  • What is it allowed to do?
  • And how consistently are those rules enforced?


Many firms discover that their existing AI governance frameworks were designed for a different era, one where AI lived inside a single interface and operated largely as a passive assistant. Agents expose the limits of that model. Permissions are implemented inconsistently. Integrations proliferate. Different tools expose the same data in different ways. Control becomes fragmented, and with it, accountability.


This is where architectural questions move to the foreground. The Model Context Protocol (MCP) has attracted attention not because it promises better answers, but because it addresses a different problem: how AI systems connect to the environments they operate within.


MCP focuses on standardizing the interface between AI applications and external resources, data sources, tools, and workflows. For law firms, its relevance lies in the possibility of reducing integration sprawl, enforcing permissions consistently, and introducing auditability into how AI systems actually function in practice.


Whether MCP itself becomes the dominant standard is almost beside the point. The underlying shift is more important. Firms are beginning to recognize that control cannot be bolted on after the fact. It has to be designed into the system.


These issues will be examined in Mastering AI with MCP: A Protocol for Controlling AI Data Access, taking place on January 21, 2026, with Morgan Llewellyn of HIKE2. The discussion is aimed squarely at firms that are already experimenting with advanced AI and are now confronting the operational consequences.


The uncomfortable middle ground firms are now in

What connects these two conversations is not technology, but operating reality.


Many firms are stuck in an uncomfortable middle ground. They have pilots that impress. They have proof-of-concepts that work under controlled conditions. But when asked whether those systems could be trusted across matters, practices, and jurisdictions, confidence drops.


The reason is structural.


  • Knowledge is treated as content rather than infrastructure.
  • Access is treated as configuration rather than design.
  • Governance is treated as oversight rather than architecture.


The result is AI that performs well in isolation but struggles when exposed to the complexity of real legal work.

What is emerging instead is a more demanding and more realistic view of legal AI:


  • Knowledge must be prepared intentionally, not simply indexed.
  • Context must be constrained deliberately, not inferred.
  • Access must be standardized, not recreated tool by tool.
  • Control must be embedded in architecture, not managed through exception handling.


This is not glamorous work. It does not produce viral demos. But it is the work required if AI is to become something firms can rely on rather than continually manage around.


Why these conversations matter now

Neither of these discussions are about chasing the next tool. Both are about confronting decisions firms are already making, often implicitly.


  • How much autonomy are we prepared to allow
  • Where do we draw boundaries and how are they enforced
  • What does “reliable” actually mean in a legal context


As AI capabilities accelerate, these questions are becoming harder to defer. Firms that address them deliberately will be better positioned to scale. Those that don’t will find themselves constrained not by regulation, but by their own systems.


RAG is not “just retrieval.” MCP is not “just integration.”

Both point toward the same destination: AI that is safe enough to scale, consistent enough to trust, and governed enough to defend.


We invite you to join these discussions.


Two events. One question: how do we build legal AI that can be trusted?


Register using codes  AGENTS (MCP) and  RAGAI (RAG).

By Inside Practice February 3, 2026
Most firms will arrive in 2026 with no shortage of AI tooling. The harder problem is what sits underneath it: whether your firm’s knowledge foundations, governance, and operating model are strong enough to make AI reliable, scalable, and defensible in practice. That’s the premise of AI x KM 2026 : positioning Knowledge Management as infrastructure, not support, and treating the “knowledge backbone” as the real differentiator in the next phase of legal transformation 
By Inside Practice February 3, 2026
The UK legal market enters 2026 at an inflection point where “data strategy” stops being a modernization project and becomes an operating philosophy. The event framing is blunt: data isn’t back-office, it’s the organizing principle for how legal work is priced, delivered, governed, and measured. Inside Legal Data: London (Apr 9, 2026) leans into that reality with a full-day agenda that connects the dots: institutional intelligence, architecture, governance, AI readiness, and the commercial outcomes leaders actually care about.
By Inside Practice January 16, 2026
Inside Legal AI Brief: 01/16/26 For the past two years, the legal industry has treated AI the way a toddler treats a marker: with wide-eyed wonder, frantic enthusiasm, and absolutely no plan for what happens when it ends up on the walls. Because legal AI has mostly been framed as a question of possibility: What can it do? How fast is it improving? Which tools should firms test? And to be clear: that phase is done. It’s over. The “AI as a fun side quest” era has ended. Because across the legal market, vendors, law firms, courts, regulators, professional bodies - AI is no longer being treated like an experiment. It’s being absorbed into the infrastructure of legal work. Which is a huge change, because “experiment” is where you can shrug and say, “Well, that didn’t work,” like you tried making sourdough in 2020. “Infrastructure” is where, if it fails, society starts emailing you like, “Hi, I can’t access my rights today, is the system down?” And that’s what changes the risk profile entirely. This is no longer a story about whether AI works. It is now a story about whether legal institutions are prepared to control it, govern it, and then, this is the real kicker, live with it . Because it’s one thing to invite AI into the building. It’s another thing to realize you’ve given it keys, a badge, and a desk right next to the confidential files. The Inflection Point We’re Already Past Here’s one of the clearest signals something fundamental has shifted: AI has crossed an internal threshold inside institutions. Vendors are no longer selling discrete tools, little “helpful” add-ons you can trial and then quietly abandon like a gym membership. They’re selling agentic systems , often paired with pricing models that are basically designed to encourage mass adoption. Litera’s rapid growth after launching no-cost access to agentic capabilities is a perfect example: remove friction, and adoption accelerates. ( Business Wire ) And courts? Courts are not just watching from a safe distance, stroking their judicial chins and murmuring, “Hmm.” U.S. judges are reportedly already using AI to summarize filings and assist with drafting. ( Wall Street Journal ) Which is… significant! Because once judges start using AI as a convenience layer, that doesn’t stay a “nice-to-have.” That changes how lawyers write, how they brief, what they emphasize, and what they assume will be “picked up.” And institutional change rarely arrives with fireworks. It arrives through normalization: one more workflow, one more “assist,” one more quiet assumption that the machine is just… part of the process now. That moment has already passed. The End of AI Euphoria and the Rise of Scrutiny Now, investment in AI hasn’t slowed. But it has sobered up. The market is moving beyond the “try everything” phase, where everyone acted like they were at a tech buffet piling shrimp onto their plate with both hands, and into a phase of discipline. Spending is being redirected toward what you need if AI is going to operate at scale without turning your firm into a liability factory: governance and oversight security and data control reliability and auditability integration with knowledge, data, and workflow systems and ROI you can actually explain to leadership and clients without sounding like a cult member And this shift mirrors what enterprise leaders are saying more broadly. McKinsey’s State of AI points to organizations focusing less on experimentation and more on operationalizing AI in ways that survive scrutiny. ( McKinsey ) Which makes sense. Because AI is no longer “innovation theater”, where everyone gets to wear a headset and say “synergy” and then go home. It is entering the same governance conversation as billing systems, pricing models, and client data platforms. Meaning: boring, consequential, and full of meetings. The true nightmare. Governance Is Rising, but Fragmenting If this were a neat story, regulation would be converging smoothly behind adoption. It is not a neat story. In the United States, AI governance is emerging through state-level activity, which means we’re heading toward the regulatory equivalent of being attacked by a flock of birds: not one clean problem, but dozens of small, frantic ones coming from different angles. You’ve got legislative developments in Texas ( Texas Legislature Online ), policy direction in California ( Governor of California ), and the likely future is layered compliance rather than one national framework. And in the UK, the signal is different but still consequential: professional bodies like the Law Society are issuing increasingly urgent guidance on pace and risk, even while enforceable standards remain fluid. ( Law Society ) So if you’re a firm operating across jurisdictions, the risk is not “no regulation.” It’s fragmentation . Which forces governance decisions now , before standards stabilize, meaning you’ll be building the plane while multiple governments are simultaneously arguing about what counts as a wing. The Quiet Risk: Reliability and Security Here’s the part that tends to get buried under shiny demos: as AI systems become more autonomous, their failure modes become less forgiving. Agentic systems introduce new attack surfaces, new manipulation risks, and new questions about containment. And the OWASP GenAI project has started formalizing these concerns, highlighting that agentic architectures demand fundamentally different security assumptions. ( OWASP GenAI ) And in legal contexts, those security and reliability questions aren’t abstract “tech issues.” They are privilege issues. Confidentiality issues. Professional liability issues. Client trust issues. Because if a system has memory, persistence, and the ability to take delegated action, the question isn’t “Can it draft a clause?” The question is: What happens when it drafts the clause, stores the clause, reuses the clause, and confidently cites the clause in a context where it absolutely should not? The industry’s focus on capability has obscured a much harder question: Can these systems be trusted, audited, and governed inside real institutions, the kind with malpractice exposure and ethical duties and clients who do not accept “the robot did it” as a defense? Workforce Redesign Is Beginning - Without a Map AI is no longer just augmenting legal work. It is starting to reshape how time, effort, and value are defined. Some firms are making that shift explicit. Ropes & Gray’s decision to credit associates for AI-related effort is a tangible signal that work design is changing. ( City A.M. ) And that’s actually a big deal, because it acknowledges the reality: if AI changes the workflow, you can’t keep measuring productivity using the same old yardstick and then act surprised when everyone’s miserable. Meanwhile, broader workforce signals, burnout, dissatisfaction, suggest institutions are redesigning work faster than they are redesigning incentives, evaluation models, and professional identity. So the risk is not “people will resist AI.” The risk is that work changes, and the definition of success doesn’t. And that creates a professional environment where the rules are unclear, the expectations are inconsistent, and everyone feels like they’re being judged on metrics designed for a different century. The Core Insight: Capability Is No Longer the Constraint Put all of this together and you get one conclusion: The limiting factor for legal AI is no longer technological capability. It is institutional readiness. Readiness to govern. Readiness to secure. Readiness to explain AI-mediated decisions to clients, courts, and regulators. Readiness to redesign work without eroding trust or professional judgment. Because the next phase of legal AI will not be won by the firms with the most tools. It will be won by the firms that treat AI as infrastructure , not magic , and invest like it. Which, unfortunately, is the least glamorous kind of investing. It’s not “Look at our futuristic demo.” It’s “We built controls, audit trails, policy, security, training, and accountability.” In other words: not a fireworks show. A foundation. And in law, the thing you build on matters a lot more than the thing you show off. The Inside Legal AI Brief, is the newly reformatted Inside Legal AI Newsletter. For more information on Inside Legal AI: www.insidelegalai.com For more information: www.insidepractice.com This brief was developed using AI tools in conjunction with proprietary internal research, expert inputs, and established editorial processes. For more information please reach out to us at contact@insidepractice.com 
Legal AI: London unites law firm leaders how to use AI in a  faster & more strategic way.
By Inside Legal AI August 12, 2025
Join Legal AI: London (Dec 3–4) to explore how law firms can use AI to boost knowledge, operations, and client service for smarter, faster outcomes.
Setting the Scene: Legal AI in the UK | Gen AI
By Inside Legal AI July 25, 2025
Generative AI is now vital for UK legal firms, with usage doubling, widespread workflow changes, and one-third budgeting specifically for GenAI licenses.
Legal AI Pathfinders Chicago
By Inside Legal AI March 12, 2024
Legal AI Pathfinders Chicago
Ai Pathfinder NYC
By Inside Legal AI January 19, 2024
As part of the program at the Legal AI Pathfinder's Assembly: New York chair, Damien Riehl posed a series of questions to attendees, focusing on their current perspectives and future outlooks regarding AI.
By Inside Legal AI December 19, 2023
Whose time is it to shine? Given the potential for generative AI to automate a significant portion of routine legal tasks, let’s pose the question – how might AI fundamentally redefine the conventional law firm structure, billing model, and value proposition?
By Inside Legal AI December 14, 2023
2023 has been a landmark year for the team at Inside Practice.
By Inside Legal AI December 14, 2023
If you missed Inflection Point: The Legal #AI Revolution Part 2 - we have just made the keynote session available - "What Does it Mean to be a Lawyer in an AI-Enabled World? "