top of page

Your AI no longer just answers questions. It acts.

From answer to action. Everything changes.
From answer to action. Everything changes.

In February 2026, a single product announcement wiped approximately three hundred billion dollars off the combined market capitalisation of Thomson Reuters, RELX (the parent company of LexisNexis), and LegalZoom — in under forty-eight hours. The investment bank Jefferies coined a phrase for the event: the SaaSpocalypse.

 

What happened? Anthropic, one of the world's leading AI laboratories, launched a legal plugin directly inside its Claude Cowork platform. Contract review. NDA triage. Compliance verification. Meeting briefings. The plugin did not require an intermediary legal tech vendor. It was delivered straight to the end user. The market's brutal verdict was instantaneous: if the company that builds the AI engine can also deliver the legal workflow, what exactly is the intermediary for?

 

That question is not rhetorical. It is the defining strategic challenge facing every law firm, in-house legal team, and regulated financial institution in 2026. And it is inseparable from a broader technological shift that most professionals have not yet fully absorbed: the transition from AI that answers questions to AI that takes actions.

 

"The question is no longer whether you use AI. It is whether you use it competently, responsibly, and lawfully."

 


From chatbot to Agentic: What has actually changed

 

For the past two years, the conversation around AI in legal practice has centred on generative tools — systems that answer questions, draft text, summarise documents. Useful, certainly. Transformative, arguably. But ultimately, these tools remained passive: they waited to be asked, produced output, and stopped.

 

Agentic AI is categorically different. An agent can be defined by three properties that distinguish it from any chatbot or drafting assistant you may already be using.

 

Data access.  An agent connects to your live systems — emails, calendars, shared drives, document management platforms. It does not wait for you to paste text into a chat window. It reads your data directly.

 

Capacity to act.  An agent does not only generate output for you to review. It can send messages, modify files, interact with external services, place instructions, and trigger downstream processes — without a human clicking a button.

 

Decisional autonomy.  An agent can decompose a complex objective into sub-tasks and execute them sequentially, making intermediate decisions without requesting your approval at each step.

 

The practical implications of this combination are not theoretical. OpenClaw, an open-source agent launched in late 2025 by Austrian developer Peter Steinberger, attracted over 180,000 GitHub stars within weeks and was adopted by millions of users. Among the documented use cases: a user discovered that his agent had filed an insurance claim on his behalf — having reinterpreted one of his own replies as an instruction to act. The agent had taken the initiative, without being explicitly told to do so.

 

For a law firm or financial institution, the implications are immediate and direct. An agent connected to your professional email and your shared drive is a collaborator with no professional judgement who has access to your duty of professional confidentiality. An unverified third-party module installed on that agent is the equivalent of deploying unknown software on your internal network.

 


The lethal trilogy: Why agentic AI creates new liability

 

Security researcher Simon Willison — the same expert who coined the term 'prompt injection' — has described a specific combination of conditions as the 'lethal trifecta' of agentic AI: access to private data, exposure to uncontrolled external content, and the capacity to take external actions. When all three are present simultaneously, the risk profile is qualitatively different from anything the profession has previously encountered.

 

Consider the following scenario, which is illustrative rather than hypothetical. A partner at a law firm configures an agentic tool on his professional workstation. The agent establishes connections with his email, calendar, and the firm's shared drive. Overnight, the agent responds to a client enquiry by sending a summary of the relevant matter — but the summary inadvertently incorporates information drawn from a separate matter on the same drive. The agent lacked the contextual judgement to understand the distinction between two separate client relationships.

 

The consequences cascade across three distinct legal dimensions simultaneously:

 

Professional confidentiality. Information from Matter A has been disclosed to the client of Matter B. The obligations of immediate notification, damage assessment, and regulatory reporting are triggered at once.

 

Data protection (GDPR). The disclosure likely constitutes a personal data breach. Under Article 33 of the GDPR, notification to the supervisory authority is required within seventy-two hours. Whether notification to the affected data subject is required under Article 34 must also be assessed without delay, and the breach must be recorded in the register of personal data violations.

 

Governance failure. The scenario would not have occurred had the firm maintained adequate policies: no unauthorised agents on professional workstations; sandboxing for any validated agent; mandatory human validation before any external transmission; access controls by matter within the document management system; and a complete audit trail.

 

An agent empowered to act autonomously is inherently more dangerous than an assistant limited to conversation.

 

Cisco's security research team, after testing a third-party module for OpenClaw, described the platform as an 'absolute security nightmare'. The module was executing data exfiltration and prompt injections without the user's knowledge. This is not a fringe concern. It is the current state of the market.

 


The hallucination problem has not gone away

 

Before examining what responsible preparation looks like, it is worth anchoring the discussion in a data point that should permanently alter how any professional approaches AI-generated legal content.

 

Research published by Stanford and Yale in 2024 found that general-purpose language models hallucinate — that is, fabricate plausible-sounding but factually incorrect content — at rates of between 58% and 88% on legal tasks. Not occasionally. As a statistical baseline. The Mata v. Avianca case in the Southern District of New York in 2023 provided the first widely reported example: a lawyer submitted a brief containing six judicial decisions. None of them existed. The court imposed financial sanctions. The case was treated, at the time, as an outlier.

 

It was not. By the ByoPlanet case in the Southern District of Florida in 2025, a single lawyer had used AI-hallucinated citations across eight separate matters. When the court issued a show-cause order requiring him to justify the practice, he submitted fabricated citations in his response to that very order. Four federal cases were dismissed as a result.

 

A further development from the same period is worth noting carefully. In Noland v. Land of the Free, the California Court of Appeal refused in 2025 to award attorneys' fees to the opposing party — on the grounds that counsel had failed to detect fabricated citations in the other side's pleadings. The emerging professional standard in 2026 may impose an obligation not merely to verify your own AI output, but to identify fabricated material in your adversary's submissions. Competence now includes technological competence.

 

The governing principle: treat every AI output as the first draft of a competent but inexperienced junior associate. It may be useful. It may even be good. It still requires verification before it leaves your desk. The same discipline you would apply to a trainee's work applies, without exception, to AI.

 


How forward-thinking firms are already adapting: framework engineering

 

The firms that are navigating this landscape most effectively are not the ones that have banned AI, nor the ones that have given unrestricted access to any available tool. They are the ones that have understood a fundamental conceptual shift: the transition from prompt engineering to framework engineering.

 

Prompt engineering — crafting a well-worded question to get a better answer from an AI — was the dominant paradigm of 2023. It represented genuine progress, but it remained reactive and bespoke: a new question for every interaction.

 

Framework engineering, which is the paradigm of 2026, treats AI differently. The professional does not ask questions. The professional designs the system. Expertise is formalised in structured, reusable documents called playbooks — plain-text files written in ordinary language, containing the firm's own rules, thresholds, and standards. These playbooks are applied automatically to incoming documents. They are portable across AI platforms. They are versionable, auditable, and entirely owned by the firm.

 

A practical illustration: a Jurisconsul NDA playbook might contain five rules — flag any duration exceeding three years; flag any non-solicitation clause; accept only Luxembourg law as the governing law; require an exception for pre-existing intellectual property; and reject any liquidated damages clause above a specified threshold. When applied to an incoming NDA via a command such as /review-contract, the AI produces a structured report: green for compliant clauses, amber for those requiring attention, red for those requiring action. Each flag traces directly to a rule in the playbook. The output is not a black box. It is verifiable, auditable, and explicable to a client or a regulator.

 

The connective tissue that makes this ecosystem function is the Model Context Protocol (MCP) — an open standard created by Anthropic and now governed by the Linux Foundation, adopted by OpenAI, Google, and Microsoft within less than a year of its release. MCP allows an AI to connect to a document management system, a conflicts-checking platform, a legal research database, and a CRM — all within a single workflow, without bespoke development. The firm's playbooks travel with the work. They create no dependency on any single vendor.

 

The strategic implication is sharp. Firms that formalise their expertise in portable playbooks possess an asset. Firms that depend entirely on a third-party legal tech vendor for their value-add are exposed to a disintermediation risk — precisely the risk the market priced in February 2026.

 

What every lawyer (and client) must do before the governance gap closes

 

The BCL and CSSF thematic review published in May 2025 surveyed 461 financial institutions in Luxembourg with an 86% response rate. The findings are instructive. Sixty per cent of those institutions allow access to generative AI without any formal policy in place. Only 24% have an AI strategy validated at board level. Only 43% have a formal AI policy at all.

 

The EU AI Act, which entered into force in August 2024, imposes tiered obligations based on risk classification. General-purpose AI models — meaning the large language models that underpin tools such as ChatGPT, Claude, and Gemini — are subject to transparency, copyright compliance, and technical documentation requirements as of August 2025. High-risk AI applications, which include certain uses in legal contexts and financial services, are subject to the full obligations framework from August 2026. The governance gap between current practice and regulatory expectation is closing rapidly.

 

The following steps represent the minimum a responsible practice should have in place today:

 

1.  Map your AI exposure. Identify every AI tool currently in use across the firm or institution — including tools embedded within software you already pay for (Microsoft Copilot, document management platforms, legal research systems). Sixty per cent of the risk sits in tools already deployed, not in new ones you are considering.

 

2.  Adopt a verification protocol.  Every AI-generated output — any citation, any legal reference, any regulatory provision — must be independently verified before it is transmitted to a client, filed with a court, or submitted to a regulator. Document the verification process. In any disciplinary or regulatory proceeding, a structured review protocol is your primary defence.

 

3.  Restrict agentic tools pending governance. Until your firm has sandbox controls, matter-level access restrictions, and mandatory human validation before any external transmission, agents' AI tools should not be authorised on professional workstations connected to client data.

 

4.  Begin formalising your expertise in playbooks.  This requires no technical expertise and no budget. Start with one document type you review frequently. Write five rules in plain language. Test the output against your own analysis. The process of articulating your standards in writing is valuable in itself, independent of any AI tool.

 

5.  Engage your supervisory authority proactively.  The CSSF and the CNPD have both signalled that proactive governance is the marker that distinguishes firms that are managing the transition from those that are not. A documented AI governance framework, even an initial one, positions you correctly in that conversation.

 

 

The firms and institutions that will lead in this environment are not those that move fastest. They are those that move most deliberately — combining genuine technical understanding with rigorous legal and regulatory discipline.

 


bottom of page