Vibe-Coded Legal AI and the EU AI Act: What Builders Need to Know
A Clifford Chance senior associate recently demonstrated that a single lawyer, working evenings and weekends, could vibe-code a suite of AI-powered legal tools sophisticated enough to rival commercial platforms. His open-source repositories have been forked over a hundred times.
A dedicated community platform, Vibecode.law, launched weeks later. BRYTER and LexisNexis have both added vibe-coding capabilities to their products.
All of this indicates the barrier to building bespoke legal AI has collapsed.
Perhaps the next question we ought to be asking is: what happens next?
Every one of those tools; the contract analyser built in a weekend, the RAG pipeline over the firm's precedent library, the due diligence triage agent, the regulatory change monitor; is potentially an AI system within the meaning of the EU AI Act. And the person or organisation that built it is potentially a provider, with the full weight of Chapter III obligations attached.
The legal tech industry is celebrating a revolution in who can build software. The EU AI Act does not care who built the software. It cares what the software does.
The build-not-buy trend is real and accelerating
The economics are genuinely compelling, in that a senior associate or innovation lead with domain expertise and access to Open Code, Claude, or Cursor can now prototype a working legal AI tool in days.
The tool will fit the firm's exact workflow and use the firm's own data, and cost a fraction of a six-figure SaaS licence. What's more (and most critically) it solves the translation problem that has plagued legal tech procurement for years: the person who understands the process is the person building the tool.
It's worth noting that this shift is not only limited to law firms:
Legal tech companies themselves are increasingly building bespoke AI features on top of foundation models rather than licensing them; corporate legal departments are building internal automation; RegTech startups are wrapping LLMs with domain-specific retrieval and fine-tuning.
This practice is widespread.
The Act does not distinguish between vibe-coded and venture-backed
Article 3(1) of the EU AI Act defines an AI system as "a machine-based system that is designed to operate with varying levels of autonomy and that may exhibit adaptiveness after deployment and that, for explicit or implicit objectives, infers, from the input it receives, how to generate outputs such as predictions, content, recommendations, or decisions that can influence physical or virtual environments".
That definition is technology-neutral and process-neutral. It does not ask how the system was built, how long it took, or how much it cost.
- A RAG pipeline over a firm's case law database that recommends relevant precedents is an AI system.
- A contract review tool that classifies clauses by risk level is an AI system.
- An intake triage agent that routes matters based on inferred complexity is an AI system.
The question is not whether these tools are AI systems.
The question is what role the builder occupies under the Act, and whether the system is high-risk.
Most bespoke legal AI builders are providers, not deployers
The AI Act distinguishes between providers (who develop or commission AI systems and place them on the market or put them into service) and deployers (who use AI systems under the authority of providers). The distinction matters because providers bear the heaviest compliance obligations: risk management, data governance, technical documentation, logging, human oversight architecture, accuracy and robustness testing.
Deployers have a lighter burden.
If you license (ie, pay for) Harvey or Luminance and use it as directed, you are a deployer. The vendor is the provider. The vendor handles Annex IV documentation, Article 12 logging, and the rest of Chapter III.
But when you build your own tool, there is no vendor.
You are the provider. And Article 25 makes this explicit: if you substantially modify an AI system, you become a provider for the purposes of the modified system.
Fine-tuning a foundation model, building a custom retrieval layer, adding domain-specific post-processing, or orchestrating multiple models in an agentic workflow all constitute substantial modification under any reasonable reading of the provision.
The Clifford Chance associate's tools, the Vibecode.law community projects, and every bespoke RAG pipeline in every innovation lab in the City occupy this space.
The organisations behind them are providers. Most have not yet reckoned with what that means.
Several legal AI use cases are high-risk
Not every AI system under the Act triggers the full Chapter III regime. Only high-risk systems do. The classification turns on Annex III, which lists specific use cases across eight categories.
Several common legal AI applications map directly to Annex III:
Administration of justice and democratic processes (Annex III, paragraph 8). AI systems intended to assist judicial authorities in researching and interpreting facts and the law, and in applying the law to a concrete set of facts. A tool that analyses case law and recommends legal arguments is not far from this description. The boundary between "legal research assistance" and "interpreting facts and the law" is not as clear as builders might hope.
Access to and enjoyment of essential private services (Annex III, paragraph 5(b)). AI systems used to evaluate creditworthiness or establish credit scores, except where used for detecting financial fraud. Contract analysis tools used in lending decisions, insurance underwriting, or financial due diligence could fall here.
Employment, workers management and access to self-employment (Annex III, paragraph 4). AI systems used for recruitment, screening, filtering, or evaluating candidates. Legal AI tools used in HR-adjacent legal workflows; employment tribunal prediction, settlement valuation, or workforce risk assessment; may be caught.
The classification analysis is fact-specific. It depends on the tool's intended purpose, its deployment context, and the significance of its output in downstream decisions. This is precisely the sort of assessment that cannot be resolved by reading a blog post. It requires examining the specific system.
Crucially, most bespoke legal AI tools have never been through this analysis.
The builder did not pause to ask whether the contract classifier they vibe-coded on Saturday is a high-risk system under Annex III. The question simply did not arise because the build-and-ship cycle now moves faster than the compliance assessment cycle.
What high-risk classification means in engineering terms
If a bespoke legal AI tool is classified as high-risk, the provider must implement and maintain:
A risk management system (Article 9) that identifies foreseeable risks, estimates their likelihood and severity, adopts risk mitigation measures, and tests the system against those measures. This is not a document; it is an operational system that gates deployment. If releases never fail on risk thresholds, the risk management system is advisory rather than operational, and that is an exposure.
Data governance (Article 10) covering training, validation, and testing datasets. For RAG-based systems, this extends to the retrieval corpus: is the data representative, free from material errors, and appropriate for the intended purpose?
Technical documentation (Article 11 and Annex IV) across nine structured categories: general description, elements and development process, monitoring and functioning, appropriateness of performance metrics, a detailed description of the risk management system, changes over the lifecycle, harmonised standards applied, the system's intended purpose, and design specifications. Annex IV documentation must be maintained and kept current. It cannot be backfilled as a one-off exercise; it must be generated alongside the system's development. If you want to understand what each section actually requires, our post on Annex IV obligations for engineers sets out the full mapping.
Automatic logging (Article 12) capable of recording events throughout the system's lifecycle. Logs must enable post-deployment monitoring, support incident investigation, and be retained for a period appropriate to the system's intended purpose, with a floor of six months (Article 19). For a practical treatment of what to log, how long to keep it, and how to reconstruct a decision chain, see our post on Article 12 logging for engineers.
Human oversight measures (Article 14) that allow natural persons to understand the system's capabilities and limitations, monitor its operation, and intervene or override when necessary. For agentic systems, where outputs trigger downstream actions without human review, the oversight architecture is a genuine design challenge.
Accuracy, robustness, and cybersecurity (Article 15) including resilience to errors, faults, inconsistencies in input data, and attempts to exploit vulnerabilities.
A vibe-coded tool built in a weekend has none of this. Neither does a prototype that graduated to production because it "worked well enough." And that is the gap.
The compliance retrofit is the hard part
The legal tech conversation about vibe coding focuses on how easy it is to build. The compliance conversation, when it happens at all, focuses on what obligations apply. The gap between those two conversations is the engineering work required to retrofit compliance infrastructure into a system that was never designed for it.
This is not a documentation exercise. You cannot write your way to Article 12 compliance. Structured, queryable, tamper-evident logging must be instrumented into the system's architecture. Risk management must gate deployment through CI/CD. Documentation must be generated from system state, not authored after the fact. Human oversight must be designed into the product interface, not bolted on as a governance policy.
For teams that built their tool in a week, this retrofit represents a fundamentally different category of work. It is infrastructure engineering, not prompt engineering. And it requires understanding both the regulatory requirements and the system architecture deeply enough to bridge them.
The Omnibus does not rescue you
Some engineering leaders are banking on the Digital Omnibus proposal to push the high-risk deadline to December 2027. Even if it passes (and in two of three plausible legislative scenarios, it does not pass before August 2026), the Omnibus adjusts timing. It does not reduce scope. Every obligation described above still applies.
More practically: enterprise clients are already pricing regulatory risk into procurement. If a law firm's bespoke AI tool is used in client-facing work, the client may ask for evidence of compliance infrastructure before any regulator does. The commercial clock runs ahead of the regulatory clock.
What to do about it
The answer is not to stop building. The vibe-coding trend represents a genuine and valuable shift in how domain experts create tools. Lawyers building their own AI applications is a net positive for the quality and specificity of legal technology.
But the build-fast culture needs a compliance layer, and for most organisations that layer does not yet exist. The practical steps are:
- Audit what you have built. Catalogue every AI-powered tool in use, whether it was built by the innovation team, a partner's weekend project, or a vendor integration that was customised beyond its original scope. Many organisations do not have a complete inventory.
- Classify each system. Determine whether it falls within the Act's scope, what role you occupy (provider or deployer), and whether it maps to an Annex III high-risk category. This analysis is fact-specific and requires examining the system's architecture, intended purpose, and deployment context.
- Prioritise the high-risk systems. For systems classified as high-risk, assess the gap between current state and Chapter III requirements. The largest gaps are typically in logging (Article 12), technical documentation (Article 11/Annex IV), and risk management (Article 9).
- Retrofit compliance infrastructure. This is engineering work: structured audit logging, documentation-as-code pipelines, deployment gates tied to risk thresholds, human oversight interfaces. It cannot be done at the governance layer alone.
- Establish a build-compliance cadence. For new tools, integrate classification and compliance assessment into the build process. The goal is not to slow down building; it is to ensure that compliance infrastructure ships alongside the tool, not months after it reaches production.
The window is closing
The legal tech industry is living through a transition that the AI Act's drafters did not anticipate. The Act was written with a mental model of intentional AI product companies: a firm decides to build a system, classifies it, documents it, and places it on the market through a deliberate process.
The reality in 2026 is that AI systems are being built in weekends, forked from GitHub, customised over lunch, and deployed before anyone has asked whether they are in scope. The vibe-coding movement has made this easier, faster, and more widespread than ever. That is genuinely exciting, and it is genuinely risky.
The organisations that thrive will be the ones that build fast and build compliantly. The two are not in tension; they require different skills, and the compliance skill set is the one most build-fast teams are missing.