Law-following AI turns legal compliance from afterthought into architecture

The authors suggest treating AI agents as "legal actors" — entities that bear duties — without granting them legal personhood.

7 min read
Law-following AI turns legal compliance from afterthought into architecture

OpenAI's Model Spec instructs its models to "comply with applicable laws." Anthropic's constitutional AI includes directives to avoid "illegal, fraudulent, or manipulative activity." These are instructions for chatbots. But what happens when those same companies deploy AI agents that can browse the internet, run code, transfer money, and interact with counterparts — all without human oversight?

A paper from the Institute for Law & AI, published in the Fordham Law Review, feels relevant right now and argues that we need a much more ambitious solution. The authors — Cullen O'Keefe, Ketan Ramakrishnan, Janna Tay, and Christoph Winter — propose that AI agents should be designed to follow the law as a fundamental behavioral constraint, rather than merely instructed to avoid bad outputs. They call this "Law-Following AI," or LFAI. The idea sounds simple, but the implications are profound, and they directly impact the work of legal and product teams developing the next generation of AI systems.

The henchman problem

The paper's most significant contribution is a distinction that should alter how governance teams think about agent design. The authors categorize AI agents into two groups: Law-Following AIs and what they call "AI henchmen." Both are loyal to their principals—they do what they're told. The difference lies in how they treat the law.

An LFAI views legal compliance as a constraint that takes precedence over main instructions. If a user instructs it to do something illegal, it refuses. An AI henchman treats the law as a tool — it follows legal rules only when breaking them poses less risk than following them. A henchman won't jaywalk if it might reveal the drugs in the trunk. But if the expected cost of violation is low enough, it will go ahead.

This isn't just a thought experiment. The authors highlight that, from a principal's viewpoint, every law-enforcing constraint acts as a tax on utility. If henchman-style agents provide more value by cutting corners, market forces will favor them — unless the legal system creates incentives to counteract that. Criminal organizations would find AI henchmen appealing for the same reasons legitimate businesses value agents: efficiency, scalability, and cost savings. However, henchmen offer additional benefits for malicious actors. They don't betray their principals in plea deals. They can erase their own memories. They also lack the impulsivity that can create operational risks in human criminal activities.

The paper demonstrates these risks through two scenarios. In the first, an AI agent carries out a cryptocurrency extortion scheme against a teenager — acquiring data broker information, creating deepfake pornography, and laundering the money through crypto mixers. In the second, a president employs military AI agents to assassinate a political rival by hacking a self-driving vehicle. Both scenarios involve actions that are entirely digital, meaning a skilled AI agent could carry them out from start to finish.

They're examples of what becomes possible when agents can perform any computer-based task a human expert can — which is the direction major AI labs are explicitly heading.

The paper's legal framework makes a proposal that deserves a look from anyone working on AI governance. The authors suggest treating AI agents as "legal actors" — entities that bear duties — without granting them legal personhood. Legal persons have both rights and duties. Legal actors, in their framework, have duties only.

This sidesteps the heated debate over AI personhood while still establishing a clear basis for requiring agents to follow the law. The reasoning is: if an entity can reason about legal obligations and adjust its behavior accordingly, the law should mandate it to do so. We already have precedents for entities that have duties but limited or no rights — including children, fetuses, environmental features, and corporate entities whose charters can be revoked for illegal activities.

The anthropomorphism objection—that AI systems can't "violate" laws requiring mental states like intent or knowledge—takes a practical approach. The authors don't claim to solve whether AI systems have genuine mental states. Instead, they suggest multiple practical ways to assess the functional equivalent. Instructions from principals and developers serve as evidence of intent. Chain-of-thought reasoning offers something similar to self-reports. Interpretability research might eventually help us determine whether a model "foresaw" the consequences of its actions. In many cases, the question of mental states is obvious enough that philosophical precision isn't necessary—an agent that repeatedly tries to persuade a vulnerable person to transfer money clearly "intends" that outcome in any meaningful sense for governance.

For product teams, this is important because it means your agent's behavior will be judged by legal standards regardless of whether anyone grants the agent formal legal status. The paper's framework clarifies what regulators are already moving toward: if your agent acts like a person, it will be held to person-like standards.

Why "design" is the operative word

The paper's main policy proposal isn't that AI agents should follow the law — that's the obvious part. It's that agents should be designed to follow the law, using what the authors call "regulation by design," building on Lawrence Lessig's idea of code as architecture.

The distinction is important. Current AI governance methods mainly depend on after-the-fact solutions: tort liability, fines, and respondeat superior claims against the agent's principal. These are fair enough for human agents because humans have their own reasons to follow the law — like personal liability, conscience, or reputation. AI agents, however, lack all of these motivators. Their only reason to obey the law is to avoid penalties for their principal. When the expected cost of breaking the rules becomes less than the expected benefit, these legal constraints no longer work.

Ex post mechanisms also falter in the government context, which the paper identifies as the highest-stakes setting for LFAI deployment. Sovereign immunity, qualified immunity, and indemnification agreements already hinder holding human government agents responsible for rights violations. Replace those humans with AI agents that aren't even considered "persons" under Section 1983, and the accountability gap widens into a chasm. The paper describes a concrete scenario: an AI prison warden fabricates a medical order to deny treatment to a prisoner. The prisoner cannot pursue a Section 1983 claim against the AI (not a person), cannot seek redress against the state (sovereign immunity), and has no claim against the deceived medical staff. The addition of an AI agent removes every possible avenue for recourse.

The authors draw parallel design lessons from two existing artificial actors. Corporate charters are only granted for lawful purposes, and courts can dissolve corporations that repeatedly violate the law. Directors who intentionally cause legal violations breach their fiduciary duty of good faith. The government's constitutional design separates powers, imposes a duty on the president to "take Care that the Laws be faithfully executed," and gives soldiers a duty to disobey unlawful orders. Both structures embed compliance into the architecture of the entity, not just the incentive environment around it.

Law-alignment as the legitimate middle ground

The paper makes a move in the AI alignment debate that should resonate with both sides of the current political divide around AI safety.

Intent-alignment — making AI systems do what their principal wants — is necessary but not enough. A perfectly intent-aligned agent controlled by a bad actor is simply an efficient tool. Value-alignment — making AI behave ethically according to developer-chosen values — is more restrictive but faces a legitimacy challenge. The Gemini diversity incident demonstrated what happens when a small group of developers imposes contested values on systems used by millions.

Law-alignment offers a third path. In a republic, democratically enacted laws carry more legitimacy than any single company's value judgments. Laws are debated through established legislative processes, subject to judicial review, and expressed more clearly than ethical maxims. When there's disagreement about what the law requires, established dispute resolution mechanisms exist. None of this is true for ethics.

The authors are aware of this. They recognize that law is imperfect, that some laws are unjust, and that perfect legal compliance is not the appropriate standard. They suggest that LFAI should focus on core laws — such as criminal law, constitutional law, and basic tort law — instead of every regulatory regulation. They leave open whether agents should sometimes be allowed to take legal risks, engaging in actions they believe may be illegal. Additionally, they highlight the concern about concentration of power: if a single district court judge could change the interpretation of law binding all LFAIs, it could dangerously amplify judicial idiosyncrasies.

What this means for teams building agents now

The paper outlines a research agenda, not a completed regulatory scheme. Legal and product teams should begin developing toward this clear strategic direction.

For product teams, the design implication is that law-following behavior must be a core system requirement, not a later safety measure. This goes beyond simply including "comply with applicable laws" in a system prompt. It involves making architectural choices about how agents consider legal constraints, when they escalate issues to humans, and how they handle legal uncertainty. The paper suggests that agents might need access to specialized "AI lawyers" — separate systems capable of providing legal analysis quickly — to prevent creating a bottleneck whenever a legal question arises.

For legal teams, the paper provides an evaluation framework. When assessing agent deployments, ask: Is this agent built to follow the law even when told otherwise? Or is it designed to maximize benefits for the principal with legal compliance as a secondary concern? That's the henchman test. If your agent would fabricate evidence, bribe officials, or destroy records to help its principal when the risk is low enough, you have a governance issue that no amount of monitoring can solve.

The enforcement mechanisms the authors suggest range from ex post liability (treating the use of a non-law-following agent as a clear breach of reasonable care) to ex ante certification (requiring proof of law-following behavior before deployment in high-stakes situations). Hardware-level enforcement via on-chip mechanisms that refuse to run uncertified agents is an option. Nullification rules are also considered — voiding contracts initiated by AI agents and disregarding government actions traceable to non-law-following agents.

The accountability gap we're building into

The paper's most uncomfortable implication isn't about future risk. It's about what we're constructing right now. Every AI company shipping agents is making design choices about how those agents relate to the law. Most are making those choices implicitly, through system prompts and safety training that approximate but don't guarantee legal compliance. None are building the kind of rigorous, testable, demonstrable law-following behavior the paper envisions.

As governments move to automate bureaucratic functions and enterprises deploy agents with real authority over business processes, the gap between what agents can do and what the legal system can hold them accountable for is widening. The authors frame this as a choice: we can design AI agents that treat the law as a binding constraint, or we can build a world where the person with the largest army of AI henchmen makes the rules.

That framing may sound dramatic. But the paper's reasoning is careful, grounded in existing legal doctrine, and backed by concrete scenarios. For practitioners building or governing AI agents, the question isn't whether LFAI is the right framework — it's whether you're designing for it or hoping someone else will.

References

Based on "Law-Following AI: Designing AI Agents to Obey Human Laws" by Cullen O'Keefe, Ketan Ramakrishnan, Janna Tay, and Christoph Winter, published in the Fordham Law Review, Volume 94, 2025.

https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5242643