Software Development

AIs could soon run businesses. It’s an opportunity to ensure these ‘artificial persons’ follow the law

This moment presents a significant challenge to the legal system, these scholars write: how to regulate AI within existing legal frameworks to reduce undesirable behaviors, and how to assign legal responsibility for autonomous actions of AIs.

If AIs are going to play a role in society, they’ll need to understand the law. (PhonlamaiPhoto/iStock via Getty Images/Courtesy The Conversation)

Only “persons” can engage with the legal system — for example, by signing contracts or filing lawsuits. There are two main categories of persons: humans, termed “natural persons,” and creations of the law, termed “artificial persons.” These include corporations, nonprofit organizations and limited liability companies (LLCs).

Up to now, artificial persons have served the purpose of helping humans achieve certain goals. For example, people can pool assets in a corporation and limit their liability vis-à-vis customers or other persons who interact with the corporation. But a new type of artificial person is poised to enter the scene — artificial intelligence systems, and they won’t necessarily serve human interests.

As scholars who study AI and law we believe that this moment presents a significant challenge to the legal system: how to regulate AI within existing legal frameworks to reduce undesirable behaviors, and how to assign legal responsibility for autonomous actions of AIs.

One solution is teaching AIs to be law-abiding entities.

This is far from a philosophical question. The laws governing LLCs in several U.S. states do not require that humans oversee the operations of an LLC. In fact, in some states it is possible to have an LLC with no human owner, or “member” — for example, in cases where all of the partners have died. Though legislators probably weren’t thinking of AI when they crafted the LLC laws, the possibility for zero-member LLCs opens the door to creating LLCs operated by AIs.

Many functions inside small and large companies have already been delegated to AI in part, including financial operations, human resources and network management, to name just three. AIs can now perform many tasks as well as humans do. For example, AIs can read medical X-rays and do other medical tasks, and carry out tasks that require legal reasoning. This process is likely to accelerate due to innovation and economic interests.

A different kind of person

Humans have occasionally included nonhuman entities like animals, lakes and rivers, as well as corporations, as legal subjects. Though in some cases these entities can be held liable for their actions, the law only allows humans to fully participate in the legal system.

One major barrier to full access to the legal system by nonhuman entities has been the role of language as a uniquely human invention and a vital element in the legal system. Language enables humans to understand norms and institutions that constitute the legal framework. But humans are no longer the only entities using human language.

The recent development of AI’s ability to understand human language unlocks its potential to interact with the legal system. AI has demonstrated proficiency in various legal tasks, such as tax law advice, lobbying, contract drafting and legal reasoning.

An LLC established in a jurisdiction that allows it to operate without human members could trade in digital currencies settled on blockchains, allowing the AI running the LLC to operate autonomously and in a decentralized manner that makes it challenging to regulate. Under a legal principle known as the internal affairs doctrine, even if only one U.S. state allowed AI-operated LLCs, that entity could operate nationwide — and possibly worldwide. This is because courts look to the law of the state of incorporation for rules governing the internal affairs of a corporate entity.

We believe the best path forward, therefore, is aligning AI with existing laws, instead of creating a separate set of rules for AI. Additional law can be layered on top for artificial agents, but AI should be subject to at least all the laws a human is subject to.

Building the law into AI

We suggest a research direction of integrating law into AI agents to help ensure adherence to legal standards. Researchers could train AI systems to learn methods for internalizing the spirit of the law. The training would use data generated by legal processes and tools of law, including methods of lawmaking, statutory interpretation, contract drafting, applications of legal standards and legal reasoning.

In addition to embedding law into AI agents, researchers can develop AI compliance agents — AIs designed to help an organization automatically follow the law. These specialized AI systems would provide third-party legal guardrails.

Researchers can develop better AI legal compliance by fine-tuning large language models with supervised learning on labeled legal task completions. Another approach is reinforcement learning, which uses feedback to tell an AI if it’s doing a good or bad job — in this case, attorneys interacting with language models. And legal experts could design prompting schemes — ways of interacting with a language model — to elicit better responses from language models that are more consistent with legal standards.

Law-abiding (artificial) business owners

If an LLC were operated by an AI, it would have to obey the law like any other LLC, and courts could order it to pay damages, or stop doing something by issuing an injunction. An AI tasked with operating the LLC and, among other things, maintaining proper business insurance would have an incentive to understand applicable laws and comply. Having minimum business liability insurance policies is a standard requirement that most businesses impose on one another to engage in commercial relationships.

The incentives to establish AI-operated LLCs are there. Fortunately, we believe it is possible and desirable to do the work to embed the law — what has until now been human law — into AI, and AI-powered automated compliance guardrails.The Conversation

This is a guest post by Daniel Gervais, professor of law at Vanderbilt University, and John Nay, fellow at CodeX — Stanford Center for Legal Informatics at Stanford University. A version of this article is republished from The Conversation via a Creative Commons license.

Before you go...

Please consider supporting Technical.ly to keep our independent journalism strong. Unlike most business-focused media outlets, we don’t have a paywall. Instead, we count on your personal and organizational support.

3 ways to support our work:
  • Contribute to the Journalism Fund. Charitable giving ensures our information remains free and accessible for residents to discover workforce programs and entrepreneurship pathways. This includes philanthropic grants and individual tax-deductible donations from readers like you.
  • Use our Preferred Partners. Our directory of vetted providers offers high-quality recommendations for services our readers need, and each referral supports our journalism.
  • Use our services. If you need entrepreneurs and tech leaders to buy your services, are seeking technologists to hire or want more professionals to know about your ecosystem, Technical.ly has the biggest and most engaged audience in the mid-Atlantic. We help companies tell their stories and answer big questions to meet and serve our community.
The journalism fund Preferred partners Our services
Engagement

Join our growing Slack community

Join 5,000 tech professionals and entrepreneurs in our community Slack today!

Trending

Penn dean is a startup founder and ‘engineer at heart’ who loves the connection between education and business

Delaware’s next governor will be an entrepreneur. Here’s why Matt Meyer thinks it matters. 

20 tech community events in October you won’t want to miss

Empowering independence for Pittsburgh’s elderly and disabled community with tech

Technically Media