Machines controlled by artificial intelligence will become as common as electricity and smartphones. But what happens when autonomous cars, robotic surgeons and thinking industrial devices make deadly mistakes?
IN a futuristic kitchen on display at a recent tech fair in Las Vegas, robot arms from an electronics company could be seen doing what we thought were exclusively human tasks – cooking a meal, making coffee, loading a dishwasher. While this vision of high-tech helpers may be appealing to anyone who hates tidying up a kitchen, what happens when your robot barista spills hot coffee on you?
It’s an age-old legal precept that holds – at least in principle – in all human societies: If you injure someone, you’re liable. But the rise of high-tech robotics and AI (artificial intelligence) poses new quandaries for judicial authorities in this matter. What if a robot causes a serious or fatal accident? A judge can’t order it to pay compensation or put it in prison. There’s growing discussion of the legal ramifications of AI in judicial and political circles, as well as industrial concerns and insurance implications.
Experts agree that within a few decades, the use of self-driving cars, intelligent industrial robots and other autonomously acting machines will be as common as electricity and smartphones are today. Practical questions of civil law still dominate, mainly with respect to fully or semi-autonomous cars.
“In order for highly and fully automated driving to be widely accepted by society, the ultimate bearer of liability must always be clarified,” says Joachim Mueller, director of property and casualty insurance for Munich-based insurance services provider Allianz Deutschland AG. Since 2017, Germany’s road traffic act has stipulated what kind of data a vehicle must – and may – store.
“What still isn’t regulated, though, is how the lawfully authorised parties can technically and organisationally get access to this data,” Mueller says. “This is a matter of equal interest to law enforcement authorities and insurers. “Many questions arise here. Does there have to be a time consuming read-out of data at the car dealership for all parties concerned?
“How will it be ensured that incriminating data isn’t destroyed? What happens to the data if a vehicle is scrapped, and who deletes the data if the vehicle is sold?” Allianz is pushing to keep car manufacturers from having sole control over the data.
“In my view, the data should be in the hands of a neutral, independent third party – a data trustee – to enable all authorised parties to get access to them under the same legal conditions,” Mueller says. The rise of robots has consequences for criminal law too. Legal scholars are already considering far-reaching questions since criminal law is based on the principle of individual guilt. Even the most intelligent machine has neither consciousness nor a conscience, however.
“Directed at individual persons, criminal law has difficulty keeping up with the development of autonomous machines and artificial intelligence,” notes Susanne Beck, a professor of criminal law and philosophy of law at the University of Hanover in Germany. “In criminal law, the operator of a machine normally bears responsibility,” she says. “If Google gives you false information and you base a decision on it, you’re responsible.” The situation is simpler in civil law.
“In some countries a car owner is already liable for accidents in which he or she may not have been involved,” points out Nicolas Woltmann, an assistant at the Research Centre for Robot Law at the University of Wuerzburg in Germany.
“A company, too, can be liable in civil law” – in contrast to criminal law – he says.
“Only people can commit crimes, not corporate bodies or machines.”
But would it be appropriate to penalise a person when AI is at the wheel and in control? “The whole point of these machines is not having to act and decide yourself,” Beck says.
“If you still bear full criminal liability as the driver, then you don’t need an autonomous vehicle, because you’ve got to concentrate just as you would if you were driving yourself. I see that as a problem.” To address the problem, ethical and legal experts have introduced the concept of “meaningful human control”.
“Before a person is held criminally accountable, a close look must be taken at whether he or she exercised any control over the machine,” Beck says. “To a considerable extent we’ll probably have to forgo judicial decisions on accidents caused by machines in the future,” she predicts, conceding that “it can definitely be unsettling for a society if no sentences are possible.”
Her proposal: “Other solutions will have to be sought, for example victim-offender mediation or solutions in other legal systems – civil law, for instance.” Most experts aren’t in favour of fundamental changes to criminal law. “Legal scholars have tentatively begun to discuss whether we have to abandon the principle that only people can behave culpably,” Woltmann says. “Our prevailing opinion is that at the present stage of technological development, there’s no need to discard the established legal norm.” But there’s indeed “a grey area where we can’t forecast how jurisprudence will develop in the future”, he adds.
“For society at large, a grey area may be acceptable if the number of (traffic) accidents falls as self-driving cars become more prevalent owing to their overall benefits for society.” This would be no consolation to accident victims and their families, though. “On an individual level, it’s certainly unsatisfying if criminal liability can’t be established after a serious accident,” Woltmann says. “That’s a dilemma.”