In the previous post (Part I: Automation & Ethics:Applying Trolleyology), we discussed the introductory part to this article; the
inherent ethical dilemma in the trolley problem and how that’s relevant to the
current scenario of automated devices and self-driving cars. We discussed
certain situations, including the Moral Machine by MIT, and our discussion was
primarily ethical. In this post, we’ll carry forward the discussion from an
ethical one to a legal one; discussing particular significant legal
implications of the issues discussed beforehand.
We discussed ethics in the previous post and observed
that people have varying stands when it comes to categorising what particularly
is ethical and what is unethical. Despite this personal biases and difference
of opinions, every society has certain ethical standards and it enforces the
same by way of punishing the unethical. Theft is punished, trespassing is
punished, defaming is punished; killing is punished, too, be it five workmen or
a single one. Your being confused as to moral reasoning isn’t a valid excuse
in the eyes of the law, and the law will punish what it ‘considers’ to be
wrong.
But, what if AI commits a wrong? Whom to hold
responsible when an artificially intelligent device or self-driving car causes
some damages or injuries? Will the coder/designer/manufacturer be liable? Or
the owner? Or the person who customised it? Big question, eh.
Legal personhood
of AI
Examining the legal liability of AI, we can start by
examining the possibilities of a legal personality of AI in the first place. Can
an artificially intelligent device/robot be considered a juristic person in the
eyes of law? If not, what would be its legal status, then? Can it still be
treated merely as a movable property despite the fact that it has an
intelligence and ability to learn things its own way?
The logical starting point would be to consider the
probability of AI falling within the definition of mere property, like a
watch or a remote controller. When it comes to products, the manufacturer is
held liable by the principle of strict liability. However, with recent developments, there has been a shift in the
judicial approach and liability is held only in cases where the defendant
manufacturer/designer could reasonably foresee the damage or has failed to take
reasonable care to avoid negligence.
Deducing on this line of reasoning, we can realise
that the manufacturer or designer of an artificially intelligent being/device
can rarely foresee the course of action its creation would take, and hence
should be exempted from the strict liability – hence AI not falling strictly
within the definition of a property per se.
The next point of examination would be whether AI
could be a person – the option on the opposite end of the logical extremity. It
requires little or no examination, in my opinion. Every one of us would agree on
this point – that although artificially intelligent devices possess some degree
of intelligence, they can in no way be considered ‘equivalent’ to humans.
So, AI doesn’t qualify for a human, and neither does it
fall within the categorisation of a property. What’s left? A middle way: Romans
invented this when confronted with the question as to whether slaves were
humans. They didn’t want to confer upon the slaves complete personhood; but
they also understood that slaves had their own minds and wouldn’t necessarily
follow the master always. They invented this new legal status – ‘quasi-persons’.
On a sliding scale, with humans like you and
I being on one end; and the non-living things like trees, pens, and pencils on
the other; and the corporations, juristic persons falling somewhere in the
middle; AI also finds a place somewhere in the middle. Where exactly in the
middle? Well, that’s a difficult question to answer objectively. Some tests
like the Legal Turing test have been proposed to address the same, but they
somehow fail when it comes to practical applicability. The best bet would be to
consider this issue on a case-to-case basis, depending upon the facts of the
case like the intention of the designer, foreseeability of the damage et
cetera.
Legal liability of
AI
Now that we have an answer of sorts regarding the
legal personhood of AI, we can proceed to discuss the legal liability of
artificially intelligent devices. And Gabriel Hallevy from Israel makes the job
easier for us with his work. A little background for non-law people here:
criminal liability has two basic prerequisites: mens rea (the criminal intent, or the guilty mind) and actus reus (the criminal act). In other
words, a man, or any person, can be said to be criminally liable only when both
the elements are present – that is, he has criminal intent and commits an act
to actualise that intent.
There can be a number of instances of various
combinations of these elements, and each of those instances calls for different
degree of liability. Of those many, Gabriel points out three possible scenarios
that can be applicable to the AI; the first one being perpetrator via another. Perpetrator via another is the situation
when a mentally deficient person or an animal, who inherently lacks mens rea
and is hence innocent, commits a crime under the instructions of someone else. The
instructor here is held liable. The second one is natural probable consequence which relies on the possibility that
the programmer could naturally see the probable consequence and decides
liability accordingly. The third one is direct
liability, and this one is special: under this set of circumstances, the AI
itself is liable – not the owner.
AI liability is no longer just a theoretical
discussion or philosophical deliberation. Way back in 2008, the Indian
judiciary faced the case of Avnish Bajaj v State (famously known as the Bazee.com case, wherein the courts
recognised the automation involved in codes and software for which the owner
could not be held liable. Time and again courts across the globe have come
across cases where they have admitted that certain things aren’t ripe for a decision
yet, unfortunately.
So, if the AI is liable in certain cases to the
exclusion of all other stakeholders, including the owner himself, the next big question
that arises is, whom to punish? Whom to give judgements to? Whom to accuse? The
conventional theories of punishments obviously aren’t going to apply to AI – we
can’t think of retribution or reform by punishing an AI system. So, whom to punish?
Well, nobody knows
the answer for now.