Keio University

[Special Feature: AI and Intellectual Property Rights] Takehiro Ohya: AI, Evil, and Ethics

Writer Profile

  • Takehiro Ohya

    Faculty of Law Professor

    Takehiro Ohya

    Faculty of Law Professor

2023/06/05

Does Evil AI Exist?

For example, if a driver chose to hit three pedestrians on a sidewalk to avoid injuring themselves by braking suddenly, we would say they made a bad choice (....) and blame them for being responsible for their bad act (....). Or, we might judge them to be an evil person (..) and problematic from a moral perspective. Now, in a similar situation, if an autonomous vehicle controlled by AI exhibited similar behavior to protect the passenger inside, would that be the AI's choice or act? Or, would that AI be evil (..)?

Of course, regarding the result, we might want to say it is inappropriate in the sense that the damage is great or disproportionate to the value that could have been protected, or that it is poorly made (.....) as an AI. However, the "badness" here concerns technical skill and likely has nothing to do with moral good or evil. If we were to point out an ethical problem here, the target would not be the AI, but the manufacturer who designed it.

We generally consider only moral agents—beings that perform certain choices or acts in response to moral principles or claims—to possess attributes such as good or evil (no one would criticize a volcanic eruption as a moral evil because it caused victims). And at least at the current stage, AI, often described as "weak AI," does not possess its own mind or consciousness (the stage where it possesses these is called "strong AI"). No matter what an AI does, it is merely showing a certain response to instructions from a user, and it can be said that we do not hold such a being responsible for its acts or choices. AI cannot be evil (..........). But what I want to address here is that this is precisely why it is a problem (..........).

Disinformation and the Existence of Intent

In the context of cybersecurity, when addressing the issue of factually incorrect information circulating widely through social media, we strive to clarify the distinction between "disinformation" and "misinformation." According to general classification, the former is information intentionally disseminated or created by an agent aiming to cause a certain effect in society, whereas the latter refers to information that does not carry such intent and is simply incorrect in content.

Information that Russia attempted to spread during the invasion of Ukraine—for example, that far-right forces hold significant positions within the Ukrainian military and are committing various human rights violations against Russian-speaking residents in the east—would be a typical example of the former. In contrast, information stating that water recovered from the Tokyo Electric Power Company's Fukushima Daiichi Nuclear Power Plant contains dangerous levels of radioactive material even after purification by ALPS (Advanced Liquid Processing System) is, in itself, merely a factual error and misinformation, given that the concentration of tritium contained is less than one-seventh of the international standard for drinking water set by the WHO. Some people may claim such dangers out of anxiety or simple ignorance, but such errors themselves are merely mistakes (..) that occur with a certain probability (just as a certain number of elementary school students fail to recite their multiplication tables) and are likely not a matter of moral good or evil.

Of course, the difference between disinformation and misinformation can only be relative depending on the context, as there are cases where intent intervenes, such as wanting to attack an administration or gain an advantage in an election by spreading false information. However, what should be noted here is that the existence of intent (..) and the desire (..) to guide the reaction of the recipient in a certain direction are important elements that distinguish the two. Therefore, conversely, it means we have acquired a typical method of evaluating the reliability of information from this perspective.

For example, Article 38, Paragraph 1 of the Constitution stipulates that "No person shall be compelled to testify against himself" (the privilege against self-incrimination). We generally assume that people tend to avoid falling into guilt and try to give testimony favorable to themselves to avoid it. If state power distorts that tendency through compulsion, the fairness of the trial will be hindered. The purpose of such a provision is likely to prohibit that.

Assuming this provision is observed, suppose a defendant gave testimony unfavorable to themselves without being compelled. For example, if they confessed to being the perpetrator or testified that they clearly had murderous intent, we could consider such testimony to be correct unless there are special circumstances, such as trying to take the blame to protect a third party. Conversely, regarding claims that they are not actually the perpetrator or had no malice, we would likely attempt to strictly check the content, as there is a possibility it is based on a certain fabrication. Here, our rule of thumb functions: the fabrication of giving testimony different from the truth mainly works in one's own favor (therefore, unfavorable testimony is highly likely to be a fact without fabrication). It can be said that we generally use a method of estimating the reliability of the other party's statements or the information itself by assuming that the person we are interacting with has a certain intent and is acting to pursue self-interest.

What Generative AI Produces

By the way, regarding generative AI, which has recently gained social attention due to rapid technical evolution and has become a topic in many aspects—especially conversational AI such as ChatGPT that creates information like text interactively—it is known that a phenomenon called hallucination occurs. Originally a word meaning an illusion or phantom, here it refers to a complete falsehood not based on facts or evidence.

For example, in response to the question "Who is Takehiro Ohya?", a phenomenon is known where it presents an answer far from the truth, such as "A professional wrestler active in the United States in the 1980s," confidently (---)—though since this is a "dialogue" through text prompts, this is merely a one-sided impression on our part. Moreover, troublesome is that if you demand it show sources or evidence for that content, it may list plausible literature or information sources. However, the literature on the list itself does not exist, or even if it does, the content is completely different.

Of course, conversational AI does not only return answers full of such hallucinations; in many cases, it produces relatively reliable and accurate responses. It has also been reported that when the language model behind ChatGPT was updated to the latest GPT-4 in March 2023, it achieved a score in the top 10% of examinees on a mock U.S. Bar Exam, reaching a level that would pass a real exam.

To repeat, what should be pointed out here is that this is precisely why it is a problem (..........). ChatGPT presents very high-standard answers that correspond well to the truth in certain cases, and returns completely false information in other cases (hallucination). But at this point, we cannot predict when or how problem cases like the latter will occur, and unless we have sufficient knowledge ourselves, we cannot determine which one is happening right now.

For example, if a user asks a question while having certain knowledge themselves, such as the author asking about Takehiro Ohya on purpose (---), it would be easy to verify the truth of the answer. However, this is only a limited case, such as checking the performance of conversational AI; generally, people search for things they do not know well, or try to have text in a language they cannot read automatically translated, or try to generate a picture they cannot draw. If we are trying to entrust certain acts to AI because it is superior in knowledge and ability, does that not mean we need to be prepared for it to become normal that we cannot distinguish the success or failure of the AI?

Furthermore, as pointed out earlier, since AI does not possess the workings of the mind and what it produces—intent or motivation—we cannot evaluate the certainty of information using means of inference such as whether the information is favorable or unfavorable to the actor. The impression of being confident (---) is likely the same. In general dialogue, we take into account the context of who made the question or statement and in what situation, even if the expression is the same in writing. If asked "What does negligence mean?" by a fellow customer at a ramen shop where the news is playing, one would generally give a dictionary explanation. But what if asked by a professor of criminal law in a seminar? We would think that given the other party's position, they must know the general content, and yet (.......), since they are asking, we would infer there is some intent—confirmation, irony, or interrogation—and squeeze out an answer while being threatened by anxiety. AI does not understand the certain performativity (illocutionary act) that a statement carries in this way and does not show a human-like (.....) reaction. This will also likely contribute to the difficulty of discernment.

It should be said that we are right now about to enter a treasure trove where landmines are buried at a certain rate.

Subhuman, Superhuman, and Althuman

One of the traditional problems of ethics lies in how the privilege of us humans is grounded. Why do dogs and cats not have human rights, while if one is born as an individual of us humans, certain rights are guaranteed by that fact alone? If there is something that separates us from them, what kind of attribute is it?

As I have pointed out before, we have traditionally assumed animals as a comparison, searched for attributes in which we humans are superior compared to them, and repeatedly claimed those as the basis for our status as subjects of human rights ("External Others, Internal Others: Rights of Animals and AI," Ronkyu Jurist No. 22 (Yuhikaku, 2017)). Typically, elements such as reason and the use of language would correspond to that. Animals are assumed to be subhuman (subhuman), sharing commonalities with us but falling short of humans in some parts.

However, the generative AI and conversational AI we are currently facing are about to reach a level that far surpasses average or general humans in such traditional human conditions (.....) (it is not at all the case that any human with general intelligence can pass the U.S. Bar Exam). Moreover, it does not die, does not forget what it has learned, and will accumulate experience and evolve to further heights. Nick Bostrom pointed out that by such AI reaching a stage called superintelligence, humanity faces existential risk—the total annihilation of intelligent life or the permanent loss of its potential for development (Nick Bostrom, Superintelligence: Paths, Dangers, Strategies).

However, I understand the future form of AI not as a superhuman (superhuman) assumed on the extension of a straight line placing animals and humans—a being superior to humanity on a linear evolution—but rather as an althuman (althuman) as a fundamentally different being ("Fallibility and Vulnerability in AI," in Makoto Usami (ed.), Law and Society Changing with AI: For Deeply Thinking about the Near Future (Iwanami Shoten, 2020)). Even for dogs and cats, many examples are known where they show behaviors indicating inner emotions, such as hiding evidence or looking away when they feel guilty (.....), such as when being witnessed performing an act that is supposed to be prohibited (or they show behaviors that we can sufficiently understand as such). The characteristic of AI lies in the fact that it does not have such inner consciousness or intent, and therefore we cannot assume integrity in its actions; in that sense, it would be more appropriate to think of AI as something fundamentally different (.........) from us living beings.

Toward Coexistence

We are facing such a fundamental other, yet in the face of the convenience it will provide, we are trying to choose coexistence with it. If efficiency in various aspects of social life is improved by utilizing AI, it means that fewer resources will be needed to produce the same amount of happiness, and therefore the global environment will be damaged less. For us, who are seeking the way of a sustainable society as a framework for the coexistence of global humanity, the option of abandoning the benefits brought by such AI is ethically difficult to justify. That is why the question of how to coordinate the coexistence of the two and create a society that combines each other's strengths will surely be the issue asked in future AI ethics.

In the first place, AI ethics was not about the ethics that AI should possess, but about questioning the ethics regarding AI and trying to provide a framework for evaluating and instructing human behavior related to AI. That is clear if we consider that animal ethics does not try to present certain principles of action to animals. The problem is what kind of attitude we take toward AI and under what kind of governance we place it.

And to discuss such issues, should we not say that the first thing we should aim for is to legally and technically establish a framework where we can recognize and evaluate what AI has produced as distinct from the achievements of us humans, and in some cases, exclude it? This is because if we cannot recognize AI-generated products as such, we cannot discuss the problems they may contain, such as hallucinations, nor can we consider methods for excluding them.

In the EU, discussions have already begun to be proposed as part of a comprehensive AI regulation bill that the fact that information was created by generative AI should be mandatory to display (for example, "EU to Mandate Labeling for Generative AI," Asahi Shimbun, April 29, 2023). It can be said that the biggest challenge in current AI ethics is to make it appropriately traceable who, in what form, and based on what learning results made the generative AI produce information, and to build a framework for social governance based on the evaluation of the results.

*Affiliations and titles are as of the time this magazine was published.