Abstract
In this paper, I draw on Hannah Arendt’s notion of ‘banality of evil’ to argue that as long as AI systems are designed to follow codes of ethics or particular normative ethical theories chosen by us and programmed in them, they are Eichmanns destined to commit evil. Since intelligence alone is not sufficient for ethical decision making, rather than strive to program AI to determine the right ethical decision based on some ethical theory or criteria, AI should be concerned with avoiding making the wrong decisions, and this requires hardwiring the thinking activity as a prerequisite for decision making.
Similar content being viewed by others
Notes
Professor Stuart Russell of University of California Berkley believes that it is imperative that human values be translated and programmed into AI systems. (Goldhill, 2015)
In contrast, “implicit ethical agents” behave according to some designed functionality.
Asimov’s laws of robotics:
1. A robot may not injure a human being, or, through inaction, allow a human being to come to harm.
2. A robot must obey the orders given it by human beings except where such orders would conflict with the first law.
3. A robot must protect its own existence as long as such protection does not conflict with the first or second law. (Asimov, 1984)
This is the title of a book by Hannah Arendt.
Martin Heidegger would disagree with this, pointing out that “thinking only begins at the point where we have come to know that Reason, glorified for centuries, is the most obstinate adversary of thinking.” (cited in Mini, 1994)
References
Allen, C., & Wallach, W. (2011). Wise machines? On the Horizon, 19(4), 253–358.
Anderson, M., & Anderson, S. L. (2007). Machine ethics: Creating an ethical intelligent agent. AI magazine, 28(4), 15–15.
Anderson, S. L. (2008). Asimov’s “three laws of robotics” and machine metaethics. AI & SOCIETY, 22(4), 477–493.
Arendt, H. (1961). Between Past and Present. The Viking Press.
Arendt, H. (1971). Thinking and moral considerations: A lecture. Social Research, 51(1/2), 7–37.
Arendt, H. (1978). The Life of the mind. Harvest.
Arendt, H. (2003). Responsibility and Judgment. In J. Kohn (Ed.).
Arendt, H. (2006). Eichmann in Jerusalem: Penguin Books.
Armstrong, S., Sandberg, A., & Bostrom, N. (2012). Thinking inside the box: Controlling and using an oracle AI. Mind and Machine, 22(4), 299–324.
Arnold, T., & Scheutz, M. (2018). The “big red button” is too late: an alternative model for the ethical evaluation of AI systems. Ethics and Information Technology, 20, 59–69.
Asimov, I. (1984). The Bicentennial Man‟ in I. Asimov, The bicentennial man and other stories. In: Doubleday, New York.
Bostrom, N. (2002). Existential risks: Analyzing human extinction scenarios and related hazards. Journal of Evolution and technology, 9.
Botică, D. A. (2017). Artificial Intelligence and the Concept of “Human Thinking”. In Business Ethics and Leadership from an Eastern European, Transdisciplinary Context (pp. 87-94): Springer.
Bringsjord, S., & Taylor, J. (2012). The divine-command approach to robot ethics (pp. 85–108). The ethical and social implications of robotics.
Brundage, M. (2014). Limitations and risks of machine ethics. Journal of Experimental & Theoretical Artificial Intelligence, 26(3), 355–372.
Crawford, K., & Calo, R. (2016). There is a blind spot in AI research. Nature, 538, 311–313.
Critchley, S. (2014). Ethics of Deconstruction: Edinburgh University Press.
Davis, E. (2015). Ethical guidelines for a superintelligence. Artificial Intelligence, 220, 121–124.
Dehaene, S., Lau, H., & Kouider, S. (2017). What is consciousness, and could machines have it? Science, 358.
Dellinger, A. J. (2015). Tim Wu says Google is degrading the Web to favor its own products. The Daily Dot. http://www.dailydot.com/technology/google-search-tim-wu-yelp/
Di Nucci, E. (2019). Should we be afraid of medical AI? Journal of Medical Ethics, 45(8), 556–558.
Etzioni, A., & Etzioni, O. (2016). AI assisted ethics. Ethics and Information Technology, 8(2), 149–156.
Etzioni, A., & Etzioni, O. (2017). Incorporating ethics into artificial intelligence. Journal of Ethics, 21, 403–418.
Fazi, M. B. (2019). Can a machine think (anything new)? Automation beyond simulation. AI & Society, 34(4), 813–824.
Fröding, B., & Peterson, M. (2020). Friendly AI. Ethics and Information Technology. https://doi.org/10.1007/s10676-020-09556-w
Goldhill, O. (2015). Human values should be programmed into robots, argues a computer scientist. Quartz. https://qz.com/538260/human-values-should-be-programmed-into-robots-argues-a-computer-scientist/
Greene, J. D. (2014). Beyond point-and-shoot morality: Why cognitive (Neuro)science matters for ethics. Ethics, 695-726.
Guarini, M. (2006). Particularism and the classification and reclassification of moral cases. IEEE Intelligent Systems, 21(4), 22–28.
Heidegger, M. (1999). Contributions to Philosophy: From Enowning: Indiana University Press.
Himma, K. E. (2009). Artificial agency, consciousness, and the criteria for moral agency: What properties must an artificial agent have to be a moral agent? Ethics and Information Technology, 11(1), 19–29.
Johnson-Laird, P. N. (2009). Human and Machine Thinking.
Lin, P., Abney, K., & Bekey, G. (2011). Robot ethics: Mapping the issues for a mechanized world. Artificial Intelligence, 175, 942–949.
Mini, P. V. (1994). Existentialism and Keynes. In John Maynard Keynes (pp. 148-163): Springer.
Misselhorn, C. (2019). Artificial Systems with Moral Capacities? A research design and its implementation in a geriatric care system. Artificial Intelligence. https://doi.org/10.1016/j.artint.2019.103179
Monin, A. S. (1992). On the definition of the concepts thinking, consciousness, and conscience. Proceedings of the National Academy of Sciences, 89(13), 5774–5778.
Moor, J. H. (2006). The nature, importance, and difficulty of machine ethics. IEEE Intelligent Systems, 21(4), 18–21.
Nath, R., & Sahu, V. (2020). The Problem of Machine Ethics in Artificial Intelligence. AI & SOCIETY, 35(1), 103–111.
Nietzsche, F. (2009). Twilight of the Idols. http://www.lexido.com/ebook_texts/twilight_of_the_idols.aspx
Penrose, R. (1991). The Emperor's New Mind: Concerning Computers, Minds, and the Laws of Physics: Penguin.
Pereira, L. M., & Saptawijaya, A. (2016). Programming machine ethics (Vol. 26): Springer.
Rapaport, W. J. (1993). Because mere calculating isn’t thinking. Mind and Machine, 3, 11–20.
Schlagel, R. H. (1999). Why not artificial consciousness or thought? Mind and Machine, 9, 3–28.
Searle, J. R. (1990). Is the brain’s mind a computer program? Scientific American, 262(1), 25–31.
Searle, J. R. (2014). What your computer can’t know. The New York review of books, 9.
Sharkey, A. (2020). Can we program or train robots to be good? Ethics and Information Technology, 22, 283–295.
Stuart, S. A. J. (2007). Machine consciousness: Cognitive and kinaesthetic imagination. Journal of Consciousness Studies, 14(7), 141–153.
Tajalli, P., & Segal, S. (2019). A nietzschean re-evaluation of values as a way of re-imagining business ethics. Business Ethics: A European Review, 28(2), 234–242. https://doi.org/10.1111/beer.12213
Torrance, S. (2008). Ethics and consciousness in artificial agents. AI & SOCIETY, 22(4), 495–521.
Vandenberg, B. (1999). Levinas and the ethical context of human development. Human Development, 42, 31–44.
Vanderelst, D., & Winfield, A. (2018). The Dark Side of Ethical Robots. Paper presented at the Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society.
Watson, D. (2019). The rhetoric and reality of anthropomorphism in artifcial intelligence. Mind and Machine, 29, 417–440.
Wolchover, N. (2015). Concerns of an artificial intelligence pioneer. Quanta. April 21. In.
Yudkowsky, E. (2007). Levels of organization in general intelligence. In B. Goertzel & C. Pennachin (Eds.), Artificial General Intelligence (pp. 389–501). Springer.
Zalta, E. N. (2019). Hannah Arendt. Metaphysics Research Lab, Stanford University.
Funding
There has been no financial support for this work.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
There are no known conflicts of interest associated with this work.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Tajalli, P. AI ethics and the banality of evil. Ethics Inf Technol 23, 447–454 (2021). https://doi.org/10.1007/s10676-021-09587-x
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10676-021-09587-x