J. Savulescu's "The God Machine" as a Moral Agent and the Problem of Responsibility

 
PIIS023620070026101-8-1
DOI10.31857/S023620070026101-8
Publication type Article
Status Published
Authors
 
Affiliation: Saint-Petersburg State University
Address: 7-9 Universitetskaya nab., Saint-Petersburg 199034, Russian Federation
Affiliation: Saint-Petersburg State University
Address: 7-9 Universitetskaya nab., Saint-Petersburg 199034, Russian Federation
Journal nameChelovek
EditionVolume 34 Issue 3
Pages24-40
Abstract

This article is devoted to the issue of moral responsibility in connection with artificial intelligence technologies. In recent years, artificial intelligence has been actively developing towards greater autonomy, which makes the philosophical analysis of artificial moral agency extremely relevant. The relevance of the work is also determined by the increasing spread of artificial intelligence in professional areas, including those related to the adoption of responsible managerial, financial, etc. solutions. The main object of research in the article is a thought experiment — the project of the so-called “The God Machine” by the Western philosopher J. Savulescu, which allows us to turn to such an important aspect as the possibility of the influence of artificial intelligence on the development of human moral consciousness, since modern technologies, along with projects of biological moral human enhancement, claim the ability to create artificial moral agents. The authors give a comprehensive analysis of this project, correlating it with other concepts of modern analytical philosophy of consciousness, incl. Н. Frankfurt. The theoretical focus of the study is directed to the concepts of responsibility and freedom as key in formulating the criteria of moral agency. At the same time, the article considers the features of the implementation of the criteria of moral agency in relation to artificial intelligence. The article assesses the viability of J. Savulescu's project, in particular: to what extent artificial intelligence like the The God Machine can be considered as a moral agent, whether a person can acquire the characteristics of a moral, responsible and free person with the help of such an artificial mediator, and to what extent such projects contribute to the moral enhancement of mankind. It also points to the risks of paternalistic interference in the development of morality. Separately, the phenomenon of moral dilemmas is considered in connection with the problems of embedding artificial intelligence in the life of modern society, as well as the logical aspects of decision-making in the context of moral conflicts. The conclusion is formulated about the key importance of the category of responsibility for modeling the relationship between man and artificial intelligence in machine ethics.

Keywordsmoral responsibility, moral agency, artificial intelligence, biological moral enhancement, freedom, moral dilemmas, J. Savulescu, H. Frankfurt
Received28.06.2023
Publication date28.06.2023
Number of characters30301
Cite  
100 rub.
When subscribing to an article or issue, the user can download PDF, evaluate the publication or contact the author. Need to register.

Number of purchasers: 0, views: 145

Readers community rating: votes 0

1. Alchurron K.E., Gerdenfors P., Makinson D. Logika teorii izmeneniya: funkcii revizii i sokrasheniya cherez chastichnoe peresechenie [The Logic of the Theory of Change: Revision Functions and Reduction through Partial Intersection]. “Normativnye sistemy” i drugie raboty po filosofii prava i logike norm. St.-Petersburg: Izdatelskij Dom SPbGU Publ., 2013. P. 318–343.

2. Garvardt M., Perova N. Universalism and Relativism of the Moral Agency of Biological Moral Enhancement (on the Example of Political Activity)]. Discourses of Ethics. 2022. N1(13). P. 69–84.

3. Larionov I.Yu. Kto neset otvetstvennost za primenenie oruzhiya s iskusstvennym intellektom? [Who is Responsible for the Use of Weapons with Artificial Intelligence?] // Internet i sovremennoe obshestvo: sbornik tezisov dokladov [Elektronnyj resurs]. Trudy XXV Mezhdunarodnoj obedinennoj nauchnoj konferencii «Internet i sovremennoe obshestvo» (IMS-2022), St.-Petersburg, 23–24 iyunya 2022 g. SPb: Universitet ITMO, 2022. http://ojs.itmo.ru/index.php/IMS/issue/view/78,svobodnyj.

4. Larionov I., Markov B. Moral and Anthropological Risks of Human Bioenhancement Technologies // Discourses of Ethics. 2022. № 4(16). P.25–46.

5. Mill J. O svobode [On Liberty]. Nauka i zhizn. 1993. N 11. P.10–11.

6. Razin A.V. Etika iskusstvennogo intellekta [The Ethics of Artificial Intelligence]. Filosofiya i obshestvo. 2019. N 1. P. 57–73.

7. Searle J. Soznanie, mozg i programmy [Minds, Brains, and Programs]. Analiticheskaya filosofiya: stanovlenie i razvitie. Moscow: Dom intellektual'noj knigi, Progress-Tradiciya Publ., 1998.

8. Filosofiya otvetstvennosti [Philosophy of Responsibility]. Ed by E.N. Lisanyuk, V.Yu. Perova. St.-Petersburg: Nauka Publ., 2014.

9. Frankfurt H. Alternativnye vozmozhnosti i moralnaya otvetstvennost [Alternate Possibilities and Moral Responsibility]. Filosofiya. Zhurnal Vysshej shkoly ekonomiki. 2017. Vol. I, N 4. P.129–140.

10. Allen C., Varner G., Zinser J. Prolegomena to any future artificial moral agent. Journal of Experimental & Theoretical Artificial Intelligence. 2000. Vol.12, N 3. P. 251–261.

11. Anderson M., Anderson S.L. Machine Ethics: Creating an Ethical Intelligent Agent. AI Magazine. 2007. N 4(28). P.15–26.

12. Awad E., Dsouza S., Kim R. et al. The Moral Machine Experiment. 2018. Nature. Vol. 563. P. 59–64.

13. Dennett D. When HAL Kills, Who’s to Blame? Computer Ethics // Stork, David, HAL’s Legacy: 2001's Computer as Dream and Reality. MIT Press, 1998.

14. Floridi L., Sanders J. Artificial Evil and the Foundation of Computer Ethics. Ethics and Information Technology. 2001. N 3(1). P. 56–66.

15. Frankfurt H. Freedom of the Will and the Concept of a Person. The Journal of Philosophy. 1971. Vol. 68, N 1. P. 5–20.

16. Harris J. Moral Enhancement and Freedom. Bioethics. 2011. Vol. 25. P.102–111.

17. Himma K.E. Artificial Agency, Consciousness, and the Criteria for Moral Agency: What Properties must an Artificial Agent have to be a Moral Agent? Ethics and Information Technology. 2009. N. 11. P. 19–29.

18. Lara F. Why a Virtual Assistant for Moral Enhancement When We Could have a Socrates? Science and Engineering Ethics. 2021, N 42. P. 27.

19. Moral Dilemmas and Moral Theory / ed. Mason H.E. N.Y.: Oxford University Press, 1996.

20. Nadeau J.E. Only Androids Can Be Ethical // Thinking about Android Epistemology / eds. Ford K. and Glymour C. MIT Press, 2006. P. 241–248.

21. Savulescu J., Persson I. Moral Enhancement, Freedom and the God Machine. The Monist. 2012. N 95(3). P. 399–421.

22. Sullins J.P. When Is a Robot a Moral Agent. International Review of Information Ethics. 2006. Vol.6 (12). P. 23–30.

23. Thomson J.J. The Trolley Problem. The Yale Law Journal. 1985. Vol. 94, N 6. P.1395–1415.

24. Turing A. Computing Machinery and Intelligence // Mind. 1950, LIX (236). P. 433–460.

Система Orphus

Loading...
Up