AI-Based Decisions and Disappearance of Law

Vol.16,No.2(2022)

Abstract

Based on the philosophical anthropology of Paul Ricoeur, the article examines, using the example of AI-based decisions, how the concept of responsibility changes under the influence of artificial intelligence, what a reverse effect this conceptual shift has on our moral experience in general, and what consequences it has for law. The problem of AI-based decisions is said to illustrate the general trend of transformation of the concept of responsibility, which consists in replacing personal responsibility with a system of collective insurance against risks and disappearing of the capacity for responsibility from the structure of our experience, which, in turn, makes justice and law impossible.


Keywords:
Artificial Intelligence; AI-based Decisions; Responsibility; Experience of Law; Paul Ricoeur

Pages:
241 – 267
Author biographies

Yulia Razmetaeva

Yaroslav Mudryi National Law University

Department of Theory and Philosophy of Law, Associate Professor

Natalia Satokhina

Yaroslav Mudryi National Law University

Department of Theory and Philosophy of Law, Associate Professor
References

[1] Angwin, J. et al. (2016) Machine Bias. There’s Software Used Across the Country to Predict Future Criminals. And it’s Biased Against Blacks. ProPublica, 23th May. Available from: https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing [Accessed 24 October 2020].

[2] Arendt, H. (1998/1958) The Human Condition. Chicago and London: The University of Chicago Press.

[3] Bărbulescu v. Romania, 61496/08, [2016] ECHR 61, [2017] ECHR 742, [2017] ECHR 754.

[4] Bleske-Rechek, A. et al. (2010) Evolution and the Trolley Problem: People Save Five Over One unless the One is Young, Genetically Related, or a Romantic Partner. Journal of Social, Evolutionary, and Cultural Psychology, 4 (3), pp. 115-127.

[5] Cofone, I. (2019) Algorithmic Discrimination is an Information Problem. Hastings Law Journal, 70, pp. 1389-1444.

[6] Da Silva, S., Matsushita, R. and De Sousa, M. (2016) Utilitarian Moral Judgments are Cognitively Too Demanding. Open Access Library Journal, 3 (2). Available from: http://dx.doi.org/10.4236/oalib.1102380.

[7] Groher, W., Rademacher, F.-W. and Csillaghy, A. (2019) Leveraging AI-based Decision Support for Opportunity Analysis. Technology Innovation Management Review, 9 (12), pp. 29-35. Available from: http://doi.org/10.22215/timreview/1289.

[8] Floridi, L. and Sanders, J. (2004) On the Morality of Artificial Agents. Minds and Machines, 14, pp. 349-379. Available from: https://doi.org/10.1023/B:MIND.0000035461.63578.9d.

[9] Foot, P. (1967) The Problem of Abortion and the Doctrine of Double Effect. Oxford Review, 5, pp. 5-15.

[10] Habermas, J. (2003) The Future of Human Nature. Cambridge: Polity Press.

[11] Kant, I. (2003) The Critique of Pure Reason. Available from: https://www.gutenberg.org/files/4280/4280-h/4280-h.htm [Accessed 23 October 2020].

[12] Klenk, M. (2020) Digital Well-being and Manipulation Online. In C. Burr and L. Floridi (eds.) Ethics of Digital Well-Being: A Multidisciplinary Approach. Dordrecht: Springer, pp. 81-100.

[13] Mazur, J. (2019) Automated Decision-making and the Precautionary Principle in EU Law. Baltic Journal of European Studies, 9 (4), pp. 3-18. Available from: https://doi.org/10.1515/bjes-2019-0035.

[14] Moor, J. (2006) The Nature, Importance, and Difficulty of Machine Ethics. IEEE Intelligent Systems, 21 (4), pp. 18-21.

[15] Phillips-Wren, G. and Jain, L. (2006) Artificial Intelligence for Decision Making. In B. Gabrys, R.J. Howlett and L.C. Jain (eds.) Knowledge-Based Intelligent Information and Engineering Systems. KES 2006. Lecture Notes in Computer Science, vol. 4252. Berlin and Heidelberg: Springer, pp. 531-536. Available from: https://doi.org/10.1007/11893004_69.

[16] Ricoeur, P. (1994) Oneself as Another. Chicago and London: The University of Chicago Press.

[17] Ricoeur, P. (2000) The Just. Chicago and London: The University of Chicago Press.

[18] Sharkey, A. (2017) Can We Program or Train Robots to Be Good? Ethics and Information Technology. Available from: https://doi.org/10.1007/s10676-017-9425-5.

[19] Susser, D., Roessler, B. and Nissenbaum, H. (2019) Technology, Autonomy, and Manipulation. Internet Policy Review, 8 (2). Available from: https://doi.org/10.14763/2019.2.1410.

[20] Thomson, J. (1985) The Trolley Problem. Yale Law Journal, 94, pp. 1395-1415. Available from: http://dx.doi.org/10.2307/796133.

[21] Vo v. France, 53924/00, [2004] ECHR 326, (2005) 40 EHRR 12.

[22] Weinreb, Lloyd L. (2004) A Secular Theory of Natural Law. Fordham Law Review, 72 (6), pp. 2287-2300. Available from: https://ir.lawnet.fordham.edu/cgi/viewcontent.cgi?article=3990&context=flr [Accessed 23 October 2020].

[23] Williams, B., Brooks, C. and Shmargad, Y. (2018) How Algorithms Discriminate Based on Data They Lack: Challenges,Solutions, and Policy Implications. Journal of Information Policy, 8, pp. 78-115.

Metrics

602

Views

275

pdf views