% IMPORTANT: The following is UTF-8 encoded.  This means that in the presence
% of non-ASCII characters, it will not work with BibTeX 0.99 or older.
% Instead, you should use an up-to-date BibTeX implementation like “bibtex8” or
% “biber”.

@INPROCEEDINGS{Heinrichs:1025942,
      author       = {Heinrichs, Jan-Hendrik},
      title        = {{AMA}s, function creep and moral overburdening},
      reportid     = {FZJ-2024-03220},
      year         = {2023},
      abstract     = {AMAs, function creep and moral overburdeningDesigning
                      artificially intelligent systems to function as moral agents
                      has been suggested as either a necessary or at least the
                      most effective way to counter the risks inherent in
                      intelligent automation [2]. This suggestion to generate
                      so-called Artificial Moral Agents (AMAs) has received ample
                      support [see the contributions in 7] as well as critique
                      [9]. What has yet to be sufficiently taken into account in
                      the discussion about AMAs is their effect on the moral
                      discourse itself. Some important considerations in this
                      regard have been presented by Shannon Vallor [8] and by
                      Sherry Turkle [3] under the heading of moral deskilling. The
                      core claim of the moral deskilling literature is that the
                      employment of artificially intelligent systems can change
                      spheres of activity which used to require moral skills in
                      such a way that these skills lose their relevance and thus
                      cease to be practiced.This contribution will argue that
                      deskilling is just one among several changes in the moral
                      landscape that might accompany the development and
                      dissemination of AMAs. It will be argued that there are two
                      complementary trends which AMAs might initiate, both of
                      which might fall under the heading of functions creep as
                      defined by Koops: „Based on this, function creep can be
                      defined as an imperceptibly transformative and therewith
                      contestable change in a data-processing system’s proper
                      activity.” [4, 10]. These developments are a) the
                      moralization of spheres of action previously under little or
                      no moral constraints and b) changing – maybe rising –
                      and increasingly complex standards of moral conduct across
                      spheres of action.a) The former trend – moralization of
                      additional spheres of action – occurs when in the process
                      of (partially) automating a certain task moral precautions
                      are implemented, which have not been a part of previous
                      considerations. A common example is the explicit, prior
                      weighing of lives which automated driving systems are
                      supposed to implement, but which typically do not play a
                      role in a real driver’s education, much less their
                      reaction during an accident. Automatization – be it of
                      cars, government services or any other sphere of activity
                      – typically requires actively revising or maintaining the
                      structures of a given process and therefore generates the
                      option to include moral considerations where there
                      previously were none or few. The availability of established
                      AMA-techniques is likely to influence stakeholders to
                      include such moral considerations, whether for product
                      safety reasons or for mere commercial advantage.b) The
                      latter trend – the change of standards of moral conduct
                      – is an effect of the requirements of intelligent
                      automatization on the one hand and of professionalism on the
                      other. Compared to humans, AMAs and their algorithmic
                      processes employ different processes of behavioural control
                      in general and of observing moral constraints in particular
                      [6]. Thus, even if automation of tasks tries to mimic
                      pre-established moral behaviour and rules there will be
                      differences in the representation, interdependence, and
                      acceptable ambiguity of moral categories in the human
                      original and the AMA implementation. Furthermore, the
                      implementation of moral constraints in automated systems
                      already is a professional field which tries to live up to
                      extremely high standards – as exemplified by the
                      increasingly complex approaches to algorithmic fairness
                      [11]. Accordingly, it is to be expected that the field will
                      aim to implement high moral standards in their products,
                      sometimes beyond what would be expected of human agents in
                      the same sphere of activity.While both trends seem to be
                      positive developments at first hand, there is relevant risk
                      that changing and complex moral standards in more and more
                      spheres of action overburden an increasing number of people,
                      making them increasingly dependent on external guidance or
                      setting them up for moral failure and its consequences. This
                      development combines two effects which have previously been
                      identified in literature. On the one hand, it seems to be a
                      case of what Frischmann and Selinger called ‘techno-social
                      engineering creep’ [4], that is the detrimental change of
                      our collective social infrastructure and standards through
                      individual, seemingly rational choices. By individually
                      implementing AMA-algorithms in appliances in several spheres
                      of activity for reasons of safety (or market share), we
                      change the moral landscape throughout. On the other hand, it
                      is similar to what Daniel Winkler has identified as one of
                      the threats of human enhancement, namely that standards
                      suited for highly functioning individuals will overburden
                      the rest of the community[5]. The present trends differ from
                      Winkler’s analysis insofar as it might affect all human
                      agents and not just those who forgo some form of cognitive
                      improvement.Clearly, the proliferation and change of moral
                      standards does not merely carry risks. It has the potential
                      to generate relevant social benefits, which need to be
                      considered in a balanced account. However, a purely
                      consequentialist analysis of these trends would lack regard
                      for the important dimension of procedural justification.
                      While changes in the moral landscape often occur without
                      explicit collective attention or explicit deliberation,
                      there seem to be minimal conditions of the legitimacy, such
                      as not providing any affected party with good reasons to
                      withhold their consent [1]. The current contribution will
                      sketch the consequentialist and procedural constraints on
                      the trends identified above and try to spot a path between
                      relinquishing the design of AMAs and alienating human agents
                      from moral discourse.References:1. Scanlon, T., What we owe
                      to each other. 1998, Cambridge, Mass.: Belknap Press of
                      Harvard University Press. ix, 420 p.2. Wallach, W. and C.
                      Allen, Moral machines. Teaching robots right from wrong.
                      2009, Oxford ; New York: Oxford University Press. xi, 275
                      p.3. Turkle, S., Alone Together: Why We Expect More from
                      Technology and Less from Each Other. 2011, New York: Basic
                      Books.4. Frischmann, B. and E. Selinger, Re-Engineering
                      Humanity 2018, Cambridge: Cambridge University Press.5.
                      Wikler, D., Paternalism in the Age of Cognitive Enhancement:
                      Do Civil Liberties Presuppose Roughly Equal Mental Ability?,
                      in Human Enhancement, J. Savulescu and N. Bostrom, Editors.
                      2010, Oxford University Press: Oxford / New York.6. Liu, Y.,
                      et al., Artificial Moral Advisors: A New Perspective from
                      Moral Psychology, in Proceedings of the 2022 AAAI/ACM
                      Conference on AI, Ethics, and Society. 2022, Association for
                      Computing Machinery: Oxford, United Kingdom. p. 436–445.7.
                      Anderson, M. and S.L. Anderson, eds. Machine Ethics. 2011,
                      Cambridge University Press: Cambridge.8. Vallor, S., Moral
                      Deskilling and Upskilling in a New Machine Age: Reflections
                      on the Ambiguous Future of Character. Philosophy $\&$
                      Technology, 2015. 28(1): p. 107-124.9. van Wynsberghe, A.
                      and S. Robbins, Critiquing the Reasons for Making Artificial
                      Moral Agents. Science and Engineering Ethics, 2019. 25(3):
                      p. 719-735.10. Koops, B.-J., The Concept of Function Creep.
                      Management of Innovation eJournal, 2020.11. Fazelpour, S.
                      and D. Danks, Algorithmic bias: Senses, sources, solutions.
                      Philosophy Compass, 2021. 16(8): p. e12760.},
      month         = {Dec},
      date          = {2023-12-15},
      organization  = {5th Conference on "Philosophy of
                       Artificial Intelligence" PhAI 2023,
                       Erlangen (Germany), 15 Dec 2023 - 16
                       Dec 2023},
      subtyp        = {Invited},
      cin          = {INM-7},
      cid          = {I:(DE-Juel1)INM-7-20090406},
      pnm          = {5255 - Neuroethics and Ethics of Information (POF4-525)},
      pid          = {G:(DE-HGF)POF4-5255},
      typ          = {PUB:(DE-HGF)6},
      url          = {https://juser.fz-juelich.de/record/1025942},
}