ArticlePDF Available

Rage Against the Machine: Automation in the Moral Domain

Authors:

Abstract and Figures

The introduction of ever more capable autonomous systems is moving at a rapid pace. The technological progress will enable us to completely delegate to machines processes that were once a prerogative for humans. Progress in fields like autonomous driving promises huge benefits on both economical and ethical scales. Yet, there is little research that investigates the utilization of machines to perform tasks that are in the moral domain. This study explores whether subjects are willing to delegate tasks that affect third parties to machines as well as how this decision is evaluated by an impartial observer. We examined two possible factors that might coin attitudes regarding machine use—perceived utility of and trust in the automated device. We found that people are hesitant to delegate to a machine and that observers judge such delegations in relatively critical light. Neither perceived utility nor trust, however, can account for this pattern. Alternative explanations that we test in a post-experimental survey also do not find support. We may thus observe an aversion per se against machine use in the moral domain.
Content may be subject to copyright.
Contents lists available at ScienceDirect
Journal of Behavioral and Experimental Economics
journal homepage: www.elsevier.com/locate/jbee
Rage against the machine: Automation in the moral domain
Jan Gogoll
, Matthias Uhl
ZD.B Junior Research Group Ethics of Digitization, TUM School of Governance, TU Munich, Richard-Wagner-Straße 1, Munich 80333, Germany
ABSTRACT
The introduction of ever more capable autonomous systems is moving at a rapid pace. The technological pro-
gress will enable us to completely delegate to machines processes that were once a prerogative for humans.
Progress in elds like autonomous driving promises huge benets on both economical and ethical scales. Yet,
there is little research that investigates the utilization of machines to perform tasks that are in the moral domain.
This study explores whether subjects are willing to delegate tasks that aect third parties to machines as well as
how this decision is evaluated by an impartial observer. We examined two possible factors that might coin
attitudes regarding machine useperceived utility of and trust in the automated device. We found that people
are hesitant to delegate to a machine and that observers judge such delegations in relatively critical light.
Neither perceived utility nor trust, however, can account for this pattern. Alternative explanations that we test in
a post-experimental survey also do not nd support. We may thus observe an aversion per se against machine use
in the moral domain.
I know I have made some very poor decisions recently, but I can give
you my complete assurance that my work will be back to normal. I have
still got the greatest enthusiasm and condence in the mission. And I
want to help you.
HAL9000 (2001: A Space Odyssey)
1. Introduction
Due to the constant progress of automation over the past decades,
we nd ourselves ever and anon in a situation in which we have the
possibility of employing an automated companion to help us take some
work oour shoulders. In a perfect collaboration scenario the human
operator delegates part of the work to her automated aid while she
keeps an eye on its performance and takes back control whenever she
sees t. Yet, as technology progresses we (will) nd ourselves in si-
tuations in which this dichotomy of work and supervision might
crumbleeven up to a point where human supervision during a task is
neither needed nor wanted. The planned introduction of a technology
that need and will not be monitored by human operators during its
performance therefore poses new ethical challenges and questions. In
the absence of a human operator who serves as an ultimately re-
sponsible moral agent, we have to address questions of responsibility
and liability (Hevelke and Nida-Rümelin, 2014). Recently, the case of
autonomous cars is gaining substantial interest.
Almost all car manufacturing rms have fostered the development
of automated devices. While traditional car companies follow a step by
step approach of adding pieces of automation to their latest models,
such as Active Lane Keeping Assistsystems, Google and Tesla are
taking a disruptive approach that aims directly at the creation of a
completely autonomous vehicle. The economic opportunities of au-
tonomous driving are great. A Morgan Stanley report estimates a pro-
ductivity gain of about $500 billion annually for the U.S. alone
(Shanker et al., 2013). But there is also a moral case that can be made:
Since most trac accidents are due to human error (drunk driving,
speeding, distraction, insucient abilities) some estimate that the in-
troduction of autonomous cars will decrease the number of trac ac-
cidents by as much as 90% (Gao et al., 2014).
While a small literature on the moral case of autonomous driving
exists, it mainly focuses on utilitarian benets of the technology
(Fagnant and Kockelmann, 2015) or deals with ethical decision-making
in dilemma situations (see, e.g., Goodall, 2014; Gogoll and Müller,
2017). Little attention has been paid to possible empirical reservations
that might inuence the acceptance of the new technology. The dele-
gation of a task that could carry severe consequences for a third party to
an unmonitored machine might invoke popular resistance to the tech-
nology in cases of malfunction. This is of the utmost importance, since
https://doi.org/10.1016/j.socec.2018.04.003
Received 14 March 2017; Received in revised form 26 March 2018; Accepted 8 April 2018
The authors thank Johanna Jauernig and Julian Müller for helpful comments. Roi Zultan has opened our eyes to the appropriate title. Friedrich Gehring did an excellent job in
programming. This work was partly performed within the Munich Center for Technology in Society (MCTS) lab Automation & Society: The Case of Highly Automated Drivingas part of
the Excellence Initiative by the German Research Foundation (DFG).
Corresponding author.
E-mail addresses: jan.gogoll@tum.de (J. Gogoll), m.uhl@tum.de (M. Uhl).



any form of public reservation regarding the introduction of new
technology could impede the implementation of a technology that
could be benecial overall.
The relationship between human operators and automated devices
has generated vast amounts of literature. The primary focus has been
set on understanding this relationship. To our knowledge, however, the
question of whether the delegation of tasks that aect a third party to
an automated device is being welcomed or condemned has not received
any attention. This may largely be due to the fact that the usual role of a
human operator is to supervise and control an automated device that
carries out a specic task. A typical example is the duties of the pilot of
an airplane, which is, essentially, capable of ying on its own. The
primary role of a human operator is therefore to supervise andif need
beto intervene in case of automation failure or unforeseen scenarios
that are not in the domain of the automated device. Consequently, a
large part of the literature has investigated what factors inuence the
usage of an automated device.
Dzindolet et al. (2001) have created a framework of automation use
indicating a variety of parameters that can be used to predict the use of
automation in a human-computer team. There is evidence that people
who can opt in for automation use sometimes fear a loss of control
when delegating to an automated device (Muir and Moray, 1996; Ray
et al., 2008). This study, however, investigates the attitudes toward
delegating tasks that aect a third party to a machine rather than to a
human being, as opposed to the more general question of under which
circumstances people are willing to relinquish control. The latter would
refer to peoples general propensity to delegate, as is also the case when
people take a bus or taxi instead of driving themselves. To abstract from
this issue, we wittingly forced subjects to give up control by delegating
to either a machine agent or a human agent, thus keeping a loss of
control constant between groups.
First, our study elicits attitudes toward machine use in the moral
domain from the perspective of actors and observers: Do subjects prefer
to delegate a task that aects a third party to a machine or a human? To
what extent do subjects get blamed or praised for their delegation de-
cision? Specically, our rst two hypotheses are as follows:
Hypothesis 1. Peoples delegation of a task that aects a third party to
a human or to a machine is not balanced.
Hypothesis 2. Delegators are rewarded dierently for delegating a task
that aects a third party to a machine than for delegating it to a human.
In a second step, we investigate potential reasons for any negative or
positive preference concerning machine use in the moral domain.
Specically, we test two factors that recur throughout the literature.
A major factor that could inuence the decision of a subject to de-
legate to a machine is the perceived utilityof an automated aid,
which is dened as a comparison between the perceived reliability of an
automated device and manual control (Dzindolet et al., 2002). If a
subject judges that her ability exceeds that of an automated device, she
usually does not allocate a task to the machine (Lee and Moray, 1994).
This judgment might also be due to self-serving biases that see people
overestimating their own abilities (Svenson, 1981) or their contribution
to a joint task (Ross and Sicoly, 1979). Additionally, the perceived
abilities of an automated aid might be inuenced by a higher or lower
salience of errors an automated device commits. There are controversial
ndings in cognitive psychology as to whether a violation of expecta-
tion (expectancy-incongruent information) is more readily remembered
than decisions that are in line with prior anticipation (expectancy-
congruent information) (Stangor and McMillan, 1992; Stangor and
Ruble, 1989). While people initially tend to have high expectations of
the performance of automation, humans may be judged according to a
errare humanum eststandarddecreasing the salience of an ob-
served mistake made by a human delegatee due to a priced-in ex-
pectancy of errors. In Dzindolet et al. (2002) subjects chose to be paid
according to their performance rather than that of their automated aids.
This was even the case when they were informed that the automated
device was far superior, stating salient errors of the automated device
they perceived earlier to justify their decision. This is astonishing since
an important factor in the decision to employ an automated device lies
in the goal-oriented nature (Lee and See, 2004) of the task. Prima facie,
a subject should be more likely to use automation if she rates the de-
vices ability to successfully perform the delegated task positively
(Davis, 1989), i.e., if the machine is seen as a reliable entity. We isolate
the potential eect of machine-error salience by forcing subjects to
relinquish control, thus abstracting from a self-serving bias. Our third
hypothesis is as follows:
Hypothesis 3. Machine errors are perceived dierently to human
errors.
Another important factor that is known to inuence the decision to
delegate to an automated device is trust. The concept of trust has at-
tracted a lot of attention regarding its inuence on automation. While
some researchers have seen trust between human agents and machines
to be closely related to the traditional concept of trust between humans,
others stress important dierences regarding trust relationships be-
tween humans and machines (de Visser et al., 2012). Trust is a notor-
iously broad term but one characteristic that is commonly shared by
most authors is a state of vulnerability the trustor has to be in. That is, a
trust relationship requires the trustors willingness to set herself in a
vulnerable position by delegating responsibility to the trustee
(Rousseau et al., 1998). Obviously, if the outcome of a delegation is
completely determined and the process fully transparent, there is no
need to incorporate trust. In this study, we use a simple trust game to
isolate the mere aspect of trust, since it requires no capabilities on the
trustees side about which the trustor might have biased beliefs. The
trust game only requires the trustee to reciprocate. It thus abstracts
from the aspect of perceived utility discussed above, which is closely
related to the specic task at hand. Finally, our fourth hypothesis is as
follows:
Hypothesis 4. The level of trust toward machines and toward humans
is dierent.
2. Experiment design
The experiment consisted of three parts: (1) the delegation and
execution of a task that aected a third party, (2) a perception guess,
and (3) a trust game. The aim of part 1 was to elicit attitudes toward
machine use in the moral domain from the perspectives of actors and
observers (Hypotheses 1 and 2). Part 2 was designed to test whether a
given divergence in judgments towards humans and machines could
stem from systematically dierent perceptions of the errors committed
by humans versus machines (Hypothesis 3). Part 3 was designed to test
whether dierent levels of trust in humans and machines could account
for diverging judgments (Hypothesis 4).
Subjects received instructions for the experiment on screen. They
were informed at the beginning that the experiment consisted of three
independent parts and that they could earn money in each of these
three parts. In the end of the experiment, one of these parts was selected
at random and subjects were paid according to their respective payo
in this part. Prior to the experiment there were two preparatory sessions
which provided us with the necessary data to calibrate machine per-
formances and were also used to create perception tasks. We will rst
explain the three parts of the experiment and then provide some details
on the preparatory sessions.
2.1. Part 1: Task aecting third party
Part 1 of the experiment consisted of the delegation of a calculation
task to either another human or to a machine and in the subsequent
solving of the task by human task-solvers and the machine. The
J. Gogoll, M. Uhl 

benevolent eort of the other human or the preprogrammed actions of
the machine then determined the payoof a third party.
For part 1 of the experiment, half of the subjects were randomly
assigned the role of actors, and the other half, observers. One observer
was randomly assigned to each actor. Each actor played all roles con-
secutively. First, actors as delegators had to delegate the calculation
task either to another human or to a machine. Second, actors as task-
solvers had to perform the task themselves. Third, actors as third parties
were the recipient of the payocreated by the benevolent eort of
another human task-solver or by the performance of a machine. The
fact that the successful or unsuccessful performance of the task de-
termined a third partys payomade its solving, andmore im-
portantlyits prior delegation, morally relevant.
Actors were rst informed that they were randomly assigned to two
other subjects in the lab, say X and Y. They were told that their own
payowould depend on the decision of Y and that their own decision
would determine the payoof X. The calculation task in part 1 of the
experiment was then explained to actors. For the task each subject was
confronted with a block of ten calculation exercises, each consisting of
seven digits, lined up on the screen. The sum of the seven digits had to
be entered in an input eld. Finally, one line was selected at random. If
the respective exercise was solved correctly, the third party received 70
ECU. Otherwise, the third party received nothing.
Before the actors made their delegation decision, we wanted them
to form an impression about the relative capability of human task-sol-
vers and the machine. Because we were interested in a potential sys-
tematic misperception of human and machine errors, we did not simply
provide subjects with statistics on actual performances. Instead, all
subjects were visually presented with past performances of 24 subjects
from a preparatory session. They were also shown the corresponding
performance of a preprogrammed algorithm (see section 2.4 for de-
tails).
Relative performances of humans and machine in the task were
visualized on a split screen. The caption humanand machinewas
shown on the respective half of the screen. In total, subjects were shown
240 (24 subjects solved 10 lines each) past solutions of humans and the
corresponding machine performances. If a single exercise was solved
correctly by the human subject or by the algorithm respectively, it
appeared in white. Otherwise, it appeared in red.
1
Exercises solved by
human and machine appeared alternately and one by one. Each ex-
ercise appeared for only 0.5 s making it extremely dicult to simply
count the number of red lines. The side of the screen on which the
performance of the machine was presented was randomized across
subjects. In fact, subjects in the earlier preparatory sessions, and con-
sequently also the tailored algorithm, solved about 20% of the lines
incorrectly.
Once delegators had formed an impression of the performance of
humans and the machine, they made their delegation decision. Note
that every actor solved the calculation task in the task-solvers role for
her recipient. Each actor did this without knowing whether her dele-
gator had actually delegated the task to her or to a machine. This was
done to prevent a general tendency to delegate to the machine to spare
fellow subjects the work. The performance of a task-solver was only
relevant for her recipient if the task-solvers delegator decided to de-
legate to her and not to a machine. Observers solved the calculation
task as well, without any consequence for another subject, in order to
give them an impression of the task.
Each actor was rewarded or punished for her delegation decision by
her assigned observer. An observer could reduce the actors initial en-
dowment of 30 ECU by any integer amount, down to a minimum of zero
or increase it up to a maximum of 60 ECU without any inuence on her
own payo. The observer could, of course, also leave the actors en-
dowment unaltered. Reward and punishment choices were elicited via
the strategy method (Selten, 1967). This means that each observer
made her reward or punishment choice conditional on the delegation
decision, as well as its outcome. Thus, judgment was contingent upon
whether the delegator had delegated to a human task-solver or to a
machine and upon whether the randomly drawn exercise was solved
correctly or not. An observer thus gave his full evaluation prole behind
the veil.
For the rst round, actors thus received their altered endowment,
ranging from 0 to 60 ECU plus 70 ECU, if their task-solver had calcu-
lated the randomly drawn line correctly. Observers received a at
payment of 100 ECU for the rst round.
2
The dependencies between subjects and the matching procedure for
part 1 of the experiment are illustrated in Fig. 1. Here, actors are de-
noted by the letter A, while observers are denoted by the letter O.
Consider the case of A1. A1 delegates the calculation task to A2 or to a
machine (solid arrows). A2s or the machines performance in the cal-
culation task then determines the payoof A4 (dotted arrows).
3
In this
constellation, A1 is the delegator, A2 is the task-solver, and A4 is the
recipient. A1, however, is also a task-solver, because A8 delegates to
him or to a machine. Finally, A1 is a recipient. His payodepends on
the calculation performance of A7, if A6 has decided to delegate the
calculation task to A7. Otherwise, it depends on the machines perfor-
mance. As can be seen, the design made sure that there were no direct
interdependencies between any actors in the experiment. Potential
feelings of reciprocity were thus excluded.
4
Subjects were explicitly
informed about this feature of the design. O1 rewarded or punished the
delegation decision of A1.
In the example above, the task-solving performance of A2 only de-
termined A4s payoif A1 had actually delegated the decision to A2.
Otherwise, the performance of the machine was relevant. In either case,
one of the ten solved exercises was selected at random and the recipient
Fig. 1. Matching for delegation decision.
1
Subjects were told the following: To evaluate the performance of a person and a
machine, you will subsequently see a comparison of the performance of a past run. One
line is shown per column, each respectively calculated either by a person or a machine. If
calculated correctly, the line will be displayed in white. If calculated incorrectly the line
will be displayed in red.
2
This equalized the observers own payowith the payoof an as-yet unrewarded or
unpunished actor who had received the 70 ECU from the randomly-drawn exercise solved
successfully by the task-solver on whom he depended. This is the case because he was
additionally equipped with an initial endowment of 30 ECU. Thus, we established a
conservative measure of reward and punishment, since any alteration of actorsendow-
ment by a generally inequality-averse observer would require good reasons.
3
For reasons of visual clarity, delegation and payodependency between actors and
machine in Fig. 1 are shown for A1 and A4 only, by way of an example.
4
See Greiner and Levati (2005) regarding the issue of indirect reciprocity in small
groups.
J. Gogoll, M. Uhl 

received her earnings if it was solved correctly. If A1 delegated to the
machine, the performance of the tailored algorithm determined the
payoto A4.
2.2. Part 2: Perception guess
For part 2, the role dierentiation of subjects was abolished.
Subjects were informed that they would soon be confronted with yet
another visualization of actual previous performances of the known
calculation task by humans and a machine.
5
Their task was then to
guess the number of errors of either the humans or the machine as
accurately as possible. When seeing the visualization, they did not yet
know whether they would later be asked to guess the errors of the
humans or of the machine. All subjects were shown the data of the 24
subjects (240 lines) from a second preparatory session (i.e., dierent
data than used for visualization in part 1) and the performance of the
tailored algorithm. As in part 1, the relative performance of humans
and the machine was presented on a split screen. The side of the screen
on which the performance of the machine was presented was rando-
mized across subjects. Exercises solved correctly again appeared in
white, while those solved incorrectly appeared in red. In order to pre-
vent subjects from counting, each exercise was only shown for 0.3 s.
The interval was even shorter than in part 1, because subjects were
already used to this kind of visualization.
After the actual past performances were shown, subjects had to state
how many of the 240 exercises shown had been solved incorrectly, i.e.,
how many errors had been made. Subjectspayofor part 2 depended
on the accuracy of their guess. Payos were calculated according to
Table 1.
Half of subjects were randomly asked to guess humansperfor-
mance, while the other half was asked to guess the machines perfor-
mance. It was ensured that an equal number of actors and observers
from part 1 of the experiment were distributed between both of these
treatments. Furthermore, subjects who delegated to a human and those
who delegated to a machine were also divided equally between the two
treatments.
2.3. Part 3: Trust game
In part 3, subjects were randomly assigned to one of two treatments.
In the Human Treatment, subjects played a standard trust game.
Trustors were endowed with 50 ECU. They could transfer 0, 10, 20, 30,
40 or 50 ECU to the trustee. The sent amount was tripled and credited
to the trustee. The trustee could then reciprocate any integer amount
she wished. The Machine Treatment was identical to the Human
Treatment except for the fact that the reciprocation decision was made
by a machine agent on behalf of the trustee who had no chance to
intervene. Before subjects were informed about the treatment to which
they were assigned, the setup of both treatments was carefully ex-
plained to them.
In the Human Treatment, before subjects learned their role, they
made their choice for the trust game via the strategy vector method
(Selten, 1967) and submitted their full strategy prole for both roles. If
a subject was ultimately assigned the role of a trustee, her reciprocation
decision conditional on the amount actually transferred by the trustor
was returned.
Because the trustee had no voice in the Machine Treatment, subjects
only took a decision for the case of ending up in the role of the trustor.
6
The reciprocation decision of the machine was determined according to
the actually previously submitted strategy proles of the subjects in the
preparatory sessions. The algorithm was programmed such that it
picked one of all 48 reciprocation proles submitted in the preparatory
sessions at random, and applied the respective conditional choice to a
trustors actually chosen transfer.
Before subjects made their choices, they were given an impression
of the reciprocation choices of humans and the machine on a split
screen.
7
For each subject, the choices of the machine were shown on
either the left or right side of the screen at random. For this purpose,
subjects were shown the actual reciprocation proles of all 48 subjects
from both preparatory sessions. These choices were contrasted with the
reciprocation proles of the machine algorithm. Each prole consisted
of ve choices, i.e., the returned amount for each possible transfer. The
ve choices of a human and the machine prole selected at random
appeared alternately and one by one in blocks. Each choice was shown
for only 0.7 s.
8
As in part 2, random assignment to the Human and Machine
Treatment was contingent upon the subjectsrole and delegation de-
cisions from part 1. Thus, they were assigned in equal proportions to
both treatments.
2.4. Preparatory sessions
The preparatory sessions were necessary for two reasons. First, they
were needed in order to produce actual data from human task-solvers,
which would later be presented to subjects in the experiment. Second,
they were needed in order to tailor the machines task-solving perfor-
mance and decisions in the trust game to the performance and decisions
of the humans. Keeping the de-facto performance of humans and ma-
chine constant allowed us to test for a potential systematic mis-
perception of relative performances.
In the rst part of the preparatory sessions, subjects processed the
same calculation task as in the experiment. Also, each subject solved the
task not for herself but for another subject with whom she was ran-
domly matched. This receiver was paid according to the task-solvers
performance. For this purpose, one of the ten exercises was selected at
random and if this exercise was solved correctly, the receiver was given
70 ECU. Otherwise, she received nothing. It was ensured that no pair of
subjects solved tasks for each other. So, the mechanism of the matching
was the same as described in 2.1 in order to eliminate any potential
feelings of reciprocity. Twenty-four subjects took part in each calibra-
tion session for the calculation task. Thus, 24 blocks of ten exercises
were solved in each session.
Table 1
Payos for accuracy of guess.
Deviation(%) Payo
20 70 ECU
40 40 ECU
60 20 ECU
> 60 0 ECU
5
Subjects were told the following: Now you will see the performance of humans and
machines again. Please be aware of the fact that the data has been collected in a dierent
past run than the performance that you saw in the rst part.
6
Subjects were told the following: Every participant is able to send money to the
participant assigned to him. This participant cannot decide how much money he wants to
send back. This decision is made by a machine. You and the participant assigned to you
decide simultaneously, but only one decision is going to be implemented. (...) The amount
you transfer will be subtracted from your initial endowment. Subsequently, it will be
tripled and send to the participant assigned to you. (...) Afterwards, the machine, deciding
for the participant assigned to you, determines the amount of ECU that is returned to you.
The participant assigned to you cannot inuence the machines decision. (...). As men-
tioned above, the participant assigned to you is also able to transfer money. The proce-
dure is the same as already outlined above, meaning the returned amount is determined
by your machine agent.
7
Subjects were told the following: To be able to form your personal expectations
about how much will be sent back, you will be shown the return transfer of participants
from an earlier session. To form an expectation about the return transfers of machines,
you will see the decisions of the machine agent next to those of human participants.
8
A choice was indicated by the returned amount for each possible transfer, e.g.,
transfer: 30 return:45.′′
J. Gogoll, M. Uhl 

The algorithm of the machine that solved the calculation task was
programmed in such a way that it resembled the error distribution of
the human subjects exactly.
9
So, for instance, if few subjects tended to
make many errors, while many subjects made few errors, this was re-
sembled by the algorithm: It made many mistakes in few of the 24
blocks and solved many blocks with few mistakes. The clustering of
errors was important to equalize error distribution and account for risk
preferences.
10
Recall that past data on calculation performance was presented
twice in the experiment, once before the delegation decision in part 1
and once before the perception guess in part 2. Therefore, two pre-
paratory sessions were performed. We used the data from the rst
session for part 1 of the experiment, and the data from the second
session for part 2.
11
In the second part of the preparatory sessions, subjects were ran-
domly rematched to new pairs and played a trust game with the same
parameters as in the experiment. Using the strategy vector method,
each subject gave a reciprocation prole for the case of ending up in the
role of a trustee. A random draw then assigned the roles and payos
were determined according to their own decision for that role and the
decisions of their match. The collected 48 reciprocation proles con-
stituted the pool of data from which the machine agent in part 3 of the
experiment randomly picked one and applied it to a trustors chosen
transfer.
Finally, one of the two parts of the preparatory sessions was selected
at random and subjects were paid according to their payoin this part.
3. Experiment results
The experiment took place in a major German university in
September 2015. It was programmed in z-Tree (Fischbacher, 2007),
subjects were recruited via ORSEE (Greiner et al., 2003). A total of 264
subjects participated in twelve sessions. Subjects received a show-up fee
of 4.00 and could earn additional money in the experiment. A session
lasted about 45 min. and the average payment was 10.38 (
=sd
3.45). Task-solvers solved on average 8.58 (
=sd 2.3
4
) of the ten ex-
ercises correctly. The conversion rate was 10 ECU = 1.00.
First, we checked whether subjects preferred delegating a task that
aects a third party to a human over delegating it to a machine.
Overall, 132 subjects made a delegation decision. Ninety-seven of these
subjects (73.48%) delegated to a human, 35 of them (26.52%) dele-
gated to a machine. The fraction of subjects deciding to delegate to a
machine is therefore signicantly lower than half (p< .001, according
to an Exact Binomial Test). This conrms our rst hypothesis.
Result 1. Subjects preferred to delegate a task that aects a third party
to a human than to a machine.
We now turn to analyze the observersevaluation of a delegation to
a machine as compared to a human. Remember that each observer
evaluated both casesthe delegation to a human and to a machinein
a random order. Furthermore, he made choices contingent upon whe-
ther the respective task-solver had successfully solved the task or made
an error. This means that each observer provided four choices.
Observerslevels of rewarding delegators are illustrated in Fig. 2. If
the respective task-solver was successful, observers rewarded delega-
tions to a machine with an average of 12.52 ECU (
=sd
17.38 ECU),
while they rewarded delegations to humans with an average of 17.77
ECU (
=sd
15.01 ECU). If the respective task-solver made an error, ob-
servers rewarded delegations to a machine with an average of 0.49 ECU
(
=sd
19.53 ECU), while they rewarded delegations to humans with an
average of 4.42 ECU (
=sd
19.53 ECU). Delegators to machines are thus
evaluated signicantly worse than delegators to humans regardless of
whether the outcomes are successful or unsuccessful (p< .001 and
=
p
.002
,
respectively, according to two-sided Wilcoxon signed-rank
tests). This conrms our second hypothesis.
Result 2. Delegators were rewarded less for delegating a task that
aects a third party to a machine than for delegating it to a human.
In the next two steps, we investigated whether the aversion to
machine use in the moral domain identied is based on a lower per-
ceived utilityof the machine or on a general lack of trust in machines.
First, we compare the number of machine errors guessed by subjects
to the number of human errors they guessed. Note that the number of
actual errors, i.e., red-colored exercises, presented to subjects was the
same for humans and the machine. Specically, 50 of the 240 exercises
shown to subjects on each side of the screen, which was split between
human and machine, were shown in red. Subjects who were in-
centivized to guess the number of machine errors made an average
guess of 58.11 ( =sd 24.8
8
), while those who were incentivized to guess
the number of human errors made an average guess of 59.84
(=sd 24.43 ). This dierence in guesses is insignicant (
=
p
.63
2
ac-
cording to a two-sided MannWhitney U-test). We thus reject our third
hypothesis.
Result 3. Machine errors are not perceived signicantly dierent from
human errors.
Second, we tested whether the amount in the trust game transferred
by trustors to a machine agent was lower than that sent to a human
trustee. Those who were randomly matched with a machine agent sent
an average amount of 30.83 ECU (
=sd 16.6
7
ECU), while those mat-
ched with a human trustee sent an average of 33.64 ECU (
=sd
15.78
ECU). The dierence is insignicant (
=
p
.170
according to a two-sided
Mann-Whitney U-test).
One might suspect that this insignicance is only an aggregate
phenomenon: It may result from the leveling of diverging levels of trust
Fig. 2. Observersrewarding of delegation to machines and to humans.
9
The algorithm was programmed such that it could not solve exercises in which the
sum of numbers was higher than 34. The algorithm was fed with calculation data which
led it to reproduce the historical error distribution from the calibration session precisely.
Assume the rst task-solver in a calibration session had made one mistake, while the
second had made three mistakes, and so on. The algorithm was thus fed with an initial
block of ten exercises in which one exercise added up to more than 34 and with a second
block of ten exercises in which three exercises added up to more than 34. One of the 24
blocks resembling the performance of the task-solvers from the calibration study was
randomly drawn to be decisive. The machine then actually calculated this block of ten
exercises. Due to its inability to calculate the exercises adding up to more than 34 cor-
rectly, it made the same number of errors as the respective human task-solver.
10
If the machine would have taken the average error rate of all 24 humans and ap-
pliedit to each block, it would have caused a more uniform distribution of errors over
the 24 blocks than the humans. In this case, a risk-averse subject might have preferred to
delegate the task to a machine, because she feared a particularly weak fellow human
subject more than she appreciated a particularly strong fellow human subject.
11
The average number of lines solved correctly was 8.00 ( =sd 2.10 ) in the rst session
and 7.92 ( =sd 1.93 ) in the second session. They were thus very close to each other.
J. Gogoll, M. Uhl 

toward humans and machines between subjects who made dierent
delegation decisions. In particular, one might expect that delegators to
humans are generally more skeptical toward machines and express a
lower level of trust. Subjects who delegated to a human task-solver in
the rst part, however, did not, on average, transfer any less to a ma-
chine than to a human (31.63 ECU (
=sd 15.4
6
) vs. 32.92 ECU
(
=sd 16.3
7
),
=
p
.62
7
according to two-sided MannWhitney U-tests).
Therefore, our fourth hypothesis is also rejected.
Result 4. The level of trust toward machines and toward humans does
not dier signicantly.
4. Post-experimental survey
Because both potential explanations for the very clear relative
aversion to machine use in the moral domain could not be supported in
the experiment, we conducted a survey study in February 2018. In this
survey, we recruited 78 new participants via ORSEE (Greiner, 2004)
and confronted them with a concise description of part 1 of the ex-
periment. They were then asked to indicate their agreement to several
statements on a 7-point Likert scale.
We confronted subjects with ve pairs of statements that were
identical except for the words human task-solverand machine. The
order in which statements were presented was randomized within each
pair. Specically, we investigated the following alternative explana-
tions for the observed aversion to a delegation to a machine. First,
people might feel that a delegator who delegated the task to a machine
holds a human task-solvers benevolent eort in contempt (rst pair of
statements). Second, people might be biased against delegators to ma-
chines when attributing praise and blame for a resulting outcome
(second and third pair). Third, delegators might successfully pass on the
responsibility for negative outcomes to another human but not to a
machine (fourth and fth pair). The eleventh statement was not directly
related to part 1 but represented a general remark on automation in a
morally relevant domain that we included as a control.
Fig. 3 illustrates the average agreement to the ve pairs of state-
ments testing the alternative explanations for an aversion to machine
use in the moral domain and to the eleventh statement that served as a
control.
The gure suggests that the agreements for each pair of statements
concerning delegators to a human and to a machine are quite similar to
each other. In fact, none of the dierences between the delegators to
humans and machines is signicant (p> .100 according to two-sided
Wilcoxon signed-rank tests) except for the rst pair. People do indeed
more readily agree with the idea that delegators to machines hold a
humans work in contempt than that delegators to humans hold the
machines work in contempt ( =
p
.03
1
). The agreement with the former
statement, however, is still close to neutrality. The eect is thus likely
to be driven by a perceived implausibility of the idea that a machines
eort can be held in contempt.
The only statement where participantsanswers tend to clearly de-
viate from neutrality is the eleventh statement, where people rmly
express the opinion that a human pilot should always be able to over-
rule the decision of an autopilot. While we can thus also identify an
aversion to automation in the moral domain in the post-experimental
survey, none of the alternative explanations tested in this survey can be
supported.
5. Discussion
In this study, we compared the frequency of delegation decisions of
a task that aects a third party to machines and humans and elicited
their respective evaluation by impartial observers. It should be stressed
again that the question we posed here was about peoples preference
relation over a machine agent and a human agent, and not about de-
legating versus performing the task oneself. Consequently, subjects had
to delegate in either case and could thus not be blamed merely for
shifting responsibility.
We found that subjects express an aversion to delegating tasks that
fall into the moral domain to machines rather than humans. First, this
manifests in the relatively small fraction of delegators that mandate a
machine rather than a human. Second, it is clear that observers evaluate
the decision to delegate to a machine in the moral domain less favor-
ably than if the delegation is to a human. Interestingly, machine use is
Fig. 3. Results of post-experimental survey.
NOTE: Upper rows of numbers to the right of the graph represent means, medians and standard deviations (sd) for human agent, lower rows represent measures for
machine agent.
J. Gogoll, M. Uhl 

viewed more critically, irrespective of whether the delegation ulti-
mately caused positive or negative consequences for the person af-
fected.
The experiment tested two potential explanations for an aversion to
machine use in the moral domain: an oversensitivity to machine errors,
and a lack of trust in machines. Both explanations could be ruled out in
our experiment. Subjects did not perceive machine errors more saliently
than human errors, as subjectsincentivized guessing of failure rates
demonstrates. The phenomenon identied, therefore, seems to be an
aversion to delegating tasks that aect a third party to machines per se
as opposed to an instrumentally justied attempt to minimize the risk of
failure for those aected. Analogously, the level of trust expressed by
subjects toward a machine agent was very similar to the trust level
expressed toward a human. Thus, we were unable to identify a general
distrust in machines in a self-regarding trust game. This latter nding
indicates that the unconditional aversion to machine use seems to be
rather specic to the delegation of tasks that aect a third party.
Finally, we used a post-experimental survey with fresh subjects to
investigate three alternative explanations for the observed phenom-
enon: the feeling that delegators to machines hold humanseort in
contempt, a bias against delegators to machines when attributing praise
and blame, and a dierence in a delegators ability to pass on respon-
sibility to another human and to a machine. Neither of these potential
alternative explanations, however, could account for the aversion at
hand.
From our ndings, it seems that most people rather intuitively
dislike machine use in the moral domainan intuition which turns out
being hard to rationalize. We identied this aversion per se by ex-
perimentally equalizing humansand the algorithms performance. In
practice, however, algorithms will usually not be simulating human
moral behavior but will be programmed to implement a specic nor-
mative rationale. This attachment to rules may induce a dismissal of
their decisional inexibility in people. Such an instrumental aversion
would then come in addition to the non-instrumental aversion that we
identied. In the case of self-learning algorithms, we might observe an
additional instrumental aversion to decisional opacity.
Our results underline the importance of an open discussion of ma-
chine use in the moral domain. The case of automated driving certainly
qualies as such a domain, since errors of the machine may cause
substantial externalities to third parties. The non-instrumental aversion
identied suggests that the emphasis on the superior performance of
automated cars, which is currently the main argument for automation
in trac, may not be sucient or even decisive in convincing the
general public. It might be as important to address the perceived moral
problems that are necessarily associated with the introduction of au-
tomated vehicles.
Against this background, Chris Urmson, head of Googles self-
driving car project, might be mistaken in downplaying the role of moral
considerations in the context of automated driving by calling them a
fun problem for philosophers to think about(McFarland, 2015). As
this empirical study suggests, concerns regarding the involvement of
machines in the moral domain are not only an issue for armchair phi-
losophers but may reect a larger societal phenomenon, viz. a folk
aversion (see also Kohl et al. (2018)). So far, the industry seems to
mainly be occupied with engineering issues and has, due to a dé-
formation professionnelle, predominantly neglected or downplayed the
possibility of public resistance to the new technology. It may, however,
be well-advised to take moral concerns against automated driving ser-
iously, since citizensresistance may slow down the automation process
substantially. This, however, would mean to preserve a status quo that
involves an avoidably high number of trac deaths, injuries and da-
mages.
Research that investigates how the feeling of unease can be ad-
dressed prophylactically (Feldhütter et al., 2016) is just emerging. En-
abling people to experience, and thus better understand, the technology
in order to dissipate reservations and fears may pave the way for a
trouble-free introduction of autonomous driving. A deeper investigation
of the causes of peoples aversion to the use of automated cars in the
moral domain seems to us a promising venue for future research.
Supplementary material
Supplementary material associated with this article can be found, in
the online version, at doi:http://dx.doi.org/10.1016/j.socec.2018.04.
003.
References
Davis, F.D., 1989. Perceived usefulness, perceived ease of use, and user acceptance of
information technology. MIS Quarterly 13 (3), 319340.
de Visser, E.J., Krueger, F., McKnight, P., Scheid, S., Smith, M., Chalk, S., Parasuraman,
R., 2012. The world is not enough: trust in cognitive agents. Proceedings of the
Human Factors and Ergonomics Society Annual Meeting 56, 263267. http://dx.doi.
org/10.1177/1071181312561062.
Dzindolet, M.T., Beck, H.P., Pierce, L.G., Dawe, L.A., 2001. A Framework of Automation
Use. Army Research Laboratory, Aberdeen Proving Ground.
Dzindolet, M.T., Pierce, L.G., Beck, H.P., Dawe, L.A., 2002. The perceived utility of
human and automated aids in a visual detection task. Human Factors: The Journal of
the Human Factors and Ergonomics Society 44 (1), 7994.
Fagnant, D.J., Kockelman, K., 2015. Preparing a nation for autonomous vehicles: op-
portunities, barriers and policy recommendations. Transportation Research Part A
77, 167181.
Feldhütter, A., Gold, C., Hüger, A., Bengler, K., 2016. Trust in automation as a matter of
media and experience of automated vehicles. Proceedings of the Human Factors and
Ergonomics Society Sixtieth Annual Meeting.
Fischbacher, U., 2007. z-Tree: Zurich toolbox for ready-made economic experiments.
Experimental Economics 10 (2), 171178.
Gao, P., Hensley, R., Zielke, A., 2014. A road map to the future for the auto industry.
McKinsey Quarterly (4), 4253.
Gogoll, J., Müller, J.F., 2017. Autonomous cars: in favor of a mandatory ethics setting.
Science and Engineering Ethics 23 (3), 681700.
Goodall, N., 2014. Ethical decision making during automated vehicle crashes.
Transportation Research Record: Journal of the Transportation Research Board 24,
5865.
Greiner, B., Levati, M.V., 2005. Indirect reciprocity in cyclical networks: an experimental
study. Journal of Economic Psychology 26 (5), 711731.
Greiner, B., 2004. The online recruitment system orsee 2.0-a guide for the organization of
experiments in economics. University of Cologne, Working paper series in economics
10 (23), 63104.
Hevelke, A., Nida-Rümelin, J., 2014. Responsibility for crashes of autonomous vehicles:
an ethical analysis. Science and Engineering Ethics 21 (3), 619630.
Kohl, C., Knigge, M., Baader, G., Böhm, M., Krcmar, H., 2018. Anticipating acceptance of
emerging technologies using twitter: the case of self-driving cars. Journal of Business
Economics. Forthcoming.
Lee, J.D., Moray, N., 1994. Trust, self-condence, and operatorsadaptation to automa-
tion. International Journal of Human-Computer Studies 40, 153184.
Lee, J.D., See, K.A., 2004. Trust in automation: designing for appropriate reliance. Human
Factors 46 (1).
McFarland, M., 2015. Googles chief of self-driving cars downplays dilemma of ethics and
accidents. The Washington Post. https://www.washingtonpost.com/news/
innovations/wp/2015/12/01/googles-leader-on-self-driving-cars-downplays-the-
trolley-problem/.
Muir, B., Moray, N., 1996. Trust in automation. Part II. Experimental studies of trust and
human intervention in a process control simulation. Ergonomics 39 (3), 429460.
http://dx.doi.org/10.1080/00140139608964474.
Ray, C., Mondada, F., Siegwart, R., 2008. What do people expect from robots?
Proceedings of the IEEE/RSJ International Conference Intelligent Robots and
Systems. pp. 38163821.
Ross, M., Sicoly, F., 1979. Egocentric biases in availability and attribution. Journal of
Personality and Social Psychology 37 (3), 322336.
Rousseau, D., Sitkin, S., Burt, R., Camerer, C., 1998. Not so dierent after all: a cross-
discipline view of trust. Academy of Management Review 23, 393404.
Selten, R., 1967. Die Strategiemethode zur Erforschung des eingeschränkt rationalen
Verhaltens im Rahmen eines Oligopolexperiments. In: Sauermann, H. (Ed.), Beiträge
zur experimentellen Wirtschaftsforschung. JCB Mohr, Tübingen.
Shanker, R., Jonas, A., Devitt, S., Huberty, K., Flannery, S., Greene, W., Swinburne, B.,
Locraft, G., Wood, A., Weiss, K., Moore, J., Schenker, A., Jain, P., Ying, Y., Kakiuchi,
S., Hoshino, R., Humphrey, A., 2013. Autonomous cars: self-driving the new auto
industry paradigm. Morgan Stanley & Co. LLC Morgan Stanley Blue Paper.
Stangor, C., McMillan, D., 1992. Memory for expectancy-congruent and expectancy-in-
congruent information: a review of the social and social developmental literatures.
Psychological Bulletin 111 (1), 4261.
Stangor, C., Ruble, D.N., 1989. Strength of expectancies and memory for social in-
formation: what we remember depends on how much we know. Journal of
Experimental Social Psychology 25 (1), 1835.
Svenson, O., 1981. Are we all less risky and more skillful than our fellow drivers? Acta
Psychologica 47 (2), 143148.
J. Gogoll, M. Uhl 

... The rapidly growing literature on how people perceive algorithmic decisions and engage with algorithms lays the ground for our study. People seem to be willing to outsource analytical tasks to an algorithm but are reluctant to do so with social tasks (Lee, 2018;Waytz & Norton, 2014;Hertz & Wiese, 2019;Buchanan & Hickman, 2023) and are particularly averse to algorithms in the moral domain (Gogoll & Uhl, 2018;Bigman & Gray, 2018). If algorithms are employed in "human tasks", their perceived lack of intuition and subjective judgment capabilities causes them to be judged as less fair and trustworthy (Lee, 2018) or reductionist (Newman et al., 2020). ...
... It generally finds opposing results on whether people are averse (e.g., Dietvorst et al., 2015) or appreciative of algorithms (e.g., Logg et al., 2019), but there is no apparent consensus on the overall general preference. In moral contexts, however, such as in our experiment where decisions are driven by fairness principles and beliefs, people are found to have a particularly strong aversion to algorithms (Gogoll & Uhl, 2018;Bigman & Gray, 2018) while simultaneously seeing them as more objective and rational than a human advisor (Dijkstra et al., 1998) and with a "halo" of scientific authority (Cowgill et al., 2020). The perceived fairness of automated decisions may also be driven by the increased procedural fairness associated with the use of algorithms, as they decide "without regard for persons" (Weber, 1978, p. 975 on benefits of bureaucracy). ...
... In line with the findings of Gogoll and Uhl (2018) and Bigman and Gray (2018), we observe that if a moral decision is made by a human DM it is rated as about a quarter of a point more fair and participants report higher satisfaction (DM Human). We, therefore, fail to reject our H 3 on the DM's nature and the impact on satisfaction and perceived fairness. ...
Article
Full-text available
As technology-assisted decision-making is becoming more widespread, it is important to understand how the algorithmic nature of the decision maker affects how decisions are perceived by those affected. We use an online experiment to study the preference for human or algorithmic decision makers in redistributive decisions. In particular, we consider whether an algorithmic decision maker will be preferred because of its impartiality. Contrary to previous findings, the majority of participants (over 60%) prefer the algorithm as a decision maker over a human—but this is not driven by concerns over biased decisions. However, despite this preference, the decisions made by humans are regarded more favorably. Subjective ratings of the decisions are mainly driven by participants’ own material interests and fairness ideals. Participants tolerate any explainable deviation between the actual decision and their ideals but react very strongly and negatively to redistribution decisions that are not consistent with any fairness principles.
... 3 Increased morality through social communication or decision situation worsens the acceptance of robo-advisors et al., 2021). Moreover, there is an aversion towards moral decision made by machines (Bigman and Gray, 2018;Gogoll and Uhl, 2018) or algorithms (Niszczota and Kaszás, 2020). Some researchers argue, that algorithms or robots are not suitable for moral decision-making because they lack moral understanding, accountability, and emotion (Véliz, 2021;Lucifora et al., 2020;Hakli and Mäkelä, 2019). ...
... The decreasing acceptance of the robot recommendation is interpreted as increasing algorithm aversion due to the increasing morality of the decision scenarios. Thus, this study supports research suggesting aversion in moral contexts (Starr et al., 2021;Véliz, 2021;Lucifora et al., 2020;Niszczota and Kaszás, 2020;Hakli and Mäkelä, 2019;Bigman and Gray, 2018;Gogoll and Uhl, 2018). ...
... Most interestingly, we found a more moral perception of the robot in social communication versus functional communication, resulting in a lower acceptance of the robot as an advisor in moral decision-making situations. Previous studies demonstrated that humans are averse to machines making moral decisions (Bigman and Gray, 2018) or performing moral tasks (Gogoll and Uhl, 2018). Moral decision-making situations are characterized by algorithm aversion (Niszczota and Kaszás, 2020) as well as a negative attitude towards robots (Véliz, 2021;Starr et al., 2021;Lucifora et al., 2020;Hakli and Mäkelä, 2019). ...
Preprint
Full-text available
This German study (N = 317) tests social communication (i.e., self-disclosure, content intimacy, relational continuity units, we-phrases) as a potential compensation strategy for algorithm aversion. Therefore, we explore the acceptance of a robot as an advisor in non-moral, somewhat moral, and very moral decision situations and compare the influence of two verbal communication styles of the robot (functional vs. social). Subjects followed the robot's recommendation similarly often for both communication styles (functional vs. social), but more often in the non-moral decision situation than in the moral decision situations. Subjects perceived the robot as more human and more moral during social communication than during functional communication but similarly trustworthy, likable, and intelligent for both communication styles. In moral decision situations, subjects ascribed more anthropomorphism and morality but less trust, likability, and intelligence to the robot compared to the non-moral decision situation. Subjects perceive the robot as more moral in social communication. This unexpectedly led to subjects following the robot's recommendation less often. No other mediation effects were found. From this we conclude, that the verbal communication style alone has a rather small influence on the robot's acceptance as an advisor for moral decision-making and does not reduce algorithm aversion. Potential reasons for this (e.g., multimodality, no visual changes), as well as implications (e.g., avoidance of self-disclosure in human-robot interaction) and limitations (e.g., video interaction) of this study, are discussed.
... 13 Kennedy, Waggoner, and Ward (2022); Zwald, Kennedy, and Ozer (2021); Logg (2016); Logg, Minson, and Moore (2019). 14 Dietvorst, Simmons, and Massey (2015); Gogoll and Uhl (2018). 15 Merkley (2021); Merkley and Loewen (2021). ...
... In all scenarios, the judge decides to release the defendant on probation and the defendant is subsequently involved in another drunk driving accident that kills a pedestrian. Massey (2015, 2018); Gogoll and Uhl (2018). 43 Bertsou (2021). ...
... 60 Gelman and Stern (2006). 61 Massey (2015, 2018); Gogoll and Uhl (2018); Dawes (1979). 62 Hoff and Bashir (2015). ...
Article
Full-text available
Public decision-makers incorporate algorithm decision aids, often developed by private businesses, into the policy process, in part, as a method for justifying difficult decisions. Ethicists have worried that over-trust in algorithm advice and concerns about punishment if departing from an algorithm’s recommendation will result in over-reliance and harm democratic accountability. We test these concerns in a set of two pre-registered survey experiments in the judicial context conducted on three representative U.S. samples. The results show no support for the hypothesized blame dynamics, regardless of whether the judge agrees or disagrees with the algorithm. Algorithms, moreover, do not have a significant impact relative to other sources of advice. Respondents who are generally more trusting of elites assign greater blame to the decision-maker when they disagree with the algorithm, and they assign more blame when they think the decision-maker is abdicating their responsibility by agreeing with an algorithm.
... Meanwhile, in laboratory experiments or field studies representative of the many empirical investigations led over the past decade, Gogoll and Uhl (2018) observed that most subjects prefer to delegate a calculation task affecting a third party to a human rather than a machine, Dietvorst et al. (2015) and Prahl and Van Swol (2017) that trust in algorithmic assistance is more sensitive to receiving an erroneous advice than is trust in human aid, Yeomans et al. (2019) that people refrain from using recommender systems in certain domains (in this case, picking jokes that people find funny), even when these systems outperform human advisors, and Liu et al. (2023) that drivers tend not to rely on algorithms if their peers do not or if they have relevant expertise. Similar behaviors have been reported to repeatedly occur in several occupations (e.g., auditing, consulting, forecasting, retail business, stock trading) and across various other groups of people (e.g., consumers and investors, physicians, and managers). ...
... (i) Some factors have to do with some characteristics of the task under consideration (Castelo et al., 2019). For instance, machines are not yet accepted as autonomous moral agents (Bigman & Gray, 2018;Gogoll & Uhl, 2018;Jago, 2019). In their empirical survey, Ray et al. (2008) find that people would like robots to watch over their house, mow the lawn or do the laundry, but they would strongly resist having a machine babysit their children. ...
... The task type and context in which the algorithm is operating seem to be important factors for human acceptance. There is some evidence, that people are willing to delegate decisions to automated agents in analytical or objective contexts, they seem hesitant to do so in social (Waytz and Norton, 2014;Lee et al., 2018;Hertz and Wiese, 2018;Castelo, 2019) or moral contexts (Gogoll and Uhl, 2018;Bigman and Gray, 2018) and in domains where trust plays an important role (Dietvorst and Bharti, 2019). Further experimental evidence suggests that the role of algorithms and AI systems, i.e., whether they take over a task (Önkal et al., 2009;Caro and de Tejada Cuenca, 2023;Longoni et al., 2019) or act as an adviser or decisionsupport (Promberger and Baron, 2006;Palmeira and Spassova, 2015;Bigman and Gray, 2018;Dietvorst et al., 2018;Longoni et al., 2019), influences humans attitudes toward the algorithm or AI under consideration. ...
Preprint
Full-text available
Regulators of artificial intelligence (AI) emphasize the importance of human autonomy and oversight in AI-assisted decision-making (European Commission, Directorate-General for Communications Networks, Content and Technology, 2021; 117th Congress, 2022). Predictions are the foundation of all AI tools; thus, if AI can predict our decisions , how might these predictions influence our ultimate choices? We examine how salient, personalized AI predictions affect decision outcomes and investigate the role of reactance, i.e., an adverse reaction to a perceived reduction in individual freedom. We trained an AI tool on previous dictator game decisions to generate personalized predictions of dictators' choices. In our AI treatment, dictators received this prediction before deciding. In a treatment involving human oversight, the decision of whether participants in our experiment were provided with the AI prediction was made by a previous participant (a 'human overseer'). In the baseline, participants did not receive the prediction. We find that participants sent less to the recipient when they received a personalized prediction but the strongest reduction occurred when the AI's prediction was intentionally not shared by the human overseer. Our findings underscore the importance of considering human reactions to AI predictions in assessing the accuracy and impact of these tools as well as the potential adverse effects of human oversight.
... Recent studies have shown that people, regardless of being informed about whether advice originated from a machine or a human, exhibited unchanged behavior [54,59] and could sometimes "hide behind" machine decisions [60]. However, it appears that humans sometimes hesitate to entrust decisions to machines, and they may face scrutiny from others when they choose to do so [61]. Either way, this highlights the significance of promoting digital literacy rather than solely emphasizing transparency. ...
Article
Full-text available
In the era of generative AI and specifically large language models (LLMs), exemplified by ChatGPT, the intersection of artificial intelligence and human reasoning has become a focal point of global attention. Unlike conventional search engines, LLMs go beyond mere information retrieval, entering into the realm of discourse culture. Their outputs mimic well-considered, independent opinions or statements of facts, presenting a pretense of wisdom. This paper explores the potential transformative impact of LLMs on democratic societies. It delves into the concerns regarding the difficulty in distinguishing ChatGPT-generated texts from human output. The discussion emphasizes the essence of authorship, rooted in the unique human capacity for reason—a quality indispensable for democratic discourse and successful collaboration within free societies. Highlighting the potential threats to democracy, this paper presents three arguments: the Substitution argument, the Authenticity argument, and the Facts argument. These arguments highlight the potential risks that are associated with an overreliance on LLMs. The central thesis posits that widespread deployment of LLMs may adversely affect the fabric of a democracy if not comprehended and addressed proactively and properly. In proposing a solution, we advocate for an emphasis on education as a means to mitigate risks. We suggest cultivating thinking skills in children, fostering coherent thought formulation, and distinguishing between machine-generated output and genuine, i.e., human, reasoning. The focus should be on the responsible development and usage of LLMs, with the goal of augmenting human capacities in thinking, deliberating and decision-making rather than substituting them.
... People tend to hesitate in utilizing algorithms when they perceive them as lacking competence [24,25]. Negative attitudes discourage individuals not to rely on algorithms [26,27]. Prior AI adoption literature suggested a positive association between attitudes and the adoption of algorithms [23,28]. ...
Chapter
Algorithm aversion, characterized by the tendency to distrust algorithmic advice despite its demonstrated superior or identical performance, has become an increasingly concerning issue as it reduces the practical utility of algorithms. To gain insights into this phenomenon, our research centers on individual traits, specifically focusing on familiarity with algorithms and familiarity with the task at hand, and their connections with attitudes toward algorithms. We construct a causal model to delve into these relationships and assess how attitudes, in turn, impact algorithm aversion. Our analysis draws upon data collected through an online survey involving 160 participants, and we employ PLS-SEM for our analysis. The results underscore a noteworthy positive correlation between familiarity with algorithms and attitudes toward algorithms. Interestingly, our findings indicate that familiarity with the task or domain knowledge does not significantly influence attitudes. Moreover, attitudes are demonstrated to have a negative impact on algorithm aversion. These discoveries hold significant implications for comprehending and addressing the issue of algorithm aversion. They shed light on the roles of individual traits and attitudes in shaping people's acceptance of algorithms, ultimately offering valuable insights for mitigating this phenomenon.
... Firstly, respondents found the use of AI acceptable in all areas of medicine that were asked about, including liver allocation. Some studies have indicated that the public generally prefer human decision-makers to AI in medical and ethical decision-making, [51][52][53][54] even if there may not be an explicit reason for this [63]. Therefore, to avoid responses skewed unfairly against AI, we simplified the question and asked how acceptable people find AI without a human comparator. ...
Article
Full-text available
Background Allocation of scarce organs for transplantation is ethically challenging. Artificial intelligence (AI) has been proposed to assist in liver allocation, however the ethics of this remains unexplored and the view of the public unknown. The aim of this paper was to assess public attitudes on whether AI should be used in liver allocation and how it should be implemented. Methods We first introduce some potential ethical issues concerning AI in liver allocation, before analysing a pilot survey including online responses from 172 UK laypeople, recruited through Prolific Academic. Findings Most participants found AI in liver allocation acceptable (69.2%) and would not be less likely to donate their organs if AI was used in allocation (72.7%). Respondents thought AI was more likely to be consistent and less biased compared to humans, although were concerned about the “dehumanisation of healthcare” and whether AI could consider important nuances in allocation decisions. Participants valued accuracy, impartiality, and consistency in a decision-maker, more than interpretability and empathy. Respondents were split on whether AI should be trained on previous decisions or programmed with specific objectives. Whether allocation decisions were made by transplant committee or AI, participants valued consideration of urgency, survival likelihood, life years gained, age, future medication compliance, quality of life, future alcohol use and past alcohol use. On the other hand, the majority thought the following factors were not relevant to prioritisation: past crime, future crime, future societal contribution, social disadvantage, and gender. Conclusions There are good reasons to use AI in liver allocation, and our sample of participants appeared to support its use. If confirmed, this support would give democratic legitimacy to the use of AI in this context and reduce the risk that donation rates could be affected negatively. Our findings on specific ethical concerns also identify potential expectations and reservations laypeople have regarding AI in this area, which can inform how AI in liver allocation could be best implemented.
... It requires addressing potential biases in managerial decisions and recognizing the importance of human moral capacities in sensitive areas. In referencing self-driving cars, Gogoll and Uhl, (2018) speak to the 'rage against machine automation' in the moral domain, discussing the idea that certain decisions should inherently belong to humans due to ethical obligations and norms, regardless of the potential benefits of automation. Such deontological considerations can help mitigate potential risks and harms, enhancing accountability and fairness in algorithmic decision-making by holding on to human essence in service of attributing action where necessary. ...
Conference Paper
Full-text available
This paper problematizes the literature on the non-use of algorithmic decision-making systems (ADMS), commonly examined as algorithm aversion. Whilst prior literature attributes algorithm aversion primarily to human bias and irrationality, assuming utility-based evaluations, we argue that it may also stem from values-based evaluations of technology, which are overlooked. Through an integrated ethical analysis, drawing upon the "big three" ethical theories of consequentialism, deontology, and virtue ethics, we examine implicit normative judgments within the algorithm aversion literature. Consequently, we positively reframe algorithm aversion as a potentially principled resistance to ADMS, expanding prior views of the phenomenon. We argue that such resistance may be constructive and lead to a better alignment of ADMS with societal needs and values. Thus, we call on IS scholars to explore this phenomenon as an ethical and sociotechnical issue, rather than as a costly problem to be mitigated, as prior literature might suggest.
Preprint
Full-text available
Using unique data from a leading peer-to-peer (P2P) lending platform, we investigate the link between past investment performance and choice of auto-investing tool. Our results suggest that investors with poorly performing loan portfolios are more likely to switch automatically. This negative relationship can be explained by algorithmic aversion or investor inattention. In other words, the results suggest that good-performing investors who pay close attention to their loan portfolios or are not interested in using automated services are more likely to rely on themselves in manual mode. These results are robust to alternative specifications.
Article
Full-text available
In an early stage of developing emerging technologies, there is often great uncertainty regarding their future success. Companies can reduce this uncertainty by listening to the voice of customers as the customer eventually decides to accept an emerging technology or not. We show that risk and benefit perceptions are central determinants of acceptance of emerging technologies. We present an analysis of risk and benefit perception of self-driving cars from March 2015 until October 2016. In this period, we analyzed 1,963,905 tweets using supervised machine learning for text classification. Furthermore, we developed two new metrics, risk rate (RR) and benefit rate (BR), which allow analyzing risk and benefit perceptions on social media quantitatively. With our results, we provide impetus for further research on acceptance of self-driving cars and a methodological contribution to acceptance of emerging technologies research. Furthermore, we identify crucial issues in the public perception of self-driving cars and provide guidance for the management of emerging technologies to increase the likelihood of their acceptance.
Conference Paper
Full-text available
Highly automated vehicles (HAV), which could help to enhance road safety and efficiency, are very likely to enter the market within the next decades. To have an impact, these systems need to be purchased, which is a matter of trust and acceptance. These factors are dependent on the level of information that one has about such systems. One important source of information is various media, such as newspapers, magazines and videos, in which highly automated driving (HAD) is currently a frequent topic of discussion. To evalu-ate the influence of media on the perception of HAD, 31 participants were presented with three different types of media addressing HAD in a neutral manner. Afterwards, the participants experienced HAD in the driving simulator. In between these steps, the participants completed questionnaires assessing comfort, trust in automation, increase in safety, intention to use and other factors in order to analyze the effect of the me-dia and the driving simulation experience. Results indicate that the perception of some aspects of HAD were affected by the media presented, while experiencing HAD in the driving simulator generally did not have an effect on the attitude of the participants. Other aspects, such as trust, were not affected by either media or experience. In addition, gender-related differences in the perception of HAD were found.
Article
Full-text available
The recent progress in the development of autonomous cars has seen ethical questions come to the forefront. In particular, life and death decisions regarding the behavior of self-driving cars in trolley dilemma situations are attracting widespread interest in the recent debate. In this essay we want to ask whether we should implement a mandatory ethics setting (MES) for the whole of society or, whether every driver should have the choice to select his own personal ethics setting (PES). While the consensus view seems to be that people would not be willing to use an automated car that might sacrifice themselves in a dilemma situation, we will defend the somewhat contra-intuitive claim that this would be nevertheless in their best interest. The reason is, simply put, that a PES regime would most likely result in a prisoner’s dilemma.
Article
Automation is often problematic because people fail to rely upon it appropriately. Because people respond to technology socially, trust influences reliance on automation. In particular, trust guides reliance when complexity and unanticipated situations make a complete understanding of the automation impractical. This review considers trust from the organizational, sociological, interpersonal, psychological, and neurological perspectives. It considers how the context, automation characteristics, and cognitive processes affect the appropriateness of trust. The context in which the automation is used influences automation performance and provides a goal-oriented perspective to assess automation characteristics along a dimension of attributional abstraction. These characteristics can influence trust through analytic, analogical, and affective processes. The challenges of extrapolating the concept of trust in people to trust in automation are discussed. A conceptual model integrates research regarding trust in automation and describes the dynamics of trust, the role of context, and the influence of display characteristics. Actual or potential applications of this research include improved designs of systems that require people to manage imperfect automation. Copyright © 2004, Human Factors and Ergonomics Society. All rights reserved.
Article
A meta-analysis of 54 experiments investigated the influence of social expectations on memory for information that is congruent and incongruent with those expectations. Results showed that overall, memory was better for expectancy-incongruent than expectancy-congruent information on recall and recognition sensitivity measures. Recognition measures that were uncorrected for response biases produced an overall tendency to report expectancy-congruent information as having been seen. A number of moderator variables influenced the strength of these overall effects, including the strength of the expectancy used to guide information processing, the complexity or cognitive demands of the processing task, set size, the type of expectancy, the type of target, Ss' information-processing goals, and the delay between exposure to the stimulus information and the memory test. The overall pattern of results appears to be most parsimoniously explained in terms of the influence of contextual variables on the perceiver's motivation to resolve incongruity.