ArticlePDF Available

Strategic Opportunities (and Challenges) of Algorithmic Decision-Making: A Call for Action on the Long-Term Societal Effects of 'Datification'

Authors:

Figures

Content may be subject to copyright.
Strategic opportunities (and challenges) of algorithmic
decision-making: A call for action on the long-term societal
effects of ‘datification’
Sue Newell
a,
, Marco Marabelli
b
a
School of Business, Management and Economics, University of Sussex, Brighton BN1 9RH, UK
b
IPM Department, Bentley University, Waltham, MA 02452, USA
article info
Article history:
Accepted 6 February 2015
Available online xxxx
Keywords:
Algorithmic decision-making
Digital traces
Sensors
Strategic
Social and ethical issues
abstract
Today, digital data are captured through a variety of devices that have the ability to
monitor the minutiae of an individual’s everyday life. These data are often processed by
algorithms, which support (or drive) decisions (termed ‘algorithmic decision-making’ in
this article). While the strategic value of these data (and subsequent analysis) for business-
es is unquestionable, the implications for individuals and wider society are less clear.
Therefore, in this Viewpoint article we aim to shed light on the tension between businesses
– that increasingly profile customers and personalize products and services – and
individuals, who, as McAfee and Brynjolfsson (2012, p. 5) suggest, are ‘walking data
generators’ but are often unaware of how the data they produce are being used, and by
whom and with what consequences. Issues associated with privacy, control and
dependence arise, suggesting that social and ethical concerns related to the way business
is strategically exploiting digitized technologies that increasingly support our everyday
activities should be brought to the fore and thoughtfully discussed. In this article we aim
to lay a foundation for this discussion in the IS community and beyond.
!2015 Elsevier B.V. All rights reserved.
Introduction
The last decade has witnessed the widespread diffusion of digitized devices that have the ability to monitor the minutiae
of our everyday lives (Hedman et al., 2013). Nolan (2012, p. 91) argues that ‘Global IT has enabled information on most
everything to flow most everywhere at stealth speed’. The data trail we leave is increasingly used by companies to manage
employees and target and personalize products and services for clients and customers, based on developing algorithms that
can make predictions about individuals by recognizing complex patterns in huge data sets compiled from multiple sources.
In this article we consider some of the observed and potential consequences of this new type of data-driven, algorithmic
decision-making, illustrating that while it can offer strategic opportunities for business and sometimes benefits for indi-
viduals, there are also costs, hence raising societal issues: as Galliers et al. (2012) indicate, there can be a difference between
how business is benefiting and how society is benefiting – or otherwise.
The IS literature has already raised social and ethical concerns associated with IT (Smith, 2002; Smith and Hasnas, 1999),
and in particular those concerns are often associated with privacy issues (e.g., see Belanger and Crossler, 2011; Chan et al.,
2005; Coll, 2014; Greenaway and Chan, 2005). However, few IS studies have linked these concerns with the digitization of
http://dx.doi.org/10.1016/j.jsis.2015.02.001
0963-8687/!2015 Elsevier B.V. All rights reserved.
Corresponding author.
E-mail addresses: sue.newell@sussex.ac.uk (S. Newell), mmarabelli@bentley.edu (M. Marabelli).
Journal of Strategic Information Systems xxx (2015) xxx–xxx
Contents lists available at ScienceDirect
Journal of Strategic Information Systems
journal homepage: www.elsevier.com/locate/jsis
Please cite this article in press as: Newell, S., Marabelli, M. Strategic opportunities (and challenges) of algorithmic decision-making: A call
for action on the long-term societal effects of ‘datification’. J. Strateg. Inform. Syst. (2015), http://dx.doi.org/10.1016/j.jsis.2015.02.001
our everyday life (exceptions include Abbas et al., 2014; Boyd and Crawford, 2014; Lyon, 2014; Slade and Prinsloo, 2013),
and fewer still have discussed this phenomenon in relation to algorithmic decision-making (one exception being
Schroeder and Cowls, 2014). Here, we focus on the consequences of ‘algorithmic decision-making’, which occurs when data
are collected through digitized devices carried by individuals such as smartphones and technologies with inbuilt sensors –
and subsequently processed by algorithms, which are then used to make (data-driven) decisions. That is, decisions are based
on relationships identified in the data, and the decision maker often ignores why such relationships may be present
(Mayer-Schonberger and Cukier, 2013). While these data-driven decisions made by businesses lead to personalized offerings
to individuals, they also result in the narrowing of their choices (Newell and Marabelli, 2014).
Given the above, we argue that algorithmic decision-making has societal consequences that may not always be positive
and, in this Viewpoint article, we aim to articulate such concerns. In so doing, we bring to the fore the issues related to algo-
rithmic decision-making and highlight the interdisciplinary nature of this topic (Chen et al., 2012; Smith et al., 2011). As we
have indicated, some work has been done to shed light on the social implications of the widespread diffusion of digital
devices in the IS community, but also in other disciplines such as sociology – as in the work of Lyon (2001, 2003, 2014),
Doyle et al. (2013), and Ball (2002, 2005) on impacts of monitoring and surveillance on society, and of Castells et al.
(2009) and Campbell and Park (2008) on societal changes determined by the diffusion of digital devices. Here, we call for
IS research that examines (and challenges) corporations (and governments) in terms of the strategic decisions that are being
made based on data that we are now constantly providing them (see also MacCrory et al., 2014), whether we realize it or not.
Next, we define some key concepts and set the boundaries of our analysis.
Big data, little data, and algorithmic decision-making
Data-driven or ‘algorithmic’ decision-making is based on collecting and analyzing large quantities of data that are then
used to make strategic decisions. Algorithmic decision-making incorporates two main characteristics: firstly, decision-
makers rely on information provided by algorithms that process huge amounts of data (often big data, as we will explain
next); secondly, the reasons behind the ‘suggestions’ made by the algorithms are often ignored by decision-makers
(Mayer-Schonberger and Cukier, 2013). We expand on both characteristics below.
Digitized technologies and data analytics
Data that originate from digitized devices are increasingly permeating our everyday lives. These digitized devices have
the ability to keep track of and record what we do. As a result, somebody else may eventually be able to use the data thus
produced – often with purposes different from those originally intended. Thus, we focus on ‘digital traces’ – all data provided
by individuals (1) during ‘IT-related’ activities, captured from social networks, online shopping, blogs, but also ATM with-
drawals, and other activities that will leave a ‘trace’ (Hedman et al. 2013; Wu and Brynjolfsson, 2009) and (2) that are cap-
tured through technologies that we use that have in-built sensors. These technologies include LBS (Location Based
Technologies) that are IT artifacts equipped with GPS systems and so have the ability to collect a user’s location such as a
smartphone with GPS – see Abbas et al. (2014) and Michael and Michael (2011) for social implications – and other surveil-
lance and monitoring devices – see the previously cited work of Lyon (2001, 2003, 2014) for privacy implications.
It is clear that the huge amount of digital trace data that are collected through the many digitized devices that we now use
to support our daily activities fall into the ‘big data’ umbrella. The big data (analytics) concept is very similar to the more
familiar (and less sexy) business intelligence that has been studied for the past decade or so (e.g., Negash, 2004; Power,
2002; Rouibah and Ould-ali, 2002; Thomsen, 2003). McAfee and Brynjolfsson (2012). Following Gartner’s (2001) definition,
it is the three Vs of big data
1
on which we focus: Volume (the amount of data determines value); Variety (data arise from dif-
ferent sources/databases and are cross-matched to find relationships), and Velocity (data are generated quickly). Big data
encompasses much more than this individually generated data trail (see Chen et al., 2012 for a broad discussion of big data
analytics) but here we focus just on this everyday digital trail that we each leave. That is, we focus on those big data that
are generated by individuals during their everyday lives (and are captured as digital traces). In other words, we focus on data
that arise as a consequence of each of us now being a ‘walking data generator’ (McAfee and Brynjolfsson, 2012, p. 5). This
attention to the digitization of our everyday life allows us to narrow the focus of our inquiry and to expand on concerns
regarding the use (and abuse) of one aspect of big data analytics that concerns algorithm-driven decision-making and associated
personalization – to which we now turn.
Algorithmic decision-making
(Big) data captured through digitized devices are processed by algorithms aimed at predicting what a person will do,
think and like on the basis of their current (or past) behaviors. These algorithms can predict particular outcomes, as with
1
The definition of big data was updated by Gartner in 2012 as they now describe the concept as ‘high volume, high velocity, and/or high variety information
assets that require new forms of processing to enable enhanced decision making, insight discovery and process optimization (Gartner, 2012). Moreover, others
have added ‘new Vs’ – e.g., veracity, variability, visualization, and value, viewing big data in terms of 5 or even 7 Vs. Here, where we stick with the original
definition (Gartner, 2001) as this reflects the essence of big data for the purposes of this article.
2S. Newell, M. Marabelli/ Journal of Strategic Information Systems xxx (2015) xxx–xxx
Please cite this article in press as: Newell, S., Marabelli, M. Strategic opportunities (and challenges) of algorithmic decision-making: A call
for action on the long-term societal effects of ‘datification’. J. Strateg. Inform. Syst. (2015), http://dx.doi.org/10.1016/j.jsis.2015.02.001
the numbers of ‘friends’ on Facebook being used to predict a person’s credit risk (http://www.google.com/patents/
US8560436) or an individual’s Facebook ‘likes’ on a college Facebook page, used to predict her/his willingness to become
a donator (http://www.nytimes.com/2015/01/25/technology/your-college-may-be-banking-on-your-facebook-likes.html?_
r=0). Interestingly, these predictions often represent a black-box: while humans must decide what to measure and produce
the algorithms to analyze the data being collected, these decisions do not necessarily involve understanding the causes and
consequences of particular patterns of behavior that are identified (Mayer-Schonberger and Cukier, 2013). Rather, it is
deemed sufficient that connections are discovered. Traditionally, making decisions has been a human-centered, knowl-
edge-based activity with humans discriminating on the basis of an understanding of theory or context (Tsoukas and
Vladimirou, 2001). By contrast, algorithmic decision-making means that discriminations are increasingly being made by
an algorithm, with few individuals actually understanding what is included in the algorithm or even why. In other words,
it is seen as being sufficient that an algorithm is successfully predictive, never mind if the reasons for the associations found
in the data from different sources are unknown. We argue that this is likely to create problems when no one in a corporation
really understands why some decisions are made. For example, one could argue that the last financial crisis was at least par-
tially a product of this problem, with the algorithms that predicted the pricing for mortgage-backed securities clearly not
taking into account all the risks while at the same time not being subject to question because the basis of the algorithm
was neither clear nor easily accessible, either to the senior managers in the financial institutions where the algorithms were
being used or to the credit rating agencies who were evaluating these products (Clark and Newell, 2013).
In sum, here we focus on data collected through digitized devices that we increasingly use to support our everyday activ-
ities. This is ‘big data’, because the three (or more) Vs of Gartner’s (2001, 2012) definition apply. In fact, data coming from
digitized technologies are high in volume because of the widespread diffusion of digital devices that allow access to social
networks at any time, as well as all other types of technologies that record what we do even if we do not ‘own’ them (e.g.,
surveillance cameras, or an ATM card machine, where the usage information goes into our bank’s database). Thus, data come
from different sources (variety). For instance, data used for making ‘algorithmic decisions’ may come from a combination of
contributions on social networks and LBS systems (e.g., a ‘check in’), or spending capacity of consumers associated with per-
sonal facts of individuals (e.g., the partner’s birthday). Data velocity is clearly another characteristic of the digitization of our
everyday life, because we are ‘walking data generators’ 24/7 and ‘More data cross the Internet every second than were stored
in the entire Internet just 20 years ago’ (McAfee and Brynjolfsson, 2012, p. 4). On this point, it is worth noting that most of
the digitized devices that collect such individual level activity data fall under the Internet of Things (IoT) umbrella (Miorandi
et al., 2012; Xi et al., 2012). However, we do not restrict our analysis to those digitized devices that are connected to the
Internet because some devices remain (for now) independent of the Internet (e.g., some OBD devices). One such example
is provided by Progressive Insurance in the USA (http://www.progressive.com), which provides a memory stick that is
plugged into a car’s on-board computer and the data must be uploaded to the insurance company rather than automatically
sent via the Internet.
Potential discriminations associated with the (ab)use of algorithmic decision-making: big and little data
The use of algorithmic decision-making associated with data coming from the digitization of our everyday lives improves
the capacity of a business to make discriminations. Thus, businesses have always discriminated in terms of to whom they
offer products and services, because products and services are targeted to different audiences (we cannot, unfortunately
all afford to buy a Bentley car). With algorithmic decision-making they are simply taking this a step further. For example,
they can now much more precisely target and personalize offerings to customers and potential customers – those predicted
to buy particular products or services. As a more specific example, a car’s computer that monitors speed, usage of brakes,
horn, lights, etc. (such as Progressive Insurance’s OnStar OBD technologies mentioned above) has the ability capture all these
details that are then sent to data centers. Computers then analyze the (big) data and insurance companies are able to use the
results to discriminate (e.g., by charging young men higher premiums because the data indicate that they – generally – drive
less safely than other categories of drivers). Such data-driven decision-making has been questioned because it can go against
the ethical principle of equal or fair treatment. This is exemplified in the recent case in the EU, where insurers are required to
no longer use statistical evidence about gender differences to set premiums. Thus, despite the fact that gender differences are
clear from the data (e.g., young male drivers are ten times more likely to be killed or injured than those – of both sexes – over
the age of 35; women live, on average, longer than men), it is considered to be discriminatory (following an EU ruling that
came into effect in December 2012) to use this trend evidence to differentiate between premiums (e.g., car insurance or actu-
arial rates) for men and women. The point about this change in the law is that it was considered to be discriminatory
because, for example, while young men in general may drive more recklessly and so be more prone to accidents, an indi-
vidual young man may not and would therefore be discriminated against when insurers set premiums based on group trends
observable in collective data.
While using big data and algorithmic decision-making to observe trends and so discriminate between groups of indi-
viduals can have social consequences that are potentially unfair, this targeting can now be taken further when data are used
not to predict group trends but to predict the behavior of a specific individual. This is sometimes described as ‘little’ data –
although it should be noted that little data are actually based on big data but are simply used in a more targeted way. Thus,
little data focuses on the everyday minutiae of specific individuals, using computing capacity to collect extremely granular
data (Munford, 2014). Drawing on the previous example of a car’s OBD, little data can now allow us to concentrate on a
S. Newell, M. Marabelli/ Journal of Strategic Information Systems xxx (2015) xxx–xxx 3
Please cite this article in press as: Newell, S., Marabelli, M. Strategic opportunities (and challenges) of algorithmic decision-making: A call
for action on the long-term societal effects of ‘datification’. J. Strateg. Inform. Syst. (2015), http://dx.doi.org/10.1016/j.jsis.2015.02.001
specific driver, and we can decide whether an individual is a good or bad driver based on the sensor data from his/her car.
Sensors have the ability to capture individual’s behaviors and are widespread. As an illustration, consider that approximately
85% of handsets now have a GPS system chipset installed (Abbas et al., 2014). By using sensor data, the insurer would not be
setting premiums based on the general trends in accident rates between groups, but instead would base their calculations on
the actual driving habits of an individual. However, if little data are more ‘objective’ in terms of discriminations made by
corporations, it probably poses more issues for societies given the observed or potential social consequences; for instance,
in terms of an individual’s privacy (Lyon, 2014) or in terms of the exploitation of the vulnerable – an issue that IS scholars
seem not to have fully addressed as yet.
It is then clear that algorithmic decision-making poses two main concerns in terms of big and little data: first, (in terms of
big data) this data trail provides the opportunity for organizations to move to algorithmic decision-making, which McAfee
and Brynjolfsson (2012) argue, is superior to traditional ‘HiPPO’ (highest-paid person’s opinion) decision-making. Algorith-
mic decision-making is, they argue, superior to human judgment-based decisions because of all the inherent biases in human
judgment (Hodgkinson et al., 2002). However, we question this assumption because making decisions on the basis of big
data (and algorithms) might create unfair discriminations. Second, we argue that monitoring an individual’s behavior poses
societal concerns since ‘‘the digital artifacts will be able to remember where they [individuals] were, who used them, the
outcomes of interactions, etc.’’ (Yoo, 2010, p. 226) and this often happens without individuals even being aware that they
are being monitored. Thus, we posit that undertaking research to assess these societal harms, so that corporations can be
held responsible and citizens become more aware, can potentially be very useful.
Below we identify three tradeoffs that involve issues associated with the use by corporations (and governments) of data
from digitized devices that support our daily activities, and in particular with the strategy of using data analytics. The first of
these considers the privacy of individuals versus security for society – an issue that is preeminent in people’s minds
following the recent terrorist attacks, particularly in Paris, in January 2015.
Tradeoffs and societal issues associated with big (and little) data
Privacy versus security
Digitized devices can improve security, and examples include the security-tracking systems adopted for prison popula-
tions, when prisoners are released but are required to wear a tracking ankle-bracelet. These systems are aimed at improving
the overall security of our society, with the sensor acting as a deterrent for prisoners to escape or commit a crime when they
are on parole. Other instances where security is enhanced by everyday digitized devices is in the capacity of sensors to trace
a stolen device, or a kidnapped child, as in the case that occurred in September 2013 in Texas, where the Houston police were
able to trace the whereabouts of a kidnapper by tracing the iPad that he had with him in his car (http://abc13.com/archive/
9242256/). A similar example relates to police authorities being able to detect a crime because it is all ‘caught on tape’, for
example with sensor-activated security cameras and, potentially, Google Glass or other camera-based devices that are now
routinely carried by many.
All these examples of companies, government agencies and private individuals using digitized technologies to increase
security come at some costs in terms of individuals’ privacy. In terms of locating a lost smartphone, it has to be the user
who, deliberately, accepts giving up her/his (right of) privacy by activating the ‘find my phone’ option (https://itunes.ap-
ple.com/us/app/find-my-iphone/id376101648?mt=8). The example of Google Glass or digital cameras worn, for example,
by cyclists or skiers to record their journey, is more complex since the privacy that a user gives up affects others’ privacy,
thus representing a shift from the individual to the societal level. In some circumstances one’s use of social software appli-
cations affects others’ privacy, as for example, for people who are tagged in somebody’s Facebook profile without them
knowing. Perhaps not surprisingly, privacy advocates have argued that in these types of exchanges consumers are justified
in expecting that the data they collect and share should remain private among those to whom it was originally disclosed –
dependent on users’ risk perceptions, as noted by Gerlach et al. (2015) – rather than being shared with third parties who may
subsequently behave opportunistically (Beldad et al., 2011; Petronio, 2002).
Thus, it is clear that improving security across society, based on digital devices, can impact on individual’s privacy. Some
companies are doing something about this. For instance, Facebook no longer allows a user’s ‘wild tagging’ and, instead, an
automatic email is sent to a user who is tagged, for approval (or at least this is a configurable option under privacy settings).
Yet, the exponential diffusion of tracking software embedded in social networks such as Facebook and the sensors and cam-
eras in many other digital devices lead us to think that it will be hard for organizations (or governments) to regulate how
individuals use responsibly technologies that enable tracking (i.e., in a way that balances security and privacy). The societal
issue is raised because the move towards using devices and applications to gain increased security comes at the expense of
reduced privacy. This points to a question about whether users (and more broadly society) want to give up some security
potential to ensure more privacy (Culnan and Williams, 2009; Velasquez, 2006). This is a decision that citizens need to
debate with their politicians (Dinev et al., 2008) and that governments in turn need to debate with businesses, since it is
businesses that collect and analyze digital traces. This is exemplified by the Lee Rigby case (the British soldier killed on a
street in London), where Facebook was accused of not helping to protect security because it did not use its analytical capa-
bility to detect and report the fact that the killer was posting that he was intending to commit just such a murder (http://
www.theguardian.com/uk-news/live/2014/nov/25/lee-rigby-woolwich-inquiry-report-published-live-coverage).
4S. Newell, M. Marabelli/ Journal of Strategic Information Systems xxx (2015) xxx–xxx
Please cite this article in press as: Newell, S., Marabelli, M. Strategic opportunities (and challenges) of algorithmic decision-making: A call
for action on the long-term societal effects of ‘datification’. J. Strateg. Inform. Syst. (2015), http://dx.doi.org/10.1016/j.jsis.2015.02.001
Other privacy/security tensions are reflected in the current debate on whether US police officers should wear cameras
following recent cases involving police officers’ improper use of force (see for instance the developments over the Michael
Brown case http://www.cnn.com/2014/12/04/us/eric-garner-ferguson-body-cameras-debate/). Here, a sensor technology
would be employed but would not actually generate data that will be processed by algorithms, since the camera records
would be reviewed only in particular circumstances. However, this and other types of sensor are pervasive (Lyytinen and
Yoo, 2002) (and invasive), and the data (e.g., the camera records) would be stored. In such circumstances, we do not know
whether in the future somebody will develop an algorithmic-based decision system to analyze the data (e.g., to assess the
performance of police officers). It is thus clear that the widespread diffusion of digitized technologies can be harmful to indi-
viduals’ privacy while justified by corporations and governments in the name of public security – a tradeoff widely discussed
by Lyon (2003, p. 79) in terms of ID cards that are supposed to improve national security in that he raises the issue of
whether as citizens we are willing to ‘‘pay the price in liberty for security’’. This tradeoff, then, raises complex social issues
because of the ready availability of these data and because of the capacity of algorithms to discriminate almost in real time –
for instance, to determine that particular categories of people (based on race, income, job, etc.) are more likely to commit a
crime, and could, therefore, be subjected to higher levels of policing and potentially also face discrimination in other areas
(Lyon, ibid). This, therefore, pits an individual’s privacy against the security of society, but also suggests broader social issues
in terms of freedom versus control, as we articulate next.
Freedom versus control
The ‘datification’ of everything means that we can use devices to constantly track every decision made and place that a
person visits (be they an employee, a citizen, or our child), and use these data to monitor and control (some now prefer to use
the term ‘nudge’) behavior (Whitman, 2011). This second tradeoff between freedom and control is more complex than the
previous one because, here, individuals can be aware that they are being controlled. This is informed control (e.g., because
they are required to carry RFID badges at the work place or maybe even have chips implanted under their skin, another
example of invasive technology – http://www.bbc.co.uk/news/technology-31042477 – or because they decide to use an elec-
tronic collection system in their car to drive through tolled roads and bridges). However, they can also be controlled without
knowing that they are being monitored. This is uninformed control. Uninformed control happens, for instance, when tracking
cookies monitor someone’s online activity, or, more generally, when ‘second hand’ use of data originating from digitized
technologies are used.
Freedom versus informed control
Surveillance based on parents tracking their children’s every move (e.g., using an application on the child’s smart-
phone) is clearly informed control and would allow parents to feel that they are in control of their children’s movements.
However, the loss of freedom (and privacy, as we have already pointed out) of those subjected to this surveillance might
have far-reaching effects, for instance in terms of children’s feelings of personal responsibility. After all, we know that
punishment is not always an effective deterrent because, once the punishment is removed, the individual often resorts
to the prior (undesirable) behavior; so, if individuals conform only because they know they are being monitored, will their
behavior change once the monitoring ceases? With constant surveillance, like punishment, while we may change behav-
ior, the beliefs about what is appropriate or inappropriate may remain (Podsakoff et al., 1982; Staples, 2013). This tension,
then, is between improved control (by business but also government or private citizens) at the expense of individuals feel-
ing that they have some freedom and autonomy – a feeling that we know has a significant influence on motivation in the
long-term (Hasan and Subhani, 2011). One such example is Hitachi’s new digital identification badge that collects data on
individual employees’ exact location within an office, records who the person has spoken to, for how long and how
energetically (http://www.cnn.com/2014/02/02/opinion/greene-corporate-surveillance). Adopting this kind of technology
as a strategy for constant monitoring may, however, affect employees’ motivation and perhaps also their capacity to pro-
duce disruptive innovation. Indeed, productivity might benefit (at least in the short term) from such an aggressive
approach to control in the workplace. However, the longer-term consequences of such constant surveillance may be more
problematic. For instance, Lyon (2003, p. 20) points out that a ‘‘surveillance system obtains personal and group data in
order to classify people and populations according to varying criteria, to determine who should be targeted for special
treatment, suspicion, eligibility, inclusion, access, and so on’’, arguing that such ‘‘social sorting’’ leads to long-term dis-
criminations. He states that ‘‘data about transactions is used both to target persons for further advertising and to dismiss
consumers who are of little value to companies’’ (ibid,; 1), leading to long-term social differences. Moreover, breakthrough
innovation, which is more risky and leads to more frequent ‘failures’ (O’Reilly and Tushman, 2004), might be jeopardized
because individuals who are constantly monitored are less likely to expose themselves to failure in front of their peers and
superiors. This suggests that those making strategic decisions about how to use this new tracking technology (whether
business, government or private individual) might want to think about reducing the amount of surveillance on employees,
customers, family members or citizens since this would be the price they might want to pay for allowing people to feel in
control of the decisions they make – in other words, being informated and not automated – to use the language of Zuboff
(1984). This supports our argument that a tradeoff emerges between control and freedom in the context of the digitization
of our everyday lives.
S. Newell, M. Marabelli/ Journal of Strategic Information Systems xxx (2015) xxx–xxx 5
Please cite this article in press as: Newell, S., Marabelli, M. Strategic opportunities (and challenges) of algorithmic decision-making: A call
for action on the long-term societal effects of ‘datification’. J. Strateg. Inform. Syst. (2015), http://dx.doi.org/10.1016/j.jsis.2015.02.001
Freedom versus uninformed control
The feeling of being controlled, as we have discussed, might lead to some unwanted consequences (e.g., loss of a sense of
responsibility or lower productivity in the work place). However, probably a more relevant societal issue emerges when con-
trol (over an individual’s freedom) is made without the individual even knowing that she/he is being controlled (when this is
not made explicit or is not requested). To this end, here we provide an example involving individuals’ online activities, where
the ‘free’ access to information is increasingly controlled as Internet companies (social networks, news, etc.) now determine
(based on algorithms) what we see. For instance, we may see many posts about the newest iPhone (6, at the time of our writ-
ing) and assume that many of our Facebook friends are posting articles about this new technology. However, the frequency
with which we see these posts may be partially due to us having clicked on an advertisement related to the iPhone 6: Face-
book’s algorithm decides that we are interested in such products and then shows us others’ posts that are related to the
iPhone 6. A consequence of such use of algorithms by corporations to decide – for the consumer – the posts, news or adver-
tising that they are exposed to, is that it may lead to a slow and often subtle manipulation of consumers’ worldviews as well
as to new forms of discrimination. Simply put, what is presented to the reader is decided by an algorithm – tapping into prior
searches – and is not based on an explicit personal choice. An example of uninformed control by a corporation that produces
worrisome societal issues is found in the account presented by Eli Pariser, who showed that ‘‘Facebook was looking at which
links I clicked on, and it was noticing that I was clicking more on my liberal friends’ links than on my conservative friends’
links. And without consulting me about it, it had edited them out. They disappeared.’’ (Pariser, 2011). In the longer term, this
manipulation by corporations of what the consuming public is exposed to – exposing us only to things that we like (or the
things that an algorithm assumes we like) – may produce societal changes. For instance, our exposure to online diversity will
be reduced, as in the example of Eli Pariser. More recently, Greg Marra, a Facebook engineer argued that, ‘‘We think that of all
the stuff you’ve connected yourself to, this is the stuff you’d be most interested in reading’’, explaining further that an algo-
rithm monitors ‘thousands and thousands’ of metrics to decide what we should see on our Facebook page. These metrics
include what device we use, how many comments or ‘Likes’ a story has received and how long readers spend on each arti-
cle/post. The assumed goal, as a New York Times article suggests, is that companies are using algorithmic decision-making
‘‘to identify what users most enjoy’’ (http://www.nytimes.com/2014/10/27/business/media/how-facebook-is-changing-the-
way-its-users-consume-journalism.html?_r=0). However, this also indicates that this practice of showing us only things that
‘fit’ with our (little) data profile, limits our possibility to choose, and might inhibit our capacity to make informed decisions
(on what we buy and even what we think).
These strategies, then, that are adopted by organizations to allow them to tailor results and personalize offerings to
individual consumers are leading to citizens (i.e., all us of who ‘surf the web’) being exposed to less and less diversity online.
A potential consequence is that we may become less tolerant to diversity, meaning that we may as a result become less able
to listen to someone who thinks differently (e.g., a Republican, in Pariser’s example). Moreover, there may be other, more
worrying consequences in the long-term that are associated with race-diversity intolerance and the increased exploitation
of the vulnerable. For example, in relation to the latter issue, if algorithms work out who is less capable of making good
financial decisions, personalized advertisements can then be sent persuading these people to take out risky loans, or
high-rate instant credit options, thereby exploiting their vulnerability. The strategic use of our own data by corporations
to personalize our Internet, in other words, is just another and potentially more pernicious way of allowing discrimination;
pernicious because the only person who has access to the outcomes of the discrimination is the individual being discriminat-
ed against (who is often not aware of the fact that they are exposed to discriminatory information – uninformed control),
making it easy for unscrupulous businesses to use personalization in a way that harms the vulnerable.
Another way to illustrate how societal concerns emerge as a consequence of businesses (and governments) using data
from the digitization of our everyday life is by articulating the tradeoff between independence and dependence, to which
we now turn.
Independence versus dependence
Citizens in many countries increasingly depend on digital devices for many activities. However, here, a tradeoff originates
from the tension between the willingness to depend on IT devices and being able to live without them (i.e., to be indepen-
dent), should the need arise. Think of our decreasing sense of direction due to our dependency on GPS systems or, were we to
consider safety issues, think of those sensor-based systems that are able to park our car – or even drive it! These driving
systems use onboard cameras and laser rangefinders to identify obstacles (or hazards, if the onboard computer controls
the car while it is ‘in motion’); then an algorithm is able to scan the surrounding environment and to identify safe zones,
avoiding for example other cars (see for instance a 2012 MIT study on these algorithms – http://newsoffice.mit.edu/2012/
mechanical-engineers-develop-intelligent-car-co-pilot-0713). In the case of car autopilots, algorithmic decision-making
has a twofold role: first, data on ‘real drivers’ are collected and analyzed so that the algorithm can make the appropriate deci-
sions (e.g., reacting as a driver would, but with the difference that the algorithm is never tired or inattentive, thus carrying a
safety advantage with respect to humans). Second, sensors embedded in cars (laser rangefinders, in this example) collect
environmental data that are analyzed in real time, so the algorithm has the ability to either assist the driver by supporting
his/her decisions (with warnings to the driver) or to make decision on its own – when the car is in full autopilot mode. Dri-
vers, thus, are somewhat ‘tempted’ to benefit from the comforts companies now design into their products using digital tech-
nology, but this necessarily takes place at the expense of our future independence. In fact, if our car computer emits a
6S. Newell, M. Marabelli/ Journal of Strategic Information Systems xxx (2015) xxx–xxx
Please cite this article in press as: Newell, S., Marabelli, M. Strategic opportunities (and challenges) of algorithmic decision-making: A call
for action on the long-term societal effects of ‘datification’. J. Strateg. Inform. Syst. (2015), http://dx.doi.org/10.1016/j.jsis.2015.02.001
warning signal while we drive on a highway, suggesting that we should slow down, we might argue that it is because the GPS
embedded in the car computer has just received a ‘traffic warning’ or, because the weather channel is broadcasting heavy
rain in minutes, or because we are about to drive through a road work area, but we do not really know the actual reason
of the warning, yet we slow down – this (again) illustrating that algorithmic decision-making incorporates advantages (in
this context, for users) but at the same time precludes a full understanding of why some decisions are being made. This limits
learning through practice (Brown and Duguid, 1991) that in the long term might modify an individual’s ability to learn new
tasks and, more generally, adapt to the workplace or to society more generally (Dall’Alba and Sandberg, 2010; Nicolini et al.,
2003).
While it is certain that there are good reasons for companies designing and for users adopting these automated sys-
tems, as we saw, this might also lead to a change in our ability to undertake particular activities without sensors, and
learn. In the example of the autopilot, once our car parks itself, will we forget how to park on our own? IT-assisted sys-
tems have been around for a while in commercial planes, but pilots are constantly trained on how to pilot a plane in case
the autopilot stops working. However, would individuals be trained on how to drive a car once such ‘autopilot’ systems
become common in private motor vehicles? This example brings to the fore the point that digital technologies and devices
(and the associated algorithmic decision-making) are increasingly influencing and even managing our lives, leaving unan-
swered the question on whether these algorithms are just supporting our activities, or whether they are actually in charge
(e.g., controlling what we do) – and if they are taking over, does this excess of control occur at the expense of our ability
to improvise and respond to emergencies? Thus, in the car example, it is clear that issues associated with safety emerge.
In fact, as car drivers who now rely on sensors, we do not have the luxury that airline pilots have, of simulators to ensure
that we maintain our skills so that we are prepared for an emergency. Nevertheless, even pilots (who, unlike private citi-
zens, are trained on how to operate aircrafts) are not free of the consequences from technology that ‘takes over’, as the US
NTSB (National Transportation Safety Board) reports in relation to some major plane accidents (see for instance the case of
Air France Flight 447 in 2009, http://spectrum.ieee.org/riskfactor/aerospace/aviation/air-france-flight-447-crash-caused-
by-a-combination-of-factors).
The negative consequences of an excess of IT dependence are associated with the risks we are exposed to when we
forget how to do certain things. Progress necessarily involves automation (and the loss of certain manual capacities),
and many innovations developed by corporations positively contribute to our quality of life (and to our safety). However,
it is digital devices and the associated algorithmic decision-making that pose issues, especially when supervising or under-
taking human activities that might involve life-threatening outcomes were the technology to stop working. Moreover,
because of the connectivity between sensor devices, there is also the potential of chaos occurring if everything stops work-
ing for everyone simultaneously. In particular, we argue, it is the diffusion of such IT automations among common citizens
that creates threats were we to become fully dependent on the technology and unable to operate without it. However, the
adoption of some of these automations is (or will become) virtually mandatory for many – creating discriminations again-
st those who do not conform. One simple example relates to US residents who, if desiring to use cars equipped with a
standard stick shift (instead of an automatic), will have to pay more, just because ‘standard’ is not a standard in the
US. On the other hand, those who can only drive automatic cars will have to pay more if they want to rent a car when
they travel overseas, because most cars will have a standard shift and there will be a premium for an automatic car. This
point raises the issue of the role of business in promoting such (automated) digitized technologies: does business have a
responsibility for thinking about such consequences and building in opportunities for learning to reduce our over-
dependence?
In sum, we argue that this tradeoff is affected by the willingness of users to give up some of the comforts that come from
dependence on IT, in the interests of preserving their ability to cope when the IT does not work as expected. Yet, digital tech-
nologies are extremely tempting, and now widely adopted. For instance past research on mobile technologies has already
shed light on users’ needs to be ‘always on’, with the consequence that a feeling of ‘dependency’ arises (Jarvenpaa and
Lang, 2005). However, here we go beyond the psychological feeling of dependency and point to the users’ need to be some-
what assisted (if not led or managed) by digital technology (that involves algorithmic decision-making) – with discrimina-
tion being the consequence of not conforming to this dependence. Companies too need to include sensor-based technologies
in their products and services to remain competitive. For example, a logistics company that does not use GPS-equipment to
determine best routing opportunities would experience difficulties in finding partners to develop a supply chain – being thus
discriminated against (again, for not conforming). However, we suggest that companies might also usefully start to think
about how they can and should, in some circumstances at least, support the development of the ability to cope with situa-
tions of technology failure, with the consequence that algorithms assist decision-makers but do not entirely take over from
human judgment. In our view, a balance must be struck, which to date seems to favor increasing dependence on IT over
being able to cope in the face of IT failure.
We have thus far identified three key tradeoffs: between privacy and security, control and freedom, and dependence and
independence, which are obviously inter-related. We do not claim that these are the only tensions that are relevant; how-
ever, they do allow us to provide concrete examples of strategic opportunities for businesses as well as societal issues emerg-
ing from corporations’ (and governments’) exploitation of data coming from the widespread use of digital technologies that
support – and impact – our everyday lives. In the next section we discuss the more general social issues arising from these
tensions.
S. Newell, M. Marabelli/ Journal of Strategic Information Systems xxx (2015) xxx–xxx 7
Please cite this article in press as: Newell, S., Marabelli, M. Strategic opportunities (and challenges) of algorithmic decision-making: A call
for action on the long-term societal effects of ‘datification’. J. Strateg. Inform. Syst. (2015), http://dx.doi.org/10.1016/j.jsis.2015.02.001
Social consequences of digital technology and algorithmic decision-making
While in the past knowledge and learning have been recognized as path-dependent (Cohen and Levinthal, 1990; Zahra
and George, 2002), in this era of widespread diffusion of digital technologies that capture our everyday activities, our aware-
ness about things appears to be not so much path-dependent as determined by our past actions and algorithmic rules. For
example, the algorithm EdgeRank is used by Facebook to weight ‘likes’ and modify an individual’s Facebook page as a result,
therefore manipulating the ‘wisdom of the crowd’ (Kittur et al., 2007). While this may make sense from a marketing perspec-
tive for businesses (e.g., it is helpful to identify a customer’s interests), it poses concerns for society because of the potential
broader and longer-term social impacts. More specifically, our examples for each tension suggest that businesses (and at
times governments and private individuals) are generally in favor of a more secure society over an individual’s privacy, of
a more controlled population (employees, customers and citizens) over individual freedom – leaving more and more people
increasingly dependent upon technology, at the expense of personal independence.
To consider these issues, we start from the premise that digital trace data is here to stay; companies will increasingly
include tracking software and sensors in the products and the services they offer, and so collect masses of data on our every-
day habits with a view to using these data to develop algorithms that drive decision-making. In fact, whether data are gath-
ered from social networks, an ATM transaction, or from a sensor-based device, there are many aspects associated with
companies using such data that many users want, hence it is unlikely to ‘go away’. As a result, businesses will keep exploiting
big and little data potentials to profile customers, please social network users, and grant (commercial) opportunities to those
who, at best, accept being controlled, reducing their need to learn while giving up some privacy. On the one hand, individuals
benefit from corporations’ use of big/little data analytics – one can save some money on an insurance policy, access a free
show if willing to watch commercials, or just be pleased to see that everybody thinks her/his way (see the Facebook experi-
ment by Pariser, above). On the other hand, businesses are aware that improving their knowledge about employees and cus-
tomers will lead to more control of employees and more (addressed and effective) sales, and therefore more profits. And it is
this enticement by the business world that leads people to assume that they have to give up some privacy/freedom/indepen-
dence, whether this is because it is a way to access a line of credit to buy a house or because they want to use social networks
to fulfill their social needs (Kane, 2014).
As we previously pointed out when we provided our definition of algorithmic decision-making, this might lead to very
superficial understandings of why things happen, and this will definitely not help managers, as well as ‘end users’ build
cumulative knowledge on phenomena. Since decisions are made following an algorithm, how the algorithm came up with
a particular result is unknown; as a result, there will be very little opportunity to learn from mistakes. Ironically, therefore,
decision-makers might be losing the capacity to make decisions on their own, thereby making them a good example of (over)
dependence on digital technology and algorithmic decision-making (cf. our third tradeoff). However, perhaps more impor-
tant (from a societal perspective) than the lack of lessons learned, is the need to discuss the creation of new forms of dis-
crimination as a result of algorithmic decision-making and the associated personalization of information. Think again of
when algorithms determine that particular categories of people (e.g., based on race, income, job) are more likely to commit
a crime and, as a result, those concerned find difficulty in obtaining a loan or changing job, never mind being subjected to
tighter police scrutiny. This clearly violates basic privacy rights, but is justified based on the idea that it will increase security
in society. Or, think again of the control exercised by algorithms in sensor-equipped cars on teenagers: these data are used by
insurance companies to decide whether a driver is good or bad, again on the basis of an algorithm (the tradeoff between
control and freedom). Similarly, when we give new technologies the possibility to take over our learning and let our car park
and drive for us (or let our car ‘suggest’ what we should do in order to perform a perfect parallel park), our decision-making
is being driven by algorithms (the tradeoff between independence and dependence).
These tradeoffs operate together rather than independently. For instance, if we use an app. that ‘knows’ what music we
like so that, when we start driving, we do not need to search for a particular song, this is because we have enabled function-
ality on our radio/phone that is able to ‘see’ our favorite playlists, or that looks into our past purchases. For instance, iTunes
Genius works with a ‘secret algorithm’ created by Apple that compares our library of tracks to all other Genius users’ libraries
and considers complex ‘weight factors’ to then come up with the appropriate playlist for a specific user (Mims, 2010). Here,
we do not aim to go into technical details on how these algorithms work – as Apple engineer Erik Goldman said, the algo-
rithm is ‘secret’, jokingly noting that ‘if he told you how Genius works, he’d have to kill you’ (Mims, ibid.) – to highlight the
relevance and commercial value of these algorithms. However, this example reflects how, in some circumstance we are lit-
erally at the mercy of an algorithm, which makes a decision for us. What if we look for vegetarian food just because we go out
for dinner with friends who happen to be vegetarian? Does this mean that, due to the connections between databases of
large companies (or because of tracking cookies), we will be denied the opportunity of seeing advertisements for steakhous-
es on our Facebook webpage? Or will we be classified as good drivers because a sensor detects that most of the time we obey
speed limits (even if the reason is that we know that where we drive the speed limits are strictly enforced)? Or will our pref-
erences in terms of the music that we listen to be so reinforced by the automatic selections based on the algorithm that we
reduce our exposure to alternative genres?
It is clear that the three tradeoffs showcase interests and needs of individuals on the one hand, and the somewhat oppor-
tunistic strategic moves of businesses (and governments) on the other. Moreover, our discussion illustrates the relevant role
of algorithms in making decisions about an individual’s characteristics/preferences based on trends, and therefore about
8S. Newell, M. Marabelli/ Journal of Strategic Information Systems xxx (2015) xxx–xxx
Please cite this article in press as: Newell, S., Marabelli, M. Strategic opportunities (and challenges) of algorithmic decision-making: A call
for action on the long-term societal effects of ‘datification’. J. Strateg. Inform. Syst. (2015), http://dx.doi.org/10.1016/j.jsis.2015.02.001
what individuals should see and are likely to buy. Eric Schmidt (Google) said in 2010 that, ‘‘It will be very hard for people to
watch or consume something that has not in some sense been tailored for them’’. This statement involves privacy issues
(businesses will know almost everything about consumers), control issues (consumers are literally monitored and then con-
trolled with choices made for them), and dependence issues (loss of independence in making informed decisions, since the
information provided about a specific choice will be driven by online feeds – created by algorithms). We posit that IS
research is needed to examine the social issues that are emerging in relation to the strategic uses made by corporations
of data from the digitization of our everyday lives. With the intention to provide an overall picture of the ethical issues
and challenges created by this increasing trend, above we present a framework (Fig. 1) that illustrates how digital technology
(first layer) generates tradeoffs (second layer), when this technology is combined with algorithmic decision-making (third
layer), leading to tensions (fourth layer). This summary framework has the purpose of showcasing strategic opportunities
as well as societal challenges in an era of widespread diffusion of digital technology, and of supporting further interdisci-
plinary research on this topic, along with our suggested new avenues of research and potential research questions (in the
last section that follows).
Research agenda and concluding remarks
We do not know for sure the extent to which digital technology and the associated big/little data analytics are going to
impact society in the long term. However, we suggest that individuals seem to be likely to accept the ‘dark side’ of datifica-
tion through digital traces (always there), and constant monitoring through sensors because they are persuaded that the
benefits outweigh the costs. Thus, businesses (and governments) try to send to citizens the message that security is more
important than privacy (to fight terrorism, for instance). And the same businesses make us believe that if we want to quickly
find what we are looking for (whether it is a movie that we like, through Netflix, or a specific piece of information, through
Google) we need the support of algorithms, that ‘know’ us and what we want – precluding our exposure to diversity. And
finally, businesses develop digital technologies that ‘help’ us do new things more quickly, but simultaneously make us more
reliant on (and so more vulnerable to) these same technologies as well as reducing our ability to learn.
Therefore we suggest that research should be carried out that considers broad social issues associated with businesses’
(and government’s) strategic use of data, especially so because we currently have very little understanding of what the con-
sequences of corporations’ non-responsible use of these data will be for society (Tene and Polonetsky, 2013), albeit we have
suggested some negative impacts above. One way of looking at these social issues may be using an ethical dilemma lens,
where we consider individuals’ right to maintain their privacy, freedom and independence, against businesses’ right to dis-
criminate to promote sales – using cutting-edge technology such as big data analytics. We suggest that such dilemmas can
be addressed using the teleological or deontological approaches to ethics (or both). The deontological approach (utilitarian-
ism) is the best-known consequentialist theory (Bentham, 1776; Mill, 1863), and suggests that ethical behavior is one that
maximizes societal welfare while minimizing social harm (Vallentyne, 1987; Berente et al., 2011). According to this
approach, insurance companies may be right in being algorithm driven – thus applying higher premiums to those who,
according to the data analytics, are more at risk of having car accidents. However, the (contrasting) deontological approach
Fig. 1. A summary framework of the consequences of an algorithm-based world.
S. Newell, M. Marabelli/ Journal of Strategic Information Systems xxx (2015) xxx–xxx 9
Please cite this article in press as: Newell, S., Marabelli, M. Strategic opportunities (and challenges) of algorithmic decision-making: A call
for action on the long-term societal effects of ‘datification’. J. Strateg. Inform. Syst. (2015), http://dx.doi.org/10.1016/j.jsis.2015.02.001
bases ethical decisions on broad, universal ethical principals and moral values such as honesty, promise keeping, fairness,
loyalty and rights (e.g. to safety, privacy) so that the process or means by which an individual does something, rather than
the outcome, is the focus of decision-making (e.g. lying is dishonest as it is one’s duty to be honest regardless of whether this
might lead to some ultimate good), therefore the end never justifies the means (Mingers and Walsham, 2010). According to
this latter approach, discriminations should not take place if a minority is adversely and unfairly treated – never mind if
following an algorithm maximizes positive consequences for society. More specifically, here we want to identify research
questions that examine the social issues related to each of our tradeoffs, as described next.
Firstly, in terms of the tradeoff between privacy and security, one aspect that deserves attention relates to how far dif-
ferent countries (and regulators) will balance this tradeoff. From an ethical perspective, for instance, they might choose
to privilege the maximization of societal welfare (so taking a teleological approach) or to pay attention to ‘minorities’,
who are penalized by the discriminations of algorithmic decision-making (deontological approach). Information privacy
laws and regulations – how citizens perceive (the value of) privacy – are country-specific and are related to cultural and his-
torical issues (Milberg et al., 1995). One example is the recent debate about the ‘right to be forgotten’ (http://ec.europa.eu/
justice/data-protection/files/factsheets/factsheet_data_protection_en.pdf), which forced Google to delete some information
(and to implement processes to do so in the future, should people ask) from the results of its search engine in Europe, while
this issue is not perceived currently as a relevant one in other countries such as in the US. To this end, it would be interesting
to dig deeper into research questions such as: ‘How far do institutions and governments influence the balance between pri-
vacy and security associated with digital technologies that collect data on our everyday lives?’ ‘What are the historical, cul-
tural and social reasons behind the variety of approaches to digital privacy adopted by different countries?’ ‘Do these
different approaches reflect differences in citizens’ ethical and moral values?’ ‘Do (or will) social networks have the ability,
in the long term, to modify ethical and moral values about privacy in different countries?’ ‘Will the diffusion of digital
technology (and the IoT) lead to the standardization of ethical and moral values across countries, in the long-term?’
In terms of the tradeoff between freedom and control, we know very little about how far users are aware that they are
being controlled by large Internet companies (especially if we think of ‘second hand’ data), and if they are, it would be
interesting to learn about whether individuals’ needs to enact social networks (Kane, 2014) prevail over the potentially
uncomfortable feeling of being profiled (little data). Moreover, we do not have specific quantitative data that illustrates
the effectiveness of algorithmic decision-making in identifying people’s needs – for instance we know that little data has
the ability to ‘know’ (or assume) what people want to purchase on the basis of a number of digital traces, but little is know
about the actual revenues that derive from this – and whether the costs of implementing ‘smart’ algorithms and maintaining
expensive hardware that can process big data is covered by the increased sales. We should assume that businesses achieve
positive bottom lines from big data, since datification and algorithmic decision making is widely adopted and is expensive
(http://www.forbes.com/sites/ciocentral/2012/04/16/the-big-cost-of-big-data/), but we do not know, in concrete terms, the
extent to which this has improved sales, customer satisfaction, inventory management or other financial, operational and
organizational parameters. Knowing this would perhaps indicate a price for a loss of individuals’ privacy and freedom. After
all, one of the commonly cited examples of the usefulness of algorithmic decision-making was of Google being able to predict
the location of a US flu epidemic, based on searches for flu remedies, faster than the Center for Disease Control (CDC). Yet, the
story often remains untold, that they have been unable to repeat this success (http://www.theguardian.com/technology/
2014/mar/27/google-flu-trends-predicting-flu). Thus, it is important that we conduct research that looks at the benefits
for citizens of having a ‘tailored’ Internet, as against the costs of the benefits of living in an ‘Internet bubble’. And finally there
is a question about, ‘what ethical guidelines might businesses usefully adopt to manage big/little data and produce the algo-
rithms from these data’? For instance, Facebook has indicated that they have developed ethical policies for those who design
algorithms, but such policies are not disclosed to the public. It is important that we research these issues so that we under-
stand the ways in which businesses are using algorithms for discriminating so that we can enter a debate with business
about associated ethical concerns (much as, for example, was the case in relation to the use of child labor in the past). Con-
sider, for example, a monitoring system that profiles utility customers and sets different prices for gas and electricity, based
on geographical areas and demand (another example of uninformed control). In this instance, maximizing societal welfare
(cheap electricity for the majority) at the expense of minorities may well be unacceptable from an ethical standpoint (since
those who end up paying more, are likely ironically also to be the very people who may be the most vulnerable and least able
to pay). As a start in this process, we need research that sheds better light on the overall awareness of individuals in terms of
how their data are being used by businesses and whether people are happy with this, especially as this exploits the more
vulnerable in society. Thus, while big data analytics has the potential to shed light on important human and societal issues
(Markus, 2015), this should not happen at the expense of the vulnerable.
In terms of the tradeoff between independency and dependency, we think that major societal issues are associated with
the lack of opportunities for individuals to learn – and this poses issues from a knowledge creation and sharing perspective.
As we pointed out earlier in this article, knowledge develops cumulatively and, according to the practice perspective
(Feldman and Orlikowski, 2011; Schatzki et al., 2001; Sandberg and Tsoukas, 2011; Whittington, 2014) knowledge equates
with practice. However, in the context of this tradeoff, it is the algorithm that gains knowledge about the minutiae of indi-
viduals – for instance, analyzing how humans drive a car, so that it can then operate as such, while humans may not gain a
better understanding from this process. This poses relevant issues that involve both the private and work life of individuals.
For example, ‘will individuals lose their capacity to learn (even from mistakes)?’ ‘Will IT-assisted systems reach a point that
they will impair an individual’s problem solving skills and abilities in her/his everyday life?’ This issue can be taken further if
10 S. Newell, M. Marabelli/ Journal of Strategic Information Systems xxx (2015) xxx–xxx
Please cite this article in press as: Newell, S., Marabelli, M. Strategic opportunities (and challenges) of algorithmic decision-making: A call
for action on the long-term societal effects of ‘datification’. J. Strateg. Inform. Syst. (2015), http://dx.doi.org/10.1016/j.jsis.2015.02.001
we refer to the potential decreased ability of managers to make decisions on their own, due to the few opportunities to ‘prac-
tice’ decision-making processes. For instance, ‘Will an individual’s capacity to adapt to a new organization and a new job be
compromised by increased control (made by algorithms, and that leads to living in a ‘bubble’, see our previous discussion of
Pariser’s ‘experiment’), which makes people less likely to be flexible and accepting towards diversity?’ Also, there is the issue
of the legitimacy of decisions that are based on algorithms and whether they will be accepted by those affected by the deci-
sion when the reasons for the decision are not actually known, even by those developing the algorithms. Thus, ‘will business-
es face more claims of unfair discrimination in the future when people identify that they have been treated differently to
others but when the justification for this is not understood and cannot be clearly articulated by anyone?’ ‘‘The computer says
‘no’’’ (e.g., https://www.youtube.com/watch?v=AJQ3TM-p2QI), may come to be an unacceptable justification for being
discriminated against as we are increasingly confronted by this ‘rationale’.
The examination of the above social issues demands a multi-disciplinary approach that considers economic, legal,
organizational, ethical, cultural and psychological consequences ofthedigitizationofoureverydaylivesfordifferentpopulations.
In examining these issues, we would do well to remember that the ways new computing technologies (and the associated
data) are used is not neutral in terms of the consequences for the human actors who leave digital traces that are then
collected and analyzed. Corporations (and governments) have choices about how and what digital traces they collect and
measure, and about the algorithms that they develop to make decisions based on this measurement, even if these decisions
are increasingly distributed throughout a corporation rather than in the hands of the CIO (Nolan, 2012). These choices raise
fundamental social questions as we have seen. As researchers we have an opportunity – and a responsibility (cf. Desouza
et al., 2006, 2007) – to expose this empirically and theoretically and so promote an agenda of ‘responsible analytics’ that
attempts to reduce the long-term negative social consequences of this new era concerned with the digitization of society.
In conclusion, this paper is an explicit call for action. We argue that researchers as well as practitioners should take these
issues into serious consideration and articulate an interdisciplinary debate on how the datification of our everyday lives and
the associated algorithmic decision-making (and the IoT) will affect society. This consequent research agenda requires a mul-
ti-disciplinary perspective. The issues are extremely relevant, strategic research topics. Whether we are interested in finding
ways to increase business value or we are concerned with broader social issues of equality and democracy, they require
immediate action. Strategic IS scholars are interested in ‘‘the way IT ‘delivers the goods’ by providing business AND social
benefits’’ (Galliers et al., 2012, emphasis added). We would argue that minimizing social harm, even if to a minority, should
be added to this agenda.
Acknowledgements
We would like to thank Sirkka Jarvenpaa and Yolande Chan for their extremely useful feedback given to previous versions
of this manuscript.
References
Abbas, R., Katina, M., Michael, M.G., 2014. The regulatory considerations and ethical dilemmas of location-based services (LBS): a literature review. Inf.
Technol. People 27 (1), 2–20.
Ball, K., 2002. Elements of surveillance: a new framework and future directions. Inf., Commun., Soc. 5 (4), 573–590.
Ball, K., 2005. Organization, surveillance, and the body: towards a politics of resistance. Organization 12 (1), 89–108.
Belanger, F., Crossler, R.E., 2011. Privacy in the digital age: a review of information privacy research in information systems. MIS Q. 35 (4), 1017–1041.
Beldad, A., de Jong, M., Steehouder, M., 2011. I trust not therefore it must be risky: determinants of the perceived risk of disclosing personal data for e-
government transactions. Comp. Hum. Behav. 27 (6), 2233–2242.
Bentham, J., 1776. A Fragment of Government. London (Preface (2nd para)).
Berente, N., Gal, U., Hansen, S., 2011. Ethical implications of social stratification in information systems research. Inf. Syst. J. 21 (4), 357–382.
Boyd, D., Crawford, K., 2014. Critical questions for big data: provocations for a cultural, technological, and scholarly phenomenon. Inf., Commun. Soc. 15 (5),
662–679.
Brown, J.S., Duguid, P., 1991. Organizational learning and communities of practice: towards a unified view of working, learning, and innovation. Organ. Sci. 2
(1), 40–57.
Campbell, S.W., Park, Y.J., 2008. Social implications of mobile telephony: the rise of personal communication society. Sociol. Compass 2 (2), 371–387.
Castells, M., Fernandez-Ardevol, M., Linchuan Qiu, J., Sey, A., 2009. A cross-cultural analysis of available evidence on the social uses of wireless
communication technology. In: Mobile Communication and Society: A Global Perspective. The MIT Press, Cambridge, MA.
Chan, Y.E., Culnan, M.J., Greenaway, K., Laden, G., Levin, T., 2005. Information privacy: management, marketplace, and legal challenges. Commun. AIS 16,
270–298.
Chen, H., Chiang, R.H.L., Storey, V.C., 2012. Business intelligence and analytics: from big data to big impact. MIS Q. 36 (4), 1165–1188.
Clark, C., Newell, S., 2013. Institutional work and complicit decoupling across the U.S. capital markets: the case of rating agencies. Bus. Ethics Q. 23 (1), 1–30.
Cohen, W.M., Levinthal, D.A., 1990. Absorptive capacity: a new perspective of learning and innovation. Adm. Sci. Q. 35 (1), 128–152.
Coll, S., 2014. Power, knowledge, and the subjects of privacy: understanding privacy as the ally of surveillance 17(10), 1250–1263.
Culnan, M.J., Clark-Williams, C., 2009. How ethics can enhance organizational privacy: lessons from the choicepoint and TJX data breaches. MIS Q. 33 (4),
673–687.
Dall’Alba, G., Sandberg, J., 2010. Learning through practice: professional and practice-based learning, 1, 104–119.
Desouza, K.C., El Sawy, O.A., Galliers, R.D., Loebbecke, C., Watson, R.T., 2006. Beyond rigor and relevance towards responsibility and reverberation:
information systems research that really matters. Commun. Assoc. Inf. Syst. 17 (16).
Desouza, K.C., Ein-Dor, P., McCubbrey, D.J., Galliers, R.D., Myers, M.D., Watson, R.T., 2007. Social activism in information systems research: making the world
a better place. Commun. Assoc. Inf. Syst. 19, 261–277.
Dinev, T., Hart, P., Mullen, M., 2008. Internet privacy concerns and beliefs about government surveillance – an empirical investigation. J. Strateg. Inf. Syst. 17
(3), 214–233.
Doyle, A., Rippert, R., Lyon, D., 2013. The Global Growth of Camera Surveillance. Routledge Publisher.
S. Newell, M. Marabelli/ Journal of Strategic Information Systems xxx (2015) xxx–xxx 11
Please cite this article in press as: Newell, S., Marabelli, M. Strategic opportunities (and challenges) of algorithmic decision-making: A call
for action on the long-term societal effects of ‘datification’. J. Strateg. Inform. Syst. (2015), http://dx.doi.org/10.1016/j.jsis.2015.02.001
Feldman, M.S., Orlikowski, W., 2011. Theorizing practice and practicing theory. Organ. Sci. 22 (4), 1240–1253.
Galliers, R.D., Jarvenpaa, S.L., Chan, Y.E., Lyytinen, K.L., 2012. Strategic information systems: reflections and prospectives. J. Strateg. Inf. Syst. 21 (2), 85–90.
Gartner, 2001. 3D Data Management: Controlling Data Volume, Velocity, and Variety, by D. Laney <http://blogs.gartner.com/doug-laney/files/2012/01/
ad949-3D-Data-Management-Controlling-Data-Volume-Velocity-and-Variety.pdf> (accessed 06.02.15).
Gartner, 2012. The Importance of ‘Big Data’: A Definition, by Beyer M.A., Laney D. <http://www.gartner.com/document/2057415>.
Gerlach, J., Widjaja, T., Buxmann, P., 2015. Handle with care: How online social network providers’ privacy policies impact users’ information sharing
behavior. J. Strat. Inf. Syst. 24(1), http://dx.doi.org/10.1016/j.jsis.2014.09.001 (available online 12.10.14).
Greenaway, K., Chan, Y.E., 2005. Theoretical explanations for firms’ information privacy behaviors. J. AIS 6 (6), 171–198.
Hasan, S., Subhani, M., 2011. Top management’s snooping: is sneaking over employees’ productivity and job commitment a wise approach. Afr. J. Bus.
Manage. 6 (14), 5034–5043.
Hedman, J., Srinivasan, N., Lindgren, R., 2013. Digital traces or information systems: sociomateriality made researchable. In: Proceedings of the 34th ICIS.
Milan, Italy.
Hodgkinson, G.P., Maule, A.J., Brown, N.J., Pearman, A.D., Glaister, K.W., 2002. Further reflections on the elimination of framing bias in strategic decision-
making. Strateg. Manag. J. 23 (11), 1069–1073.
Jarvenpaa, S.L., Lang, K.R., 2005. Managing the paradoxes of mobile technology. Inf. Syst. Manage. 22 (4), 7–23.
Kane, G.C., 2014. Psychological stages of symbolic action in social media. In: Proceedings of the 35th ICIS December 14–17. Auckland, NZ.
Kittur, A., Chi, E., Pendleton, B.A., Suh, B., Mytkowicz, T., 2007. Power of the few vs. wisdom of the crowd: Wikipedia and the rise of the bourgeoises. WWW:
World Wide Web 1 (2), 19–28.
Lyon, D., 2001. Surveillance and Society: Monitoring Everyday Life. McGraw-Hill International – Publisher.
Lyon, D., 2003. Surveillance as Social Sorting: Privacy, Risk, and Digital Discrimination. Routledge, London.
Lyon, D., 2014. Surveillance, snowden, and big data: capacities, consequences, critique. Big Data Soc. 1 (2). http://dx.doi.org/10.1177/2053951714541861.
Lyytinen, K., Yoo, Y., 2002. Issues and challenges in ubiquitous computing. Commun. ACM 45 (12), 62–96.
MacCrory, F., Westerman, G., Alhammadi, Y., Brynjolfsson, E., 2014. Racing with and against the machine: changes in occupational skill composition in an
era of rapid technological advance. In: Proceedings of the International Conference of Information Systems (ICIS). Auckland, NZ.
Markus, M.L., 2015. New games, new rules, new scoreboards: the potential consequences of big data. J. Inf. Technol. http://dx.doi.org/10.1057/jit.2014.28
(available online 20.01.15).
Mayer-Schonberger, V., Cukier, K., 2013. Big Data: A Revolution That Will Transform How We Live, Work, and Think. Houghton Mifflin Harcourt Publishing
Company, New York, NY.
McAfee, A., Brynjolfsson, E., 2012. Big data: the management revolution. Harv. Bus. Rev. 90 (10), 60–68.
Michael, K., Michael, M.G., 2011. The social and behavioral implications of location-based services. J. Loc. Based Serv. 5 (3–4), 121–137.
Milberg, S.J., Burke, S.J., Smith, H.J., Kallman, E.A., 1995. Values, personal information, privacy, and regulatory approaches. Commun. ACM 38 (12), 65–74.
Mill, J.S., 1863. Utilitarianism, reprinted in 1906 in Chicago by the University of Chicago Press (first appearance in Fraser’s Magazine, in 1861, then collected
and reprinted as a single book in 1863).
Mims, C., 2010. How iTunes Genius Really Works, MIT Technology Review <http://www.technologyreview.com/view/419198/how-itunes-genius-really-
works/> (accessed February 02.02.15).
Mingers, J., Walsham, G., 2010. Toward ethical information systems: the contribution of discourse ethics. MIS Q. 34 (4), 833–854.
Miorandi, D., Sicari, S., De Pellegrini, F., Chlamatac, I., 2012. Internet of things: vision, applications, and research challenges. Ad Hoc Netw. 10, 1497–1516.
Munford, M., 2014. Rule changes and big data revolutionise Caterham F1 chances. The Telegraph, Technology Section, 23 February 2014 <http://www.
telegraph.co.uk/technology/technology-topics/10654658/Rule-changes-and-big-data-revolutionise-Caterham-F1-chances.html> (accessed 15.11.14).
Negash, S., 2004. Business intelligence, communication of the association for information systems, 13 (article 15) <http://aisel.aisnet.org/cais/vol13/iss1/
15>.
Newell, S., Marabelli, M., 2014. The crowd and sensors era: opportunities and challenges for individuals, organizations, society, and researchers, in:
Proceedings of 35th ICIS, December 14–17. Auckland, NZ.
Nicolini, D., Gherardi, S., Yanow, D., 2003. Introduction: toward a practice-based view of knowing and learning in organizations. In: Knowing in
Organizations: A Practice-based Approach. Barnes & Noble.
Nolan, R., 2012. Ubiquitous IT: the case of the Boeing 787 and implications for strategy research. J. Strateg. Inf. Syst. 21 (2), 91–102.
O’Reilly, C.A., Tushman, M.L., 2004. The ambidextrous organizations. Harv. Bus. Rev., April, 1–10.
Pariser, E., 2011. Beware Online ‘Filter Bubbles’ TED-2011 <http://www.ted.com/talks/eli_pariser_beware_online_filter_bubbles> (accessed 01.10.14).
Petronio, S., 2002. Boundaries of Privacy: Dialectics of Disclosure. State University of New York Press, Albany, NY.
Podsakoff, P., Todor, W., Skov, R., 1982. Effects of leader contingent and noncontingent reward and punishment behaviors on subordinate performance and
satisfaction. Acad. Manag. J. 25 (4), 810–821.
Power, D.J., 2002. Decisions Support Systems: Concepts and Resources for Managers. Quorum Books, Westport, CT.
Rouibah, K., Ould-ali, S., 2002. Puzzle: a concept and prototype for linking business intelligence to business strategy. J. Strateg. Inf. Syst. 11 (2), 133–152.
Sandberg, J., Tsoukas, H., 2011. Grasping the logic of practice: theorizing through practical rationality. Acad. Manag. Rev. 36 (2), 338–360.
Schatzki, T.R., Knorr-Cetina, K., von Savigny, E., 2001. The Practice Turn in Contemporary Theory. Routledge, London.
Schroeder, R., Cowls, J., 2014. Big data, ethics, and the social implications of knowledge production. GeoJournal <https://dataethics.github.io/proceedings/
BigDataEthicsandtheSocialImplicationsofKnowledgeProduction.pdf> (accessed 24.01.15).
Slade, S., Prinsloo, P., 2013. Learning analytics: ethical issues and dilemmas. Am. Behav. Scient. 57 (10), 1510–1529.
Smith, H.J., 2002. Ethics and information systems: resolving the quandaries. ACM SIGMIS Datab. 33 (3), 8–22.
Smith, H.J., Hasnas, J., 1999. Ethics and information systems: the corporate domain. MIS Q. 23 (1), 109–127.
Smith, H.J., Dinev, T., Xu, H., 2011. Information privacy research: an interdisciplinary review. MIS Q. 35 (4), 989–1016.
Staples, W.G., 2013. Everyday Surveillance: Vigilance and Visibility in the Postmodern Life. Rowman & Littlefield.
Tene, O., Polonetsky, J., 2013. Big data for all: privacy and user control in the age of analytics, 11 Nw. J. Technol. Intellect. Prop., 239 <http://
scholarlycommons.law.northwestern.edu/njtip/vol11/iss5/1> (accessed 02.11.14).
Thomsen, E., 2003. BI’s promised land. Intell. Enterp. 6 (4), 21–25.
Tsoukas, H., Vladimirou, E., 2001. What is organizational knowledge? J. Manage. Stud. 38 (7), 973–993.
Vallentyne, P., 1987. The teleological/deontological distinction. J. Value Inq. 21, 21–32.
Velasquez, M., 2006. Business Ethics: Concepts and Cases. Upper Saddle River, NJ, Pearson.
Whitman, G., 2011. The new paternalism: unraveling ‘nudge’. Econ. Affairs 31, 4–5.
Whittington, R., 2014. Information systems strategy and strategy-as-practice: a joint agenda. J. Strateg. Inf. Syst. 23 (1), 87–91.
Wu, L., Brynjolfsson, E., 2009. The future of prediction: how google searches foreshadow housing prices and quantities. In: Proceedings of 31st ICIS,
December 15–18. Pheonix, AZ, paper 147.
Xi, F., Yang, L.T., Wang, L., Vinel, 2012. Internet of things. Int. J. Commun. Syst. 25, 1101–1102.
Yoo, Y., 2010. Computing in everyday life: a call for research on experiential computing. MIS Q. 34 (2), 213–231.
Zahra, S.A., George, G., 2002. Absorptive capacity: a review, reconceptualization, and extension. Acad. Manage. Rev. 27 (2), 185–203.
Zuboff, S., 1984. In the Age of Smart Machine: The Future of Work and Power. Basic Books, NY.
12 S. Newell, M. Marabelli/ Journal of Strategic Information Systems xxx (2015) xxx–xxx
Please cite this article in press as: Newell, S., Marabelli, M. Strategic opportunities (and challenges) of algorithmic decision-making: A call
for action on the long-term societal effects of ‘datification’. J. Strateg. Inform. Syst. (2015), http://dx.doi.org/10.1016/j.jsis.2015.02.001
... Online labor platforms, such as Uber, increasingly employ algorithmic management (Curchod et al., 2020;Gal et al., 2020;Newlands, 2021;Wiener et al., 2021). By using sophisticated every-second monitoring and surveillance techniques, these platforms collect large amounts of data about their workforce in real-time (Kellogg et al., 2020;Newell & Marabelli, 2015;Zuboff, 2019). Machine learning algorithms then read such data, identify patterns, and guide meaningful decision-making through classifications and predictions (Benbya et al., 2021;Burrell, 2016;de Reuver et al., 2018;Faraj et al., 2018;Gregory et al., 2020;Lee et al., 2018;Pachidi et al., 2021;Schuetz & Venkatesh, 2020). ...
... Workers exposed to algorithmic management often report that they experience tensions in their work environment (Gal et al., 2020;Kellogg et al., 2020;Möhlmann et al., 2021;Page et al., 2017;Tilson et al., 2021;Wiener et al., 2021). For example, while gig workers often experience high levels of autonomy and flexibility (Rosenblat & Stark, 2016), they still feel surveilled and controlled through real-time surveillance (Newell & Marabelli, 2015;Zuboff, 2019). Workers can be penalized and even (temporarily) banned from platforms for showing behavior classified as "undesirable" (Lee et al., 2015;Möhlmann et al., 2021;Rosenblat & Stark, 2016). ...
... Furthermore, some of our findings mirror previous research which widely accepts that platform workers' behavior is influenced by the personalized, algorithmic instructions presented to them. In turn, platforms monitor and track their workers' "every move", feeding this information back into their learning algorithms (Benbya et al., 2021;Faraj et al., 2018;Gregory et al., 2020;Möhlmann et al., 2021;Newell & Marabelli, 2015 ...
Article
Full-text available
Algorithmic management may create work environment tensions that are detrimental to workplace well-being and productivity. One specific type of tension originates from the fact that algorithms often exhibit limited transparency and are perceived as highly opaque, which impedes workers' understanding of their inner workings. While algorithmic transparency may facilitate sensemaking, the algorithm's opaqueness may aggravate sensemaking. By conducting an empirical case study in the context of the Uber platform, we explore how platform workers make sense of the algorithms managing them. Drawing on Weick's enactment theory, we theorize a new form of sensemaking-algorithm sensemaking-and unpack its three sub-elements: (1) focused enactment, (2) selection modes, and (3) retention sources. The sophisticated, multi-step process of algorithm sensemaking allows platform workers to keep up with algorithmic instructions systematically. We add to previous literature by theorizing algorithm sensemaking as a mediator linking workers' perceptions about tensions in their work environment and their behavioral responses.
... However, on the other side, a number of autonomy-undermining AI-enabled phenomena has been also widely shown in the debate on AI and ethics. Such phenomena extend from AI-facilitated risks of individuals' deception, manipulation, and coercion (Susser et al. 2019;Jonjepier and Klenk 2022), especially enabled by the deployment of ML profiling techniques to exploit users' emotion and/or vulnerabilities to meet thirdparty goals above the individual's 2 (Applin and Fischer 2015;Newell and Marabelli 2015;Zarsky 2016;Helberger 2016;O'Neil 2016;Milano et al. 2020), to the narrowing of the range of informational choice options (Royakkers et al. 2018;Tiribelli 2020Tiribelli , 2023 and informational diversity (Newell and Marabelli 2015;Pariser 2011;Sunstein 2008)-both pre-conditions for autonomy (Van den Hoven and Rooksby 2018)-via personalization techniques and the construction of paternalistic algorithmic choice-architectures (Tene and Polonetsky 2013; Tiribelli 2023), up to morally unjustified opportunity losses or exclusion, from the access to a job opportunity or a college admission, up to the preclusion of an healthcare facility or of freedom itself (probation), due to gender and racial bias embedded via data in AI systems (Angwin et al. 2016;Dastin 2018;Obermeyer et al. 2019;Simonite 2020). ...
... However, on the other side, a number of autonomy-undermining AI-enabled phenomena has been also widely shown in the debate on AI and ethics. Such phenomena extend from AI-facilitated risks of individuals' deception, manipulation, and coercion (Susser et al. 2019;Jonjepier and Klenk 2022), especially enabled by the deployment of ML profiling techniques to exploit users' emotion and/or vulnerabilities to meet thirdparty goals above the individual's 2 (Applin and Fischer 2015;Newell and Marabelli 2015;Zarsky 2016;Helberger 2016;O'Neil 2016;Milano et al. 2020), to the narrowing of the range of informational choice options (Royakkers et al. 2018;Tiribelli 2020Tiribelli , 2023 and informational diversity (Newell and Marabelli 2015;Pariser 2011;Sunstein 2008)-both pre-conditions for autonomy (Van den Hoven and Rooksby 2018)-via personalization techniques and the construction of paternalistic algorithmic choice-architectures (Tene and Polonetsky 2013; Tiribelli 2023), up to morally unjustified opportunity losses or exclusion, from the access to a job opportunity or a college admission, up to the preclusion of an healthcare facility or of freedom itself (probation), due to gender and racial bias embedded via data in AI systems (Angwin et al. 2016;Dastin 2018;Obermeyer et al. 2019;Simonite 2020). ...
Article
Full-text available
Many popular artificial intelligence (AI) ethics frameworks center the principle of autonomy as necessary in order to mitigate the harms that might result from the use of AI within society. These harms often disproportionately affect the most marginalized within society. In this paper, we argue that the principle of autonomy, as currently formalized in AI ethics, is itself flawed, as it expresses only a mainstream mainly liberal notion of autonomy as rational self-determination, derived from Western traditional philosophy. In particular, we claim that the adherence to such principle, as currently formalized, does not only fail to address many ways in which people’s autonomy can be violated, but also to grasp a broader range of AI-empowered harms profoundly tied to the legacy of colonization, and which particularly affect the already marginalized and most vulnerable on a global scale. To counter such a phenomenon, we advocate for the need of a relational turn in AI ethics, starting from a relational rethinking of the AI ethics principle of autonomy that we propose by drawing on theories on relational autonomy developed both in moral philosophy and Ubuntu ethics.
... The genesis of an algorithmic AI approach for a social contract has several challenges to overcome based on recent experience. For example, the ongoing criticisms of algorithmic decision-making systems (ADS) in widespread use such as in hiring, lending, judicial and legal decisions, housing, healthcare and education [17]. ...
... Considerations here for the application of big data for algorithmic AI approaches include the need to overcome the bias issues of AI discrimination, security, ethics and colonialism [17,18]. Algorithms applied need to be: (i) safe for all affected by them, (ii) reliable and (iii) available for utility, (iv) achieve an appropriate balance between privacy and security, (v) be explainable (through black-box, white-box or constructivist approaches), (vi) allow transparency and visibility of the social consequences of AI algorithmic decision-making. ...
Article
Full-text available
The potential for artificial intelligence algorithms and game theory concepts to offer prescriptive and decision-making capability for humankind is increasingly recognized. This derives from the increasing availability of granular, multivariable, well-curated data offering analytical insights for necessarily complex human behaviors and activities. Of the multitude of situations that this decision-making aptitude presents, the application to governmental policy offers a commanding case. This would allow decisions to be made for the benefit of societies and citizens based on rigorous objective information devoid of the traditional approach of choosing policies and societal values based on the opinion of a handful of selected representatives who may be exposed to a lack of comprehensive data analysis capacity and subject to personal biases. There would need to be a critical requirement of wider socially responsible data practices here, beyond those of technical considerations and the incorporation of wider societal fairness approaches. Amongst the schools of political thought particularly acquiescent to the application by this approach would be the egalitarian approach of John Rawls. Here an Original Position’s pre-determination tool of Veil of Ignorance and ensuing Difference Principal presents a method of distributive justice that can be clearly mathematically defined in economics theory through Wald’s Maximin principle. This offers an opportunity to apply algorithmic game theory and artificial intelligence computational approaches to implement Rawlsian distributive justice that are presented and discussed. The outputs from the algorithmic acquaintance of Rawlsian egalitarianism with applicable state data, protected with appropriate privacy, security, legal, ethical and social governance could in turn lead to automated direct governmental choices and an objective Social Contract for citizens of digitally literate nations.
... Algorithm designers today are not trained to think through the systems implications of their choices, and even more complicated, the societal implications [4], [5], [6]. At the same time, because of the pace with which AI innovations are affecting the world, trainees must excel not only in algorithmic design, mathematical rigor and programming patterns and abstractions but they must also be equipped to engage with the societal implications of their innovations [7], [8], [9], [10], [11], [24]. ...
... While the opportunities and constraints of algorithmic decision-making with AI have been a trending subject of scholarly urban studies literature (Wu and Silva 2010;Newell and Marabelli 2015;Kitchin 2017;Yigitcanlar et al. 2021a), there are only a few academic studies focused on the perceptions on automated decision-making concerning cities by AI (Cui and Wu 2019;Kassens-Noor et al. 2021). These studies mostly concentrated on public perceptions (Yigitcanlar et al. 2020c;Araujo et al. 2020;Kankanamge et al. 2021;Schiff et al. 2021) or the perceptions of the stakeholders from a specific sector, most commonly health (Sun and Medaglia 2019;Lai et al. 2020), or data sources and the analytical techniques (including AI) that local governments use (Vogl et al. 2020;Watson and Ryan, 2020). ...
Article
Full-text available
Highly sophisticated capabilities of artificial intelligence (AI) have skyrocketed its popularity across many industry sectors globally. The public sector is one of these. Many cities around the world are trying to position themselves as leaders of urban innovation through the development and deployment of AI systems. Likewise, increasing numbers of local government agencies are attempting to utilise AI technologies in their operations to deliver policy and generate efficiencies in highly uncertain and complex urban environments. While the popularity of AI is on the rise in urban policy circles, there is limited understanding and lack of empirical studies on the city manager perceptions concerning urban AI systems. Bridging this gap is the rationale of this study. The methodological approach adopted in this study is twofold. First, the study collects data through semi-structured interviews with city managers from Australia and the US. Then, the study analyses the data using the summative content analysis technique with two data analysis software. The analysis identifies the following themes and generates insights into local government services: AI adoption areas, cautionary areas, challenges, effects, impacts, knowledge basis, plans, preparedness, roadblocks, technologies, deployment timeframes, and usefulness. The study findings inform city managers in their efforts to deploy AI in their local government operations, and offer directions for prospective research.
... Isso tudo acontece num cenário de capitalismo globalizado, com o ambiente digital desregulado e inserido em plataformas tornadas gigantes do setor das TICs, que tiram proveito de seu poder para ampliar atuação, se utilizando da manipulação de dados e controle de algoritmos (Newell & Marabelli, 2015;Morozov, 2018). ...
Chapter
Full-text available
This study investigates actors of media activism movements, with emphasis on communication collectives and Mídia Ninja - acronym for Independent Narratives, Journalism and Action - which erupted in 2013, along with large street demonstrations in Brazil, and became a model for other media activist initiatives. Using as main methodologies in-depth interviews and direct observation, we propose a list of differentials that characterize the collective model of these media agents (collaborative practice, tendency to horizontality, non-neutral narrative), and exclusive aspects of Mídia Ninja (collective houses and funding). It is concluded that these differentials stimulate the emergence of values in those settings, which are important for understanding how these agents act and keep their participants engaged
... Isso tudo acontece num cenário de capitalismo globalizado, com o ambiente digital desregulado e inserido em plataformas tornadas gigantes do setor das TICs, que tiram proveito de seu poder para ampliar atuação, se utilizando da manipulação de dados e controle de algoritmos (Newell & Marabelli, 2015;Morozov, 2018). ...
Book
Full-text available
O livro “Organizações e Movimentos Periféricos nas Redes Digitais Ibero-Americanas” é um esforço coletivo para retratar diferentes aspectos da atuação de forças sociais no ambiente virtual, sob o prisma das periferias, conceito em plena construção, como ressalta a investigadora Mara Rovida (2020). Estudos geográficos e sociológicos atrelam periférico/a à regiões e a indivíduos afastados dos centros urbanos e dos equipamentos sociais, marcados pela pobreza e segregação (D’Andrea, 2013). Essa mesma periferia geraria uma noção identitária de quem produz o território (Santos, 2002), a ponto de ser um local em potência, dada a dinâmica social poderosa realizada por seus sujeitos periféricos (D’Andrea, 2020). No que se refere à comunicação social, as periferias deteriam o potencial do que Rovida chama de diálogo social solidário nas bordas urbanas, (2020. 6), uma reinterpretação da Solidariedade Orgânica (Durkheim, 1977, in 2004), em dinâmica de cooperação necessária ou interdependência, e da prática jornalística como forma de interação social, ação coletiva e dependente da interação entre sujeitos (Medina, 2014).
Article
Full-text available
This article explores the synergy between applied mechanics and decisionmaking in the realms of management and commerce. Leveraging principles from physics and engineering, particularly Newtonian laws, applied mechanics provides a unique framework to understand and optimize the dynamics of organizational decision-making. The article delves into concepts such as strategic inertia, balancing forces in commerce, achieving equilibrium in market dynamics, and applying mechanics to risk management. A case study from the automotive industry illustrates the practical application of these principles. This interdisciplinary approach enhances strategic navigation, enabling organizations to make informed decisions, adapt to change, and thrive in the ever-evolving business landscape.
Chapter
The rule of law is an elusive concept, and its fluidity lends itself to multiple interpretations. Different accounts connect various core elements—‘desiderata’—under this universally recognised concept. However, there is a consensus (albeit implicit) that the rule of law is essentially a public law concept, of only marginal concern to private law. This paper departs from this understanding and suggests that this presumption is a misperception. The rule of law does not concern only the regulation of powers and arbitrariness between individuals and the State, but it operates also in the relationships between private individuals. In particular, with the flare-up in recent years in the use of machine learning algorithms to profile online users (in order to predict their behaviour and tailor recommendations and searches to their preferences), private actors (i.e. online platforms) have obtained a super-dominant position both in the collection of data and development of the technology, in the digital (eco)systems in which they operate. This paper aims to prospectively assess the duplicitous relevance that algorithmic profiling has for the protection of fundamental rights from a private law perspective (e.g. right to privacy, right to not be discriminated, freedom of expression) and for the self-appointed power of online platforms to self-regulate their contractual relationships with users, in the digital markets. Conversely, it also discusses the relevance of the rule of law for private law relationships in its function of stronghold for the protection of fundamental rights. This value, on one side, creates legal guardrails around private self-regulation of online platforms, and on the other side, it secures the respect of fundamental rights of users from algorithmic profiling by online platforms. The paper concludes with the need to re-evaluate the State’s power to limit private freedom and interfere in parties’ autonomy in cases where fundamental rights are seriously at stake.
Article
Full-text available
In algorithmic work, algorithms execute operational and management tasks such as work allocation, task tracking and performance evaluation. Humans and algorithms interact with one another to accomplish work so that the algorithm takes on the role of a co‐worker. Human–algorithm interactions are characterised by problematic issues such as absence of mutually co‐constructed dialogue, lack of transparency regarding how algorithmic outputs are generated, and difficulty of over‐riding algorithmic directive – conditions that create lack of clarity for the human worker. This article examines human–algorithm role interactions in algorithmic work. Drawing on the theoretical framing of organisational roles, we theorise on the algorithm as role sender and the human as the role taker. We explain how the algorithm is a multi‐role sender with entangled roles, while the human as role taker experiences algorithm‐driven role conflict and role ambiguity. Further, while the algorithm records all of the human's task actions, it is ignorant of the human's cognitive reactions – it undergoes what we conceptualise as ‘broken loop learning’. The empirical context of our study is algorithm‐driven taxi driving (in the United States) exemplified by companies such as Uber. We draw from data that include interviews with 15 Uber drivers, a netnographic study of 1700 discussion threads among Uber drivers from two popular online forums, and analysis of Uber's web pages. Implications for IS scholarship, practice and policy are discussed.
Article
Full-text available
Business intelligence and analytics (BI&A) has emerged as an important area of study for both practitioners and researchers, reflecting the magnitude and impact of data-related problems to be solved in contemporary business organizations. This introduction to the MIS Quarterly Special Issue on Business Intelligence Research first provides a framework that identifies the evolution, applications, and emerging research areas of BI&A. BI&A 1.0, BI&A 2.0, and BI&A 3.0 are defined and described in terms of their key characteristics and capabilities. Current research in BI&A is analyzed and challenges and opportunities associated with BI&A research and education are identified. We also report a bibliometric study of critical BI&A publications, researchers, and research topics based on more than a decade of related academic and industry publications. Finally, the six articles that comprise this special issue are introduced and characterized in terms of the proposed BI&A research framework.
Article
Full-text available
The era of Big Data has begun. Computer scientists, physicists, economists, mathematicians, political scientists, bio-informaticists, sociologists, and other scholars are clamouring for access to the massive quantities of information produced by and about people, things, and their interactions. Significant questions emerge. Will large-scale search data help us create better tools, services, and public goods? Or will it usher in a new wave of privacy incursions and invasive marketing? Will data analytics help us understand online communities and political movements? Or will it be used to track protesters and suppress speech? Will it transform how we study human communication and culture, or narrow the palette of research options and alter what 'research' means? Given the rise of Big Data as a socio-technical phenomenon, we argue that it is necessary to critically interrogate its assumptions and biases. In this article, we offer six provocations to spark conversations about the issues of Big Data: a cultural, technological, and scholarly phenomenon that rests on the interplay of technology, analysis, and mythology that provokes extensive utopian and dystopian rhetoric.
Book
For MIS specialists and nonspecialists alike, teacher and consultant Dan Power provides a readable, comprehensive, understandable guide to the concepts and applications of decision support systems. Power defines DSS broadly: interactive computer-based systems and subsystems that help people use computer communications, data, documents, knowledge, and models to solve problems and make decisions. This book covers an expanded framework for categorizing Decision Support Systems (DSS), a general managerial and technical perspective on building DSS, details and examples of the general types of DSS, and tools and issues associated with assessing proposals for DSS projects. A glossary and DSS readiness audit questions give special, ongoing value to all readers. Free eBook at https://scholarworks.uni.edu/facbook/67/