Conference PaperPDF Available

Conversational AI: Social and Ethical Considerations

Authors:

Abstract

Conversational Agents are becoming ubiquitous in our daily lives. They are used in various areas including customer service, education, medicine, and entertainment. As tools that are increasingly permeating various social domains, Conversational Agents can have a direct impact on individual’s lives and on social discourse in general. Consequently, critical evaluation of this impact is imperative. In this paper, we highlight some emerging ethical issues and suggest ways for agent designers, developers, and owners to approach them with the goal of responsible development of Conversational Agents.
Conversational AI: Social and Ethical
Considerations
Elayne Ruane, Abeba Birhane, Anthony Ventresque
School of Computer Science, University College Dublin, Ireland
Lero - The Irish Software Research Centre
{elayne.ruane, abeba.birhane}@ucdconnect.ie
Abstract. Conversational Agents are becoming ubiquitous in our daily
lives. They are used in various areas including customer service, educa-
tion, medicine, and entertainment. As tools that are increasingly perme-
ating various social domains, Conversational Agents can have a direct
impact on individual’s lives and on social discourse in general. Conse-
quently, critical evaluation of this impact is imperative. In this paper,
we highlight some emerging ethical issues and suggest ways for agent
designers, developers, and owners to approach them with the goal of
responsible development of Conversational Agents.
Keywords: Conversational Agent, Intelligent Systems, Social Impact, Ethics
1 Introduction
Conversational AI allows human users to communicate with an automatic system
using natural language. The interaction may be speech and/or text based. It may
be served to the user through messaging channels (e.g. Facebook Messenger and
Skype), through dedicated phone or web applications, integrated into a website,
or shipped as part of an operating system. Conversational AI systems have many
names depending on their capabilities, domain, and level of embodiment. These
terms include automatic agent, virtual agent, conversational agent, chatbot, or,
for very simple systems, bot. In this paper we use the term Conversational AI
to refer to any use of Machine Learning (ML) and Deep Learning (DL) models,
Natural Language Understanding and Processing (NLU & NLP) techniques, and
dialogue management systems to understand user input and generate natural
language responses. We use the term Conversational Agent (CA) to refer to
systems that have a Conversational AI component and have other features such
as a user interface (UI) to facilitate interaction and server-side features such as
the app logic and the database.
The year 2016, dubbed the “Year of the Bot” after Microsoft CEO Satya
Nadella described bots as the new apps, saw the launch of more than 30,000
chatbots on the Facebook Messenger platform alone [8] [11]. By 2018, there
were more than 300,000 active bots with 8 Billion messages exchanged every
month on the platform [5]. Much of this growth is driven by commercial in-
terests. Chatbots are an inexpensive, fast, and always-on service for answering
FAQs and completing other well-defined tasks. Although quality remains an is-
sue for more complex tasks and conversational system evaluation is an active
area of research [39], CAs have seen adoption in various social domains includ-
ing customer service and product recommendation, education support, medical
services, entertainment, social outreach, and personal organisation.
Ethical concerns inevitably arise with any technological innovation. However,
they are often considered secondary to technical development challenges, if they
are considered at all [7] [36] [51]. As with any technology that permeates our
daily lives, the development and application of conversational AI raises various
ethical questions. While some concerns, such as privacy, are an active area of
research [29], others have received less attention. This paper examines the eth-
ical challenges posed by the integration of conversational systems into human
interaction as well as the necessary cautions and measured steps that need to
be considered in developing and deploying CAs. We hope that this paper can
serve as a call to action for agent designers, developers, and owners. Section 2
motivates this work by highlighting the potential harms of Conversational AI
and Section 3 discusses relevant work from the literature. In Section 4, we iden-
tify a number of concerns and propose a way forward for critical and ethical
engagement throughout the design and development process.
2 Motivation and Contribution
The history of humanity is full of examples of technology as a force for soci-
etal and behavioural change from the earliest prehistoric stone tools through to
the invention of the computer, internet, and other advances in Information and
Communications Technology. The pace of change has accelerated, and successive
generations leading quite different lifestyles due to the impact of technological
change. The most recent example may be the wide-spread adoption and usage
of smartphones [43] [17] and social media [24] which gives individuals communi-
cation capabilities and access to information and other media that is changing
the social and political landscape.
Spurred on by the use of smartphones, the last decade has seen the adoption
and integration of CAs in our day-to-day lives. The release of Apple’s virtual
assistant Siri in 2011, shipped with the iPhone 4S, marked the start of the
ubiquity seen today where conversational AI is present in our homes, offices, and
social media platforms shaping how we interact with companies and services.
A survey (n=800) of marketing professionals by Oracle found 36% of brands
surveyed had implemented chatbots for customer service with an increase to
80% expected by 2020 [37]. According to a report from Global Market Insights,
Inc., the intelligent virtual assistant market which includes Apple’s Siri, Google
Assistant, Amazon’s Alexa, and Microsoft’s Cortana, is expected to grow from
a $1 billion valuation in 2017 to $11.5 billion by 2024 [20]. This growth directly
and indirectly impacts how individuals interact with services, consume media,
and interact online.
Such ubiquity and increasing integration makes CAs forces that shape, alter,
and impact the experience of individuals and groups. Their impact and potential
harm varies depending on the domain and target user group. Relatively simple
social bots, for example, have transformed the political landscape. In a study
that examined the impact of bots in the 2016 US presidential election, Bessi and
Ferrara (2016) [4] found that the presence of social media bots negatively affects
democratic political discussion. Although social bots are often benign and useful,
they can be used to manipulate and mislead users by spreading misinformation
which has been particularly effective on Twitter and Facebook.
Like any other AI system, CAs do not exist in a social, political, economic,
and cultural vacuum. They are developed by individuals or teams of individuals
with specific, often commercial, aims. CAs necessarily reflect the values and per-
spectives of such individuals and the interests of the respective industry. When
chatbots are rolled-out to users, they become part of the social utility where the
implications of their design can be felt by real people. However, a combination
of a lack of awareness of the technology behind these agents among the general
public, company-level confidentiality, and the emerging nature of this technol-
ogy has created an environment in which ethical concerns are not well-defined
around Conversational AI. As such, we argue those involved in the process of
developing and deploying CAs have a responsibility to critically examine the
social impact of their tools and to view such practice as an integral part of the
development process.
3 Background and Related Work
Although many major companies, research institutions, and public sector or-
ganizations have all issued guidelines for ethical artificial intelligence, recent
work [22] has discovered substantive divergence in how these are written and
interpreted, highlighting the complexity of designing guidelines for systems with
complex social impact. An emerging body of work indicates that the integra-
tion of AI systems into various social spheres brings with it a host of often
unanticipated and harmful outcomes. Furthermore, users from disadvantaged
backgrounds, such as those with disabilities or those that face racial, gender, or
other bias, may face disproportionate harm. Various studies illustrate this, as
bias is found in: detecting skin tones in pedestrians [52], predictive policing sys-
tems and justice [38], the display of STEM career ads [27], recidivism algorithms
[2], politics of search engines [21], medical applications [13], automatic speech
recognition [44], and in hiring algorithms [1]. This emerging body of work that
critically examines unfairness, injustice, bias and discrimination within various
areas of AI is invaluable. However, there are a number of ethical considerations
that are unique to machine-human conversation that have not yet become de
facto considerations in the design and development stages of building a chatbot
or other CA.
Conversational AI and Human-Computer Interaction (HCI) are active fields
of research within academia. However, most publicly deployed CAs are developed
by industry stakeholders among which there is little cross-collaboration or pub-
lication of proprietary training datasets and system architectures. This makes
critical engagement and analysis of social impact difficult. Given their ubiqui-
tous presence in various social, political, and financial spheres, we contend that
CAs might be best viewed primarily as social utilities, and not solely as corpo-
rate assets. The effect of unintended consequences as a direct result of design
decisions holds the potential to harm people. Consequently, critical engagement
is required throughout design and development.
Language is central to Conversational AI systems as a medium that facilitates
interaction. Effective and responsible design of CAs requires an understanding
of various linguistic elements of conversation as well as an awareness of wider
social and contextual factors [18] [46]. Language, as a social activity embed-
ded in historical, cultural, and social norms is not a “neutral” or “objective”
medium. Rather, it reflects existing societal values and judgements [30]. Take,
for example, how the meaning of, and the discourse around, the word “gay” has
changed since the 1950s. Language is situational and contextual - a single word
or conversation can have radically different meanings depending on context and
time. “Acceptable” norms and forms of conversing in one context might be per-
ceived as “unacceptable” or “deviant” in another. Consequently, conversation
formats, phrases, and words that are perceived as “acceptable” or “standard”
might represent the status quo, leaving anything outside the status quo either
implicitly or explicitly coded as an anomaly or outlier [6]. Decisions made dur-
ing development regarding various aspects of language such as accent, dialect,
and register can encode socially held beliefs and assumptions of, for example,
“standard language” into the system. Language registers and expressions that
are used by target user groups but not recognized by an agent are implicitly
deemed outside the “norm”. The language(s) accepted and understood by the
system reflect the accessibility of the system and this is a deliberate choice dur-
ing the design phase that may have significant knock-on effect for users after
deployment. In the process of developing CAs, these nuances of language and
conversation, and the problems that arise due to lack of awareness around them,
should take centre stage alongside the technical challenges.
Language is inherently social, cultural, contextual, and historical, which
means that the design of agent dialogue necessarily reflects a particular world-
view. As tools that exist within the social realm, socially sensitive conversations
are unavoidable. How these socially sensitive issues are responded to plays a
significant role in terms of how such sensitive issues and individuals affected by
them are perceived. Recent work [9] studied how CAs handle sensitive requests
involving sexual harassment and bullying. The authors found that while com-
mercial conversational systems often avoid answering such requests altogether
and rule-based systems usually try to deflect these topics, data-driven systems
risk responding in a way that can be interpreted as flirtatious and sometimes
counter-aggressive. Similarly, it was found that race related conversations are of-
ten deflected by chatbots [41]. Given their pervasiveness, these topics are some-
thing an open-domain agent should be designed to handle responsibly. Although
the rationale behind such design is to take a “neutral” stance, avoidance and de-
flection of complex social issues can symbolize either endorsement, trivialization,
or devaluation of the topic or an individual’s experience.
The use of CAs within mental health services is another area where critical
reflection is required. The gap between the demand for mental health services
and lack of available resources, as well as the cost efficiency and seemingly non-
judgmental nature of CAs, makes them seem an attractive solution. So far,
CAs have been bestowed with responsibilities including screening diagnosis and
treatment of mental health [19] [35] [47]. However, despite being perceived as
less-stigmatizing, CAs might actually pose harm to users due to their limited
capacity to re-create human interaction and to provide tailored treatment, es-
pecially if they are not continually audited and evaluated [25]. Mental health
services meet people at their most vulnerable. Consequently, any conversational
interactions with such users needs utmost ethical and critical attention. How-
ever, ongoing evaluation for harms and benefits, which is essential for ethical
and responsible practice, is absent in many digital platforms and apps for men-
tal health [25]. Unfortunately, this is not limited to this application domain.
Among the varied applications of CAs, one common recurring theme is a lack of
critical assessment. Evaluation of the use of CAs often mentions the importance
of ethical considerations but fails to explicitly discuss such concerns or provide
mechanisms to address them such as in [19] [35] and [54].
There have been numerous approaches proposed to implement ethical deci-
sion making for AI agents. Some argue the best approach is within the context
of Safety Engineering whereby safety mechanisms are used to mitigate harm-
ful impact of AI systems. Others argue for a Machine Ethics approach which
involves encoding ethical standards and reasoning into the AI systems them-
selves [32] [3]. In this paper, we argue for a shift in mindset that considers social
context in identifying and addressing ethical concerns specific to conversational
AI throughout the design and development process. We place responsibility on
designers and developers for cultivating awareness of these issues and how their
approaches impact the end user, as opposed to discussing general ethical ap-
proaches and focusing on agent decision-making. In the next section, we discuss
aspects of conversational AI that require critical reflection throughout the de-
sign and development phases. This is not a complete list of concerns that arise
with Conversational AI by any means. Rather, these are some concerns we have
focused on as a point of discussion with the aim of bringing forth and clarifying
implicit assumptions and the impact they may have on users.
4 Towards Ethical Conversational Agents
4.1 Plurality of approaches
Ethical concerns that emerge with Conversational AI vary markedly depending
on the application domain, target user group, and the goal(s) of the agent. As
such, an understanding of the domain and the problem that the agent aims to
solve should inform the identification of possible ethical concerns and solutions.
For example, a chatbot used within an organisation by employees for a specific
purpose will have a considerably different set of considerations than a customer
or public-facing agent that may be expected to answer general or unconstrained
queries. For responsible system design, deep understanding of the user groups
characteristics, contexts, and interests is imperative. For example, a recent sur-
vey on the use of CAs in education and associated user concerns revealed that
people were open to this technology if privacy issues are addressed but found
that there were significant differences in how adults and children viewed privacy
in this context [28]. Such insight, and its incorporation into the design of the
system, is critical for ethical and responsible design that centres the values and
interests important to users. As such, embracing contextual, flexible, and plural
methods of identifying and addressing ethical concerns is imperative. Addition-
ally, identifying solutions that are the most suitable to the specific scenario
should always be prioritized over attempting to fit some standard principles.
While failure to anticipate and mitigate potential ethical issues can result in
destructive, traumatic, or dangerous outcomes in some circumstances, emerging
issues might be easily contained and corrected in others. Consequently, there
is no one-fits-all ethical standard or principle that can be applied to all CAs.
Therefore, in the strive to develop ethical and responsible Conversational AI, we
encourage contextual and plural approaches over a set of abstract principles.
4.2 Trust and Transparency
Providing users with choices, and consequently with control, over how they prefer
to interact with an agent, is an important first step towards centring users needs
and wellbeing. Transparency about an agents status as automatic (non-human)
and the limits of its capabilities, for example, is essential in order to allow users
to make informed choices, which further contributes to users trust. Recent work
has shown that users behave and interact differently when conversing with an
automatic agent compared to interacting with another human [34]. If users are
aware that they are speaking to an automated agent or a human agent, then they
might be able to make informed decisions with regards to their own behaviour, in
particular regarding information disclosure [14]. This is especially crucial where
the user information being discussed or disclosed is sensitive, such as in banking
or education of minors, or where the implications and/or consequences of the
conversation are significant such as user health concerns.
Understanding user expectations of an agent is crucial in ensuring that user
trust in not taken advantage of. Reasonable expectations should be identified
and validated before the agent is published. For example, if a user expects a
conversation to be anonymous, then identifiable plain text conversation logs
should not be visible to individuals on the development team. Similarly, if a
chatbot has been designed to recommend products, such as the retailer H&M’s
chatbot which helps users to plan and purchase outfits, a user may expect rel-
atively unbiased information such that the chatbot will not show clothes from
other retailers but also that it won’t only show the most expensive H&M clothes
either. The user’s assumption of agent neutrality is part of a widely held but
often misguided belief that AI systems are unbiased. It can be difficult to eval-
uate the behaviour of a system such that we can validate whether the agent
recommends products based on genuine interests or needs instead of profiling
users by features such as gender, race, age, or location in a way that may harm
their opportunity for a fair purchase. Nonetheless, given the magnitude of harm
that this might cause, it is imperative to continually assess and ensure that users
are not profiled based on these sensitive features. CAs that engage with users
in a higher-risk scenario such as mental health services as opposed to clothing
or household-item purchases, have a greater social responsibility towards their
users and how the service may affect them. In any scenario, the user should be
able to trust the system not to take advantage of them and to provide the stated
service in good faith. This requires (1) explicitly detailing the agent’s motiva-
tions and explaining its behaviour in a way the target user group can understand
(2) evaluation to determine how the agent is treating various types of users, and
(3) an understanding of users concerns, expectations, and experience.
4.3 Privacy
The interaction of humans and CAs, and sometimes even the presence of virtual
agents such as in-home, always-on devices, present various ethical and legal
questions including what data is collected, who has access to it, how long the data
is stored and where and what such data is used for. Collecting user data raises
many privacy concerns, some of which have legal basis and are covered by data
protection laws that vary geographically, such as GDPR in Europe. The nature
of these ethical issues varies significantly depending on the domain in which the
agent is deployed and the level of vulnerability of the user group. However, we
propose that clear legal requirements should be viewed as a baseline, not a target,
in this area where the default approach should be to only collect and store user
data if required for delivery of the stated service and to do so in a transparent
manner. User privacy is paramount and is becoming increasingly important as
we see AI systems rolled out into more areas of society where such systems are
used to make increasingly substantial and far-reaching decisions. This makes the
concept of privacy something that should not be framed entirely as a problem
regarding the individual user but rather as a wider social concern. The individual
user is often not afforded the opportunity or does not have the resources to
negotiate terms and conditions that are written by corporations in a manner
that applies to all. How we think about and legislate privacy, therefore, should
be considered in light of how the collective might be impacted by the introduction
of AI systems. This perspective is helpful in re-conceptualizing privacy in a way
that links it to the bigger picture of collective aspirations and concerns.
A distinct concern with respect to CAs in this area is the influence that the
social relations that users develop with an agent and the way user-agent inter-
action is often perceived as anonymous [12] [14], can encourage self-disclosure
of information. Additionally, the dialogue design of an agent impacts users in-
clination to self-disclose. Self-disclosure may be encouraged to gather data with
the goal of improving user experience via personalization [40]. However, unlike
explicitly submitting data via a structured form, users may not be conscious of
how much information they have divulged via a conversation or what personal
data can be inferred from their natural language utterances. Furthermore, users
may not know how the system works on a technical level with regards to the
processing and storing of their data [31]. For these reasons, the unique context of
CAs with respect to privacy should be considered when aiming to comply with
legal requirements such as GDPR or any adopted privacy guidelines.
4.4 Agent Persona
A large part of agent design decisions relate to persona and personality, which
can be used to inform specific dialogue choices. Agent persona expressions in-
clude gender, age, race, cultural affiliation, and class. These indications may be
more explicit as the level of embodiment increases. It is important to consider
the impact of agent persona on the types of relationships users may try to ex-
plore with the agent and to determine if the design of the agent persona and
accompanying dialogue is encouraging behaviour that may be harmful. Agent
persona design can also inadvertently reinforce harmful stereotypes. Many pub-
licly available agents present as female, including popular assistants such as Siri,
Alexa, Cortana and the default female voice for Google Assistant [50]. While
female personas are often used in subservient contexts, male personas are often
found in situations perceived as authoritative, such as an automatic interviewer
[23] [45] [55]. Gendering CAs in this manner may reflect market research but
in the interests of gender equity, practices that embed and perpetuate socially
held harmful gender stereotypes should be avoided. In some domains, there is
an increased move towards androgyny such as banking agent Kai. Research has
been conducted on how users respond to androgynous agents and the effects this
has on user experience. A study that analysed college students’ perceptions of
gendered vs androgynous agents [15], found a gender-neutral agent led to more
positive views on females than a female-presenting agent did. Another similar
study by [42] found that female agents received more abuse than androgynous
agents. There is no clear consensus within the industry on this issue. Some rec-
ommend allowing users to lead the agent persona by designing the agent to
dynamically respond to how the user interacts. Others continue to gender the
agents they build in an attempt to humanize the system and increase user satis-
faction at the risk of reinforcing harmful gender bias. We recommend designing
agents to be androgynous to avoid gender stereotypes and allow users to interpret
according to their own context.
4.5 Anthropomorphism and Sexualization
Humans tend to anthropomorphize machines [26]. This kind of anthropomor-
phism is exacerbated when users can interact conversationally with a system
and especially if the system has been imbued with personality and embodied
with an avatar or in some other way. This can be seen throughout history and
occurs even when the developers themselves oppose such anthropomorphism and
over-hyping of machines. The creator of ELIZA (1964-66) Joseph Weizenbaum,
for example, explicitly insisted that ELIZA could not converse with true under-
standing. Despite this, many users were convinced of ELIZAs intelligence and
empathy [49]. Possibly a surprising element of human-computer interaction is
unsolicited romantic attention towards the agent. A good example of this is the
popular entertainment chatbot Mitsuku1which has won the Loebner Prize four
times. Steve Worswick, the creator and maintainer of Mitsuku, has described the
type of romantic attention ”she” gets and even the correspondence he receives
from users demanding her freedom [53].
Research has shown users use greater profanity with a chatbot than with a
human and are similarly more likely to harass a chatbot than a human agent
[16], even more so if the agent has a female persona [42]. Recent work [9] that
explored the capabilities of conversational agents to respond to sexual harass-
ment on the part of the user and collected 360,000 conversations found that
4% were sexually explicit, a percentage somewhat below previous research into
sexually explicit chatbot interactions. The authors argue handling these types
of conversations should be a core part of a systems design and evaluation due
to their prevalence and consequences of reinforcing gender bias and encouraging
aggressive behaviour.
Due to the prevalence of abusive messages directed at conversational agents,
unsupervised learning techniques on an unconstrained user group should be
avoided. Even with a trusted user group, oversight is required to ensure the
agent has not acquired harmful concepts or language. There are numerous ex-
amples of chatbots that have been released for use by the general public that
use unsupervised learning but quickly learn racist, homophobic, and sexist lan-
guage and have to be shut down to avoid abuse of human users. In the case of
Microsoft’s Tay bot, this took less than 24 hours [48]. Dialogue design should in-
volve response strategies for romantic attention, sexualized messages, and abuse
with the aim of protecting the user. If an agent can detect abusive language,
which is a difficult task for both social and technical reasons, it can invoke the
appropriate response strategy. This may be a non-response, a neutral response,
an in-kind response, or escalation to a human agent. In this scenario the domain
and goals of the agent are important, but the user demographic is the most in-
fluential factor when designing the agent’s response strategy [10]. For example,
it is very rare that an in-kind response, that is responding with similar tone
and content as the abusive message, will be an ethical and acceptable response
strategy. In the case of an education bot that converses with minors, escalation
to a human (maybe a teacher) is the most appropriate response. It should be
noted that a neutral response can be seen as endorsement. Engaging in use-case
centred discourse can help to elicit social values that may then be used to inform
1Mitsuku: https://www.pandorabots.com/mitsuku/
the design of a specific agent’s response strategy, especially where variation of
values across user groups is high (value pluralism) [33].
5 Conclusion
Assuming agents continue to improve in their functionality and conversational
ability, how will their ubiquity and integration in our daily lives change how we
live? Who will be most affected by the decisions of agent owners? These questions
are difficult to answer but provide perspective on the ethical issues raised in this
paper. Ultimately, there are no one-approach-fits-all answers to the concerns we
have discussed. However, designing, building, and deploying an agent into the
social sphere engenders a level of social responsibility that must be confronted
and contemplated on an agent-by-agent basis to produce agent-specific strategies
to address the ethical considerations described in this paper.
Acknowledgement. This work was supported, in part, by Science Foundation
Ireland grant 13/RC/2094.
References
1. Ajunwa, I., Friedler, S., Scheidegger, C.E., Venkatasubramanian, S.: Hiring by
algorithm: predicting and preventing disparate impact. Available at SSRN (2016)
2. Angwin, J., Larson, J., Mattu, S., Kirchner, L.: Machine bias. ProPublica (2016)
3. Baum, S.D.: Social choice ethics in artificial intelligence. AI & SOCIETY pp. 1–12
(2017)
4. Bessi, A., Ferrara, E.: Social bots distort the 2016 us presidential election online
discussion. First Monday 21(11-7) (2016)
5. Boiteux, M.: Messenger a F8 2018 (2018)
6. Bourdieu, P.: Language and symbolic power. Harvard University Press (1991)
7. Cheney-Lippold, J.: We are data: Algorithms and the making of our digital selves.
NYU Press (2018)
8. Constine, J., Perez, S.: Facebook messenger now allows payments in its 30,000 chat
bots. techcrunch. URL: https://tcrn.ch/2cDEVbk (2016)
9. Curry, A.C., Rieser, V.: #MeToo Alexa: How conversational systems respond to
sexual harassment. In: Proceedings of the Second ACL Workshop on Ethics in
Natural Language Processing. pp. 7–14 (2018)
10. Curry, A.C., Rieser, V.: A crowd-based evaluation of abuse response strategies in
conversational agents. arXiv preprint arXiv:1909.04387 (2019)
11. Dale, R.: The return of the chatbots. Natural Language Engineering 22(5), 811–817
(2016)
12. Evans, R.E., Kortum, P.: The impact of voice characteristics on user response in
an interactive voice response system. Interacting with Computers 22(6), 606–614
(2010)
13. Ferryman, K., Pitcan, M.: Fairness in precision medicine. Data & Society (2018)
14. Gentsch, P.: Conversational ai: How (chat) bots will reshape the digital experience.
In: AI in Marketing, Sales and Service, pp. 81–125. Springer (2019)
15. Gulz, A., Haake, M.: Challenging gender stereotypes using virtual pedagogical
characters. In: Gender Issues in Learning and Working with Information Technol-
ogy: Social Constructs and Cultural Contexts, pp. 113–132. IGI Global (2010)
16. Hill, J., Ford, W.R., Farreras, I.G.: Real conversations with artificial intelligence:
A comparison between human–human online conversations and human–chatbot
conversations. Computers in Human Behavior 49, 245–250 (2015)
17. Howard, R., Kehoe, J.: Mobile consumer survey 2018: The irish cut (2018)
18. Hutchby, I., Wooffitt, R.: Conversation analysis. Polity (2008)
19. Inkster, B., Sarda, S., Subramanian, V.: An empathy-driven, conversational artifi-
cial intelligence agent (wysa) for digital mental well-being: real-world data evalu-
ation mixed-methods study. JMIR mHealth and uHealth 6(11), e12106 (2018)
20. Insights, G.M.: Intelligent virtual assistant (iva) market trends share forecast 2024s
(2018)
21. Introna, L., Nissenbaum, H.: The politics of search engines. IEEE Spectrum 37(6),
26–27 (2000)
22. Jobin, A., Ienca, M., Vayena, E.: The global landscape of ai ethics guidelines.
Nature Machine Intelligence pp. 1–11 (2019)
23. Kim, Y., Baylor, A.L., Shen, E.: Pedagogical agents as learning companions: the
impact of agent emotion and gender. Journal of Computer Assisted Learning 23(3),
220–234 (2007)
24. Kim, Y., Wang, Y., Oh, J.: Digital media use and social engagement: How social
media and smartphone use influence social activities of college students. Cyberpsy-
chology, Behavior, and Social Networking 19(4), 264–269 (2016)
25. Kretzschmar, K., Tyroll, H., Pavarini, G., Manzini, A., Singh, I., Group, N.Y.P.A.:
Can your phone be your therapist? young peoples ethical perspectives on the
use of fully automated conversational agents (chatbots) in mental health support.
Biomedical Informatics Insights 11 (2019)
26. Kuipers, B., McCarthy, J., Weizenbaum, J.: Computer power and human reason.
ACM SIGART Bulletin (58), 4–13 (1976)
27. Lambrecht, A., Tucker, C.: Algorithmic bias? an empirical study of apparent
gender-based discrimination in the display of stem career ads. Management Science
(2019)
28. Latham, A., Goltz, S.: A survey of the general publics views on the ethics of using
ai in education. In: International Conference on Artificial Intelligence in Education.
pp. 194–206. Springer (2019)
29. Lau, J., Zimmerman, B., Schaub, F.: Alexa, are you listening?: Privacy perceptions,
concerns and privacy-seeking behaviors with smart speakers. HCI 2, 102 (2018)
30. Linell, P.: Rethinking language, mind, and world dialogically. IAP (2009)
31. Luger, E., Sellen, A.: Like having a really bad pa: the gulf between user expec-
tation and experience of conversational agents. In: Proceedings of the 2016 CHI
Conference on Human Factors in Computing Systems. pp. 5286–5297 (2016)
32. Meek, T., Barham, H., Beltaif, N., Kaadoor, A., Akhter, T.: Managing the ethical
and risk implications of rapid advances in artificial intelligence: a literature review.
In: 2016 Portland International Conference on Management of Engineering and
Technology (PICMET). pp. 682–693. IEEE (2016)
33. Morley, J., Floridi, L., Kinsey, L., Elhalal, A.: From what to how. an overview of
ai ethics tools, methods and research to translate principles into practices (2019)
34. Mou, Y., Xu, K.: The media inequality: Comparing the initial human-human and
human-ai social interactions. Computers in Human Behavior 72, 432–440 (2017)
35. Oh, K.J., Lee, D., Ko, B., Choi, H.J.: A chatbot for psychiatric counseling in mental
healthcare service based on emotional dialogue analysis and sentence generation.
In: 2017 18th IEEE International Conference on Mobile Data Management (MDM).
pp. 371–375. IEEE (2017)
36. O’Neil, C.: Weapons of math destruction: How big data increases inequality and
threatens democracy. Broadway Books (2016)
37. Oracle: Can virtual experiences replace reality? (2016)
38. Richardson, R., Schultz, J., Crawford, K.: Dirty data, bad predictions: How civil
rights violations impact police data, predictive policing systems, and justice. New
York University Law Review Online, Forthcoming (2019)
39. Ruane, E., Faure, T., Smith, R., Bean, D., Carson-Berndsen, J., Ventresque, A.:
Botest: a framework to test the quality of conversational agents using divergent
input examples. In: IUI. p. 64 (2018)
40. Saffarizadeh, K., Boodraj, M., Alashoor, T.M.: Conversational assistants: investi-
gating privacy concerns, trust, and self-disclosure. In: International Conference on
Information Systems. AIS (2017)
41. Schlesinger, A., O’Hara, K.P., Taylor, A.S.: Let’s talk about race: Identity, chat-
bots, and ai. In: Proceedings of the 2018 CHI Conference on Human Factors in
Computing Systems. p. 315 (2018)
42. Silvervarg, A., Raukola, K., Haake, M., Gulz, A.: The effect of visual gender on
abuse in conversation with ecas. In: International Conference on Intelligent Virtual
Agents. pp. 153–160. Springer (2012)
43. Smith, A.: Smartphone ownership–2013 update. Pew Research Center: Washington
DC 12, 2013 (2013)
44. Tatman, R.: Gender and dialect bias in youtubes automatic captions. In: Proceed-
ings of the First ACL Workshop on Ethics in Natural Language Processing. pp.
53–59 (2017)
45. Tay, B., Jung, Y., Park, T.: When stereotypes meet robots: the double-edge sword
of robot gender and personality in human–robot interaction. Computers in Human
Behavior 38, 75–84 (2014)
46. Te Molder, H., Potter, J.: Conversation and cognition. Cambridge University Press
(2005)
47. Vaidyam, A.N., Wisniewski, H., Halamka, J.D., Kashavan, M.S., Torous, J.B.:
Chatbots and conversational agents in mental health: a review of the psychiatric
landscape. The Canadian Journal of Psychiatry 64(7), 456–464 (2019)
48. Wakefield, J.: Microsoft chatbot is taught to swear on twitter. bbc news (2016)
49. Weizenbaum, J., et al.: Eliza—a computer program for the study of natural lan-
guage communication between man and machine. Communications of the ACM
9(1), 36–45 (1966)
50. West, M., Kraut, R., Ei Chew, H.: I’d blush if i could: closing gender divides in
digital skills through education (2019)
51. Whittaker, M., Crawford, K., Dobbe, R., Fried, G., Kaziunas, E., Mathur, V.,
West, S.M., Richardson, R., Schultz, J., Schwartz, O.: AI now report 2018. AI
Now Institute at New York University (2018)
52. Wilson, B., Hoffman, J., Morgenstern, J.: Predictive inequity in object detection.
arXiv preprint arXiv:1902.11097 (2019)
53. Worswick, S.: Ethics and chatbots. medium (2018)
54. Yu, H., Miao, C., Leung, C., White, T.J.: Towards ai-powered personalization in
mooc learning. npj Science of Learning 2(1), 15 (2017)
55. Zhou, M.X., Mark, G., Li, J., Yang, H.: Trusting virtual agents: The effect of
personality. Transactions on Interactive Intelligent Systems 9(2-3), 10 (2019)
... When exploring the world of ChatGPT and conversational AI in general, ethical issues are crucial since these technologies present a number of difficulties that need careful scrutiny [6]. The following four major ethical problems require special attention are: ...
... Users may accept false information if the model produces plausible-sounding but factually erroneous replies. This has significant implications for critical domains like health, economics, and education where bad advice can have real-world repercussions [6]. ...
... Interactions with AI agents like ChatGPT frequently entails disclosing private or delicate information. These interactions run the danger of being saved or utilized without the required authorization, thus breaching the privacy of users [6]. Maintaining user confidence depends critically on ensuring clear data usage regulations and safe data management procedures. ...
Chapter
This book series aims to provide a forum for researchers from both academia and industry to share their latest research contributions in the area of computing technologies and Data Sciences and thus to exchange knowledge with the common goal of shaping the future. The best way to create memories is to gather and share ideas, creativity and innovations.
... Another concern is the possibility that chatbots could be used to encourage academic dishonesty, such as by providing answers to assignments or exams. Some researchers have argued that chatbots have the potential to facilitate plagiarism by providing students with ready-made answers to assignments and assessments [24]. In [25], the authors presented the potential for chatbots to be used to cheat on online exams. ...
... One risk is the possibility that students may use ChatGPT in an unethical manner. This aligns with the findings of [24]. For example, a student might use ChatGPT to generate answers to questions that they are not able to solve on their own, or to write papers or complete assignments without doing the necessary research or understanding the material themselves. ...
... While it is unsurprising that GPT-4-the more advanced and costly version-performs better than the default version (GPT-3.5), this suggests potential ethical issues regarding tools available to users of different socioeconomic positions [37][38][39] . These potential ethical concerns can be especially relevant We believe that a comprehensive examination of the capabilities of generative AI tools, such as ChatGPT and Bard, will likely grow in value, considering their quickly increasing role in climate literacy 25,26 and their potential-yet debated-beneficial applications in the general education sector 22,27,40 within countries where it is available. ...
Article
Full-text available
In the face of climate change, climate literacy is becoming increasingly important. With wide access to generative AI tools, such as OpenAI’s ChatGPT, we explore the potential of AI platforms for ordinary citizens asking climate literacy questions. Here, we focus on a global scale and collect responses from ChatGPT (GPT-3.5 and GPT-4) on climate change-related hazard prompts over multiple iterations by utilizing the OpenAI’s API and comparing the results with credible hazard risk indices. We find a general sense of agreement in comparisons and consistency in ChatGPT over the iterations. GPT-4 displayed fewer errors than GPT-3.5. Generative AI tools may be used in climate literacy, a timely topic of importance, but must be scrutinized for potential biases and inaccuracies moving forward and considered in a social context. Future work should identify and disseminate best practices for optimal use across various generative AI tools.
Article
Since the introduction of OpenAI's ChatGPT‐3 in late 2022, conversational chatbots have gained significant popularity. These chatbots are designed to offer a user‐friendly interface for individuals to engage with technology using natural language in their daily interactions. However, these interactions raise user privacy concerns due to the data shared and the potential for misuse in these conversational information exchanges. Furthermore, there are no overarching laws and regulations governing such conversational interfaces in the United States. Thus, there is a need to investigate the user privacy concerns. To understand these concerns in the existing literature, this paper presents a literature review and analysis of 38 papers out of 894 retrieved papers that focus on user privacy concerns arising from interactions with text‐based conversational chatbots through the lens of social informatics. The review indicates that the primary user privacy concern that has consistently been addressed is self‐disclosure. This review contributes to the broader understanding of privacy concerns regarding chatbots the need for further exploration in this domain. As these chatbots continue to evolve, this paper acts as a foundation for future research endeavors and informs potential regulatory frameworks to safeguard user privacy in an increasingly digitized world.
Article
The integration of chatbot technology into human-robot interaction has emerged as a transformative approach to foster seamless communication and elevate user experiences. This paper explores the convergence of chatbots with robotics, elucidating the benefits and challenges encountered in this integration. By enabling natural language processing capabilities, chatbots facilitate more intuitive and interactive communication between users and robots. This paper examines the advantages of this integration, such as 24/7 availability, scalability, and improved efficiency, while highlighting challenges such as contextual understanding limitations, language dependencies, and ethical concerns regarding data privacy. Moreover, it delves into the implications of leveraging chatbots in diverse applications, including customer service, healthcare, and education. By comprehensively analyzing the landscape of chatbot integration in human-robot interaction, this paper aims to provide insights into maximizing the advantages while mitigating the challenges, thereby paving the way for more effective and engaging interactions between humans and robots.
Chapter
Autonomous and intelligent systems and services that use narrow artificial intelligence technologies such as statistical learning and limited inferencing (AISSN) are pervasive in our lives and industry. These systems, very far from having human-like intelligence, will offer significant potential for doing social good, achieving productivity gains and advancing science and engineering. However, AISSN systems can have unanticipated and harmful impacts. This chapter highlights the ethical challenges of AISSNs using three diverse and pervasive examples: Internet of Things, conversational AI, and semi-autonomous vehicles. We contend that AISSNs will be the norm for the foreseeable future and that artificial general intelligence will not develop anytime soon. The ethical challenges of AISSNs are addressable using human-centred “Ethical Design”, the use of widely accepted moral standards of right and wrong to guide the conduct of people in the ideation, design, development, and deployment of AISSN systems. Depending on the problem domain, multidisciplinary teams of computer scientists and engineers, sociologists, economists, ethicists, linguists, and cultural anthropologists will be required to implement humanistic design processes.
Article
Full-text available
As technology improves, artificial systems are increasingly able to behave in human-like ways: holding a conversation; providing information, advice, and support; or taking on the role of therapist, teacher, or counsellor. This enhanced behavioural complexity, we argue, encourages deeper forms of affective engagement on the part of the human user, with the artificial agent helping to stabilise, subdue, prolong, or intensify a person’s emotional condition. Here, we defend a fictionalist account of human/AI interaction, according to which these encounters involve an elaborate practise of imaginative pretence: a make-believe in which the artificial agent is attributed a life of its own. We attend, specifically, to the temporal characteristics of these fictions, and to what we imagine artificial agents are doing when we are not looking at them.
Article
Modern anlamda, tarihi 2. Dünya savaşına uzanan yapay zekâ, 2000’li yıllardan sonra günlük kullanımdaki dijital aletlere entegre edilmeye başladı. Ancak bu teknolojiler yalnızca içeriğinde yapay zekâ destekli uygulama bulunan aletlere sahip insanlar tarafından deneyimlenebildi. 30 Kasım 2022 tarihinde OpenAI firması tarafından ChatGPT adıyla kitlesel kullanıma sunulan sohbet robotu ile yapay zekâ kişisel ve profesyonel işlerde kullanılabilir hale geldi. Çalışmanın amacı, kitlelerin yapay zekâ sohbet robotuna yönelik deneyimlerini Ekşi Sözlük girdileri kapsamında sorgulayarak geleceğin iletişiminde yapay zekânın yerini tartışmaktır. Bu kapsamda, Ekşi Sözlük’te yer alan “ChatGPT” başlığından 30 Kasım 2022 – 28 Ağustos 2023 tarih aralığındaki tüm girdiler (4233 adet), Python dili ile oluşturulan algoritma ile alındı. Üçer aylık dönemlere ayrılan veri seti, metin madenciliği teknikleriyle temizlenerek duygu analizi gerçekleştirildi. Ardından dönemlik en fazla harf karakterine sahip toplamda 600 adet girdi, içerik analizi için örneklem olarak belirlendi. İçerik analizi sonrası sonuçlar, dönemsel olarak karşılaştırıldı ve yapay zekânın gelecekte, gerek profesyonel gerekse kişisel iletişim süreçlerinde kullanılabileceği bulgulandı.
Article
Full-text available
In the past five years, private companies, research institutions and public sector organizations have issued principles and guidelines for ethical artificial intelligence (AI). However, despite an apparent agreement that AI should be ‘ethical’, there is debate about both what constitutes ‘ethical AI’ and which ethical requirements, technical standards and best practices are needed for its realization. To investigate whether a global agreement on these questions is emerging, we mapped and analysed the current corpus of principles and guidelines on ethical AI. Our results reveal a global convergence emerging around five ethical principles (transparency, justice and fairness, non-maleficence, responsibility and privacy), with substantive divergence in relation to how these principles are interpreted, why they are deemed important, what issue, domain or actors they pertain to, and how they should be implemented. Our findings highlight the importance of integrating guideline-development efforts with substantive ethical analysis and adequate implementation strategies.
Article
Full-text available
We present artificial intelligent (AI) agents that act as interviewers to engage with a user in a text-based conversation and automatically infer the user's personality traits. We investigate how the personality of an AI interviewer and the inferred personality of a user influences the user's trust in the AI interviewer from two perspectives: the user's willingness to confide in and listen to an AI interviewer. We have developed two AI interviewers with distinct personalities and deployed them in a series of real-world events. We present findings from four such deployments involving 1,280 users, including 606 actual job applicants. Notably, users are more willing to confide in and listen to an AI interviewer with a serious, assertive personality in a high-stakes job interview. Moreover, users’ personality traits, inferred from their chat text, along with interview context, influence their perception of and their willingness to confide in and listen to an AI interviewer. Finally, we discuss the design implications of our work on building hyper-personalized, intelligent agents.
Article
Full-text available
Over the last decade, there has been an explosion of digital interventions that aim to either supplement or replace face-to-face mental health services. More recently, a number of automated conversational agents have also been made available, which respond to users in ways that mirror a real-life interaction. What are the social and ethical concerns that arise from these advances? In this article, we discuss, from a young person’s perspective, the strengths and limitations of using chatbots in mental health support. We also outline what we consider to be minimum ethical standards for these platforms, including issues surrounding privacy and confidentiality, efficacy, and safety, and review three existing platforms (Woebot, Joy, and Wysa) according to our proposed framework. It is our hope that this article will stimulate ethical debate among app developers, practitioners, young people, and other stakeholders, and inspire ethically responsible practice in digital mental health.
Article
Full-text available
Background: A World Health Organization 2017 report stated that major depression affects almost 5% of the human population. Major depression is associated with impaired psychosocial functioning and reduced quality of life. Challenges such as shortage of mental health personnel, long waiting times, perceived stigma, and lower government spends pose barriers to the alleviation of mental health problems. Face-to-face psychotherapy alone provides only point-in-time support and cannot scale quickly enough to address this growing global public health challenge. Artificial intelligence (AI)-enabled, empathetic, and evidence-driven conversational mobile app technologies could play an active role in filling this gap by increasing adoption and enabling reach. Although such a technology can help manage these barriers, they should never replace time with a health care professional for more severe mental health problems. However, app technologies could act as a supplementary or intermediate support system. Mobile mental well-being apps need to uphold privacy and foster both short- and long-term positive outcomes. Objective: This study aimed to present a preliminary real-world data evaluation of the effectiveness and engagement levels of an AI-enabled, empathetic, text-based conversational mobile mental well-being app, Wysa, on users with self-reported symptoms of depression. Methods: In the study, a group of anonymous global users were observed who voluntarily installed the Wysa app, engaged in text-based messaging, and self-reported symptoms of depression using the Patient Health Questionnaire-9. On the basis of the extent of app usage on and between 2 consecutive screening time points, 2 distinct groups of users (high users and low users) emerged. The study used mixed-methods approach to evaluate the impact and engagement levels among these users. The quantitative analysis measured the app impact by comparing the average improvement in symptoms of depression between high and low users. The qualitative analysis measured the app engagement and experience by analyzing in-app user feedback and evaluated the performance of a machine learning classifier to detect user objections during conversations. Results: The average mood improvement (ie, difference in pre- and post-self-reported depression scores) between the groups (ie, high vs low users; n=108 and n=21, respectively) revealed that the high users group had significantly higher average improvement (mean 5.84 [SD 6.66]) compared with the low users group (mean 3.52 [SD 6.15]); Mann-Whitney P=.03 and with a moderate effect size of 0.63. Moreover, 67.7% of user-provided feedback responses found the app experience helpful and encouraging. Conclusions: The real-world data evaluation findings on the effectiveness and engagement levels of Wysa app on users with self-reported symptoms of depression show promise. However, further work is required to validate these initial findings in much larger samples and across longer periods.
Article
We explore data from a field test of how an algorithm delivered ads promoting job opportunities in the science, technology, engineering and math fields. This ad was explicitly intended to be gender neutral in its delivery. Empirically, however, fewer women saw the ad than men. This happened because younger women are a prized demographic and are more expensive to show ads to. An algorithm that simply optimizes cost-effectiveness in ad delivery will deliver ads that were intended to be gender neutral in an apparently discriminatory way, because of crowding out. We show that this empirical regularity extends to other major digital platforms. This paper was accepted by Joshua Gans, business strategy.
Article
Objective: The aim of this review was to explore the current evidence for conversational agents or chatbots in the field of psychiatry and their role in screening, diagnosis, and treatment of mental illnesses. Methods: A systematic literature search in June 2018 was conducted in PubMed, EmBase, PsycINFO, Cochrane, Web of Science, and IEEE Xplore. Studies were included that involved a chatbot in a mental health setting focusing on populations with or at high risk of developing depression, anxiety, schizophrenia, bipolar, and substance abuse disorders. Results: From the selected databases, 1466 records were retrieved and 8 studies met the inclusion criteria. Two additional studies were included from reference list screening for a total of 10 included studies. Overall, potential for conversational agents in psychiatric use was reported to be high across all studies. In particular, conversational agents showed potential for benefit in psychoeducation and self-adherence. In addition, satisfaction rating of chatbots was high across all studies, suggesting that they would be an effective and enjoyable tool in psychiatric treatment. Conclusion: Preliminary evidence for psychiatric use of chatbots is favourable. However, given the heterogeneity of the reviewed studies, further research with standardized outcomes reporting is required to more thoroughly examine the effectiveness of conversational agents. Regardless, early evidence shows that with the proper approach and research, the mental health field could use conversational agents in psychiatric treatment.
Article
Smart speakers with voice assistants, like Amazon Echo and Google Home, provide benefits and convenience but also raise privacy concerns due to their continuously listening microphones. We studied people's reasons for and against adopting smart speakers, their privacy perceptions and concerns, and their privacy-seeking behaviors around smart speakers. We conducted a diary study and interviews with seventeen smart speaker users and interviews with seventeen non-users. We found that many non-users did not see the utility of smart speakers or did not trust speaker companies. In contrast, users express few privacy concerns, but their rationalizations indicate an incomplete understanding of privacy risks, a complicated trust relationship with speaker companies, and a reliance on the socio-technical context in which smart speakers reside. Users trade privacy for convenience with different levels of deliberation and privacy resignation. Privacy tensions arise between primary, secondary, and incidental users of smart speakers. Finally, current smart speaker privacy controls are rarely used, as they are not well-aligned with users' needs. Our findings can inform future smart speaker designs; in particular we recommend better integrating privacy controls into smart speaker interaction.
Chapter
The topic of bots is new. Back in 1966, Joseph Weizenbaum developed with ELIZA a computer program that demonstrated the possibilities of communication between a human and a computer via natural language. When replying, the machine took on the role of a psychotherapist, worked on the basis of a structured dictionary and looked for keywords in the entered text.