ArticlePDF Available

Simulation of emotions of agents in virtual environments using neural networks

Authors:

Abstract and Figures

A distributed architecture for a system simulating the emotional state of an agent acting in a virtual environment is presented. The system is an implementation of an event appraisal model of emotional behaviour and uses neural networks to learn how the emotional state should be influenced by the occurrence of environmental and internal stimuli. A part of the modular system is domain-independent. The system can easily be adapted for handling different events that influence the emotional state. A first prototype and a testbed for this architecture are presented.
Content may be subject to copyright.
Simulation of emotions of agents in virtual
environments using neural networks
Aard-Jan van Kesteren
Rieks op den Akker
Mannes Poel
Anton Nijholt
Department of Computer Science
University of Twente
Enschede, the Netherlands
e-mail : anijholt@cs.utwente.nl
Abstract
A distributed architecture for a system simulating the emotional state of an agent acting
in a virtual environment is presented. The system is an implementation of an event-
appraisal model of emotional behaviour and uses neural networks to learn how the
emotional state should be influenced by the occurrence of environmental and internal
stimuli. A part of the modular system is domain-independent. The system can easily be
adapted for handling different events that influence the emotional state. A first
prototype and a testbed for this architecture are presented.
1 Introduction
This paper presents results of ongoing research on human-computer interaction with intelligent
conversational agents in virtual environments. It reports work in a project aiming at the design and
agent-oriented implementation of a multi-user, multi-agent and multi-modal interactive
environment [Nijholt 1999].
Our interest in emotions and our objective to simulate emotional behaviour finds its primary
motivation in the hypothesis that avatars that somehow show emotions in the way they behave are
more believable than agents that lack these human qualities.
There is another motive for incorporating emotions in the design of synthetic agents; often
mentioned by AI-researchers. A. Damasio came with neurological evidence for the importance of
emotions in human decision processes [Damasio 1994]. The interaction between emotions and the
rational processes that underlie the making of decisions could possibly explain why humans are so
good in making decisions in a context of uncertainty. Answering the question whether machines can
have emotions, M. Minsky even stated: “The question is not whether intelligent machines can have
any emotions, but whether machines can be intelligent without any emotions” [Minsky 1986].
Hence the idea to give synthetic agents emotions in order to make it possible that they perform
whatever tasks they have in a more intelligent way. Whether there is some sense in this depends on
the notion of emotion one has in mind and on the analysis of the aspects involved in the process of
decision-making.
To model and to simulate all of the relevant aspects of emotions in a comprehensive system is a
project for years, so it shouldn’t come as a surprise that every group that’s conducting research on
the topic of emotional agents has made it’s own decisions as to what aspects are built into their
system, and what is left out for (hopefully) future development. Most groups [Reilly & Bates 1992,
El-Nasr & Yen 1988, Velásquez 1997] have taken a wide range of topics related to emotions (e.g.
emotion, facial expressions and emotional behaviour) into account for their emotional systems.
Until now we have only looked at the topic “simulation of emotions, as we think its best to focus
on one topic at a time.
We think of an emotional agent as an agent that has among other things (e.g. believes, desires and
intentions) an emotional state. This emotional state can be altered by stimuli from the environment
or by stimuli from internal elements of an agent (e.g. decisions, future expectations, memory, etc.).
Simulation of emotions consists of the elements that appraise the stimuli and the processes that
take care of the dynamics of the emotional state. In principle, every kind of behaviour and internal
process of an agent can make use of the emotional state. For instance, the emotional state can be
used to steer the facial expressions and the decision process. Consequently, our agents dont show
emotions yet. A user can only get an impression of what the agent feels, by looking at a
representation of the emotional state.
For our model of emotions, we have two basic concerns. We want our model to resemble the
emotional processes, as they exist within the human brain, as closely as possible and consequently
we dont want to diverge too much from the leading theories of emotions. On the other hand we
also want to develop a computational approach, which can be easily used to implement an
emotional agent.
Our research is very much ongoing and some of the elements presented in this paper havent been
properly tested yet. This paper will give an overview of how we dealt with the problem of emotions
until now and what our approach for the future will be. Furthermore an architecture for the
simulation of emotions will be introduced.
First the relevant models of emotions will be discussed shortly. After that we will globally describe
what our approach is, followed by a description of the environment that we used as a testbed.
Subsequently our distributed architecture will be introduced, by firstly explaining the architecture
as a whole and secondly by discussing the various subparts in more detail. A description of a first
prototype will be given next to further clarify the architecture and to present some first test results.
The paper will be concluded by a comparison with another model and some conclusive remarks.
2 Theories of emotion
In this section, two theories on emotion will be introduced. How we used these theories for our
model will be explained in the next section.
Event appraisal models
1
have as assumption, that its the subjective, cognitive interpretation of
events that explains the experienced emotions.
Event appraisal models mainly focus on the question how events are appraised and in which
direction (and with what velocity) the emotional state is likely to shift. We think of this as the
emotional meaning of an event. Usually event appraisal models dont model the dynamics of the
emotional state: if an event is appraised, what will then become the new emotional state? For an AI-
model of emotion this question is as important as a natural appraisal of events.
1
For instance [Ortony, Clore & Collins 1988] and [Roseman, Jose & Spindel 1990]
One of the leading event appraisal models is the model of Ortony, Clore and Collins (The OCC
Model) [Ortony, Clore & Collins 1988]. The model only looks at emotion types. Each type contains
a large number of emotional states, which may differ in intensity (e.g. the emotional states: worried,
scared and terrified belong to the emotion type fear), but may also differ because of other reasons,
such as the cause of the emotion (e.g. the emotional state heartache differs from other emotional
states belonging to the emotion type distress). By focusing on emotion types instead of emotions,
the whole field of emotions gets more comprehensible.
In order to appraise a particular event, there are three different aspects that can be focused on. Every
aspect has a different set of emotion types attached to it. We will elaborate a bit on the different
aspects by means of an example. Imagine that someone hears that his sister-in-law has just killed
one of his children. If that someone focuses on the consequences of the event, hell probably
experience distress. If he focuses on the action of the agent (the sister-in-law), hell probably
experience reproach. And if he looks at the aspects of an object (aspects of the sister-in-law) hell
probably experience hate.
Associated with each aspect is a different kind of knowledge representation. If one focuses on the
consequences of an event, goals are of importance. If one focuses on the action of an agent,
standards are of importance and if one focuses on the action of an agent, attitudes are relevant.
A strong point of the OCC model, which makes it particularly useful for designing an AI-model of
emotion, is that the model includes a complete set of variables that influence the intensity of the
emotions. The variables can be global, which means that the variable influences all of the emotion
types, or local, which means that the variable influences only some of the emotion types. We will
shortly introduce the (in our opinion) three most important local variables. The variable desirability
is important if one focuses on the consequences of an event. The variable praiseworthiness is
important if one focuses on the actions of an agent and the variable appealingness is important if
one focuses on the aspects of an object.
Besides the event appraisal models, there are more theories of emotions we found interesting and
useful. One of those is Frijdas notion of emotion [Frijda 1986]. According to him, emotion is
action readiness change, which refers to the inner disposition (and the absence) for performing
actions and achieving relational change, as it exists at a particular moment. The experience of
emotion largely consists of experienced action readiness or unreadiness: an impulse to flee, strike or
embrace; or lack of impulse, apathy, listlessness. So, emotion is not the same as the feeling of
emotion. A consequence of this is, that an explicit representation of emotions is unlikely to exist.
3 Our approach
Our research goal is to develop a human like agent that shows natural emotional behaviour. Yet, as
we didnt want to start too ambitiously we are firstly focusing on simple agents in a simulated
environment. The agents and the environment shall be introduced in the next section.
Our model is very much an implementation of the more theoretical OCC model. First of all, only
events can alter the emotional state. Furthermore, the representation of emotion consists of the same
emotion types the OCC model distinguishes, the three aspects of events also play a central role in
our model and the same variables (the properties of events that are important for emotions) as in the
OCC model will be used. Furthermore, the agents will contain a representation of goals, standards
and attitudes.
The current state of research on emotions is, that a lot of work has been done on philosophical
questions like: What are emotions?, What influences emotions?, Which emotions can be
distinguished? and What is the connection between cognition and emotion?, but no complete
quantitative models of emotions (besides some AI-models) have been developed yet. Apparently
its very difficult to describe the emotional behaviour in a complete and precise manner. On the
other hand we noticed that its relatively easy to imagine in practical situations what would be
natural behaviour.
Therefore we chose an approach in which the system has to learn about emotional processes and
rules from examples provided by a trainer. Most other research groups have used a completely
opposite approach. In their systems the designer has to define the emotional processes and rules,
from which natural emotional behaviour has to originate. Well use the common terms top-down
approach for our approach and bottom-up approach for the other approach (Figure 1 gives a
schematic representation of the two concepts).
Emotional behaviour
Emotional processes & rules
top-down
bottom-up
Figure 1 A top-down approach vs. a bottom up approach
For our approach trainingsdata is needed, which will be obtained through annotation. We realize
that annotation is a time-consuming and tedious task and therefore we defined the architecture in
such a way, that a minimal amount of trainingsdata is required.
Another advantage of a top-down approach above a bottom-up approach is, that its more flexible;
the same system can be used for more applications, which is for instance practical if one wants to
design agents with different personalities. Using a different trainingsset for each personality is an
easy way to do this. Of course the same effect may be obtained in a bottom-up approach by
changing some of the parameters, but we strongly doubt if parameters can be defined in such a way
that it offers the same flexibility as a top-down approach does.
As we want to design a system that only needs a small amount of trainingsdata, its especially
important that the system is able to generalize well. This is one of the proven qualities of neural
networks, and accordingly we will use those for our system. For the emotion domain, it isnt very
important that the system works very accurately in a quantitative way (the chosen numerical
representation of emotions is a rough estimate anyway), as long as it performs well in a qualitative
way and reasonably well in a quantitative way. Therefore relatively small neural networks can be
used, which should enhance the generalization capacities of the system even more.
As Frijdas theory of emotions clearly suggests, an explicit representation of emotions is unlikely to
exist within the human brain. We do not only think that an explicit representation of emotions is
unlikely from a cognitive point of view, but we also think that a wrongly chosen representation
might have a strong negative effect on the capacities of the system to show a wide range of
emotional phenomena. By using recurrent neural networks, the system can learn an optimal,
implicit representation of emotions itself, which we see as a major advantage of a connectionist
approach.
4 The environment and the agents
In order to develop and test a system that simulates emotions, an environment is needed, which is
inhabited by agents that can have emotions. As we want to model a broad range of emotions, the
environment and the agents must satisfy a number of requirements.
As we make use of the OCC model as basis for our AI-model of emotion, its important that the
agents have an explicit or implicit representation of goals, standards and attitudes
1
. Agents should
also be able to translate observations into terms of these three concepts. In practice this last
requirement has the following consequences:
Agents should be able to see if an event satisfies a particular goal or has a positive or
negative effect on the probability that a particular goal will be satisfied (important for the
emotion types joy and distress).
Agents must be able to reason about the future and should be able to change their
expectations of the future as a consequence of events. Expectations about goals are
important for emotion types such as hope and fear.
Agents must have some kind of memory about previous expectations and should be able to
compare new events to these previous expectations (important for the emotion types relief,
fears-confirmed, satisfaction and disappointment).
The same kind of reasoning the agents must be able to do for themselves, they must be able
to do for other agents also (important for emotion types happy-for, resentment, gloating and
pity).
Agents must be able to compare actions of themselves or of other agents to their standards
(important for emotion types pride, shame, admiration and reproach).
Finally, agents must be able to compare aspects of objects or agents to their attitudes
(important for emotion types love and hate).
The future must be partly uncertain for the agents. If its precisely known whats going to happen in
the future, emotions such as hope and fear cannot exist. A way to guarantee this is by making the
environment nondeterministic. Another important property of the environment is, that it contains
multiple agents, as some emotions depend on actions of other agents or on the consequences of
events for other agents.
As the user has to annotate natural emotional behaviour, its important that the user is able to put
himself in the position of an agent. A user can only do this, if the behaviour of the agent is
believable enough and if the user has the exact same knowledge as the agent. A final requirement of
the environment is, that the situations are complex enough, such that interesting emotional
behaviour can arise.
We set out to define and implement an environment, which satisfies all the previously stated
constraints. For the domain we were inspired by [Inoue, Kawabata & Kobayashi 1996]. In Figure 2
a picture of the domain can be seen.
1
We refer to [Reilly & Bates 1992] for a good example of an implementation of these concepts.
Figure 2 Picture of the environment
The domain is a gridworld containing grass, water pools that can be dry or contain water, apple
trees that can have apples growing and rocks, with possibly herbs growing on it. The status of the
trees, water pools and rocks constantly changes in a nondeterministic way.
Multiple agents inhabit the gridworld. One of the agents can be seen in the middle of the picture. An
agent can only see a small part of the world (the light part). Currently the agent sees one predator
(directly beneath the agent) and one tree with an apple (directly south-east of the predator). An
agent only knows where the visible agents and predators are and he knows the location of all the
trees, rocks and water pools, but the status is only known of the visible trees, rocks and water pools.
A trainer can only see the things an agent can see; in the light part, a trainer can see everything, just
like an agent, and in the dark part, a trainer sees only the locations of the trees, rocks and water
pools, as this is also knowledge that an agent possesses. Because of this, its easier for a trainer to
imagine what an agent feels. A trainer also knows how hungry, thirsty and healthy an agent is,
which events have occurred lately and what his current action is.
Agents need food and water, which can be supplied by apple trees and water pools. There are
predators that can be dangerous for the agents. An attack by a predator affects the health of an
agent. An agent can regain health by eating an herb.
To a large extent, the behaviour of a predator is random. The only exception is when a predator is in
the neighbourhood of food, water or an herb. Then it will stay there because it knows that agents
will come there sooner or later. This means that the presence of a predator is an indication that food,
water or an herb may be nearby. On the other hand the presence of food, water or an herb is an
indication that a predator may be nearby.
Agents are able to make decisions and have all the previously described capacities. There is a social
order between agents and they can choose (dependent on their character) either to follow the leader
of a group or to go their own way. Social grouping is a result of common concerns. An agent knows
how thirsty, hungry and healthy the agents in his group are.
5 The system
5.1 The architecture
In this section, the architecture for our system will be introduced. First a global overview of the
system will be given and after that the various concepts and subparts will be discussed in more
detail. For the system we have chosen a distributed architecture, as depicted in Figure 3. The
motivations for this particular architecture will be given after a brief introduction.
Emotional
State
Calculator
N
E
IV
N
E
I
V
N
E
IV
Event
E
v
e
n
t
I
n
f
o
E
v
e
n
t
I
n
f
o
E
v
e
n
t
I
n
f
o
...
Normalizer 1
Event
Appraiser 1
EIV
Event type 1
Normalizer 2
Event
Appraiser 2
EIV
Event type 2
Normalizer n
Event
Appraiser n
EIV
Event type n
Emotional
State
Figure 3 Schematic representation of the architecture
From a functional perspective, the task of the system is to convert an event with particular
properties into a new emotional state. In order to convert an event to a new emotional state, two
phases are distinguished.
The first phase is to appraise the emotional meaning of the event. This is the task of an Event
Appraiser. Each event belongs to maximally one event type (e.g. the event: An agent sees an apple
at a particular location while he has a particular desire for food belongs to the Apple_spotted
event type). For each relevant event type, one event appraiser is defined. The content of the event
information depends on the type of the event and can vary greatly from one event type to another.
To sum up, when an event occurs, first the event type of the event has to be established. Then its
exactly known which Event Appraiser has to be used and what kind of information is needed for
appraisal.
In the second phase, the Emotional State Calculator (the ESC) uses a numerical representation of
the emotional meaning of the event (firstly stored in an EIV and secondly stored in an NEIV; both
will be explained in detail later on) to calculate the new emotional state. A history of previous
emotional states and emotional events is implicitly stored within the ESC, which is implemented by
a recurrent neural network. The Normalizers form an interface between the Event Appraisers and
the ESC. The task of a Normalizer is to normalize the data in an EIV, such that the ESC can treat all
of the Event Appraiser in a uniform way. The Normalizers will also be implemented by neural
networks.
This approach has a few qualities:
Conformity with emotion theory. As mentioned before, emotion theorists are inclined to
see the appraisal of events as a separate problem. In this architecture, the appraisal of events
is explicitly a subpart of the system.
Conformity with human intuition. In our first annotation attempts, we noticed that if we
had to predict a new emotional state, given an event and the old emotional state, that we
were inclined to first look only at the event and imagine what kind of emotional
consequences this event might have and only after that we looked at the old emotional state
and predicted a new emotional state. Therefore the way the system works conforms at least
to our intuition and maybe also to the intuition of most people. This is an important
property, as this means that the annotation task (which will be explained in the section on
training) will be relatively easy.
Trainability. The different subparts can be kept relatively small, as a result of which the
various neural networks can be trained more easily and with a smaller amount of
trainingsdata.
Possibility of incremental development. This approach allows that events are added to the
system one-by-one. This is a big advantage from a software engineering point of view.
Scalability. If the domain gets more complicated (meaning more types of events), the
amount of neural networks within the system will increase, but the complexity of the neural
networks will remain the same. This means that the scalability of the system shouldnt be a
problem, if the amount of types of events increases. Whether the system is scalable in other
dimensions also (e.g. the amount of modelled emotions) is one of the open research
questions remaining.
Reusability of the ESC. The ESC is domain-independent and therefore can be reused for
different agents in different domains.
5.2 Emotion Impulse Vectors
An Emotion Impulse Vector (an EIV) is a data structure especially suited for storing the emotional
meaning of an event. The structure of an EIV has been chosen as follows:
EIV = <ei
1
, ei
2
, ..., ei
n
>
in which every emotion type e
i
has a corresponding emotion impulse ei
i
. The emotion impulse ei
i
indicates how the emotion e
i
is likely to change. For instance a high positive value for ei
i
means that
the EIV probably will have a large positive effect on the emotion e
i
.
A Normalized Emotion Impulse Vector (an NEIV) is a normalized version of an EIV. How the
normalization functions, will be explained later on.
5.3 Event Appraisal
The task of an Event Appraiser is to assess the emotional meaning of an event to construct an EIV.
Every event and domain has its own properties and therefore an Event Appraiser should be
constructed separately for every event type. This way a priori knowledge about the event and
domain can be used in an optimal way. A consequence of this is, that an Event Appraiser can be
implemented by a neural network, but can just as well be implemented by a rule-based system, a
function or something else.
Defining or training an Event Appraiser that can immediately communicate with the ESC in such a
way that the behaviour of the total system is precisely as required, would mean a full understanding
of the ESC and would at least make defining or training the Event Appraiser a very difficult task.
Therefore we introduced the Normalizers. Their task is to convert an EIV to an NEIV such that the
behaviour of the total system works as required. As the Normalizers only have to do scaling
operations, and therefore only have to learn linear or near linear functions, the Normalizers can be
simple feedforward networks with only a few hidden neurons.
A nice property of using Normalizers is that the Normalizers can easily correct quantitative errors
of the Event Appraisers and the ESC. Therefore its not very important that the Event Appraisers
and the ESC work well in a quantitative way, just as long as they work well in a qualitative way.
The inputs of an Event Appraiser are the variables defined in the OCC model or properties of an
event that can be used to calculate the variables. The current action of the agent is a special kind of
property, which can be important to appraise an event. For instance, an agent is bound to appraise
the event of spotting a predator differently if hes attacking than if hes fleeing.
5.4 The Emotional State Calculator
The task of ESC is to calculate a new emotional state given a NEIV. Although the ESC is one
neural network, conceptually it can be seen as two different parts. The first part, the core of the
ESC, is a recurrent network and models all of the emotional behaviour. The state of the recurrent
network can be seen as an implicit emotional state. We will implement the implicit emotional state
with an Elman network [Elman 1990].
The second part of the ESC is a feed forward network, which has the implicit emotional state as
input. This part, the Emotional state monitor, has as task to translate the implicit emotional state,
which is very hard to understand and use to influence the behaviour, into an explicit emotional state.
The structure of the explicit emotional state has been chosen as follows:
Explicit emotional state = <e
1
, e
2
, , e
n
>
in which e
i
denotes the intensity of an emotion type. Every e
i
refers to a different emotion type.
A schematic representation of the ESC is depicted in Figure 4.
Emotional state
......
...
NEIV
...
...
Emotional state
monitor
Explicit
emotional state
Implicit
emotional state
Figure 4 Schematic representation of the Emotional State Calculator. A
dashed line represents a time delay of 1.
5.5 Decay
The decay of an emotional state is a topic that has received special attention in most other AI-
models of emotion and is also a topic mentioned in [Ortony, Clore & Collins 1988] as being a
specific point of concern for a computational model of emotion. In our system the decay is modeled
as an event type with a corresponding Event Appraiser and Normalizer. An event of this type occurs
by definition after a fixed interval of time and sees to it that the emotional state eventually returns to
a state of rest. A property that determines the appraisal of this event is among other thing the current
(explicit or implicit) emotional state.
5.6 Training the system
The system is built out of three types of building blocks. For every type of building block a
different training strategy has to be used. The order in which the building blocks have to be trained
or defined is depicted in Figure 5.
Training the
Emotional State Calculator
Constructing an
Event Appraiser
Training a
Normalizer
Adding a new Event
Figure 5 Schematic overview of the trainingsprocess
First the Emotional State Calculator has to be trained. This is the most difficult part of the system to
train, as gathering trainingsdata is difficult. A temporal sequence of (NEIV, explicit emotional
state) pairs is needed to train the Elman network. The NEIVs can be generated randomly, but the
emotional states have to be annotated by hand.
There are two problems with this approach. First of all a reasonable amount of trainingsdata is
needed and secondly the needed data is of an abstract nature, as a result of which annotating even
one sample isnt an easy task. But we feel that the benefits of this tedious work are large enough to
be profitable. If the Emotional State Calculator has been trained, adding events to the system is easy
and an Emotional State Calculator is domain-independent and therefore only has to be trained once.
The trainingsset doesnt have to be perfect, as the learning algorithm should be able to filter out the
errors. Therefore the trainer can annotate quickly without worrying too much about errors, which
makes the annotation task a bit easier. Also, it shouldnt be overlooked that a significant part of the
difficulties of training the ESC are inherent to the complexity of the problem of emotions. A nice
property of our system is that a major part of the complexity of emotions is modelled within a
domain-independent part of the architecture, so that the same problems dont have to be solved over
and over again for every domain and agent.
After the ESC works satisfactory, event types can be added one-by-one to the system. For a new
event type, an Event Appraiser has to be trained or defined firstly. Again its only important that an
Event Appraiser works well in a qualitative way, which makes defining or training an Event
Appraiser relatively easy.
From a procedural perspective, training a Normalizer is a bit less straightforward than training the
ESC or an Event Appraiser. (EIV, NEIV) pairs are needed to train a Normalizer. The user cannot be
expected to annotate NEIVs, because this would mean among other things that the user has to know
exactly how the ESC works, which is not a realistic requirement. Therefore we propose the
following approach.
Instead of (EIV, NEIV) pairs, (Event Information, Explicit emotional state, State of the ESC) triples
have to be gathered. This can be done within the real domain. Every time an event of the relevant
event type occurs, the user has to annotate the new explicit emotional state. The event information
can be extracted from the event and the current state of the ESC can be copied easily. After the
triples have been gathered, the triples have to be converted to the (EIV, NEIV) pairs. It is easy to
convert the event information to an EIV, as the Event Appraiser is already trained or defined. An
NEIV can be found by using the inverse of the ESC. If this inverse is known, the explicit emotional
state (the output of the ESC) and the state of the ESC are enough information to calculate an NEIV.
Of course the inverse is not known, but using an iterative improvement algorithm (e.g. a simulated
annealer), can solve this final problem. This algorithm has to find the NEIV that explains the
explicit emotional state the best. If needed, a second criterion can be that the NEIV has to resemble
the EIV as much as possible.
Using this approach makes training a Normalizer easy for a user. Annotating an explicit emotional
state is easy and as the Normalizers are usually small networks, only a small amount of
trainingsdata is needed.
6 A first prototype
A first prototype of the system has been implemented. In this prototype the various subparts were
kept as simple as possible, such that we could examine whether the subparts are able to cooperate
and whether the system as a whole is capable of simulating the required behaviour. One of the
consequences of this is, that the ESC has not been implemented by the proposed Elman network
yet.
In the first version only two emotion types are modelled, joy and distress, and seven event types.
The two emotion types joy and distress were considered as one emotion type, the joy/distress
emotion type. If the intensity of this emotion type is negative, the agent experiences distress and if
the intensity is positive, the agent experiences joy. So, an EIV, an NEIV and the explicit emotional
state, all consist of only one value.
The ESC has been implemented as a simple linear function (and therefore the inverse is known).
The previous emotional state is one of the variables of the function. Although this simple function
suits us well in this phase of the research, we are convinced that such a simple approach wont be
sufficient in the future and that the described approach with an Elman network is a more powerful
and a cognitively more credible approach.
The Event Appraisers were kept simple too. We will clarify this part of the architecture a bit more,
by focusing on the New_predator_spotted event type in detail. The events belonging to this event
type occur when the agent sees a predator for the first time.
As mentioned before, the variables from the OCC model are the only factors that influence the
intensity of an emotion type. Although more than one variable is defined in the OCC model, that
influence the emotion types joy and distress, only the variable desirability determines the appraisal
value in this version. The event information of the New_predator_spotted event type consists of
two properties: The current health of the agent (health) and the amount of predators the agent has
seen previously this turn (#seen_previously). If the health of the agent is low, the desirability of the
event should be lower, than if the health is high, as being near a predator while health is low is very
dangerous. The property #seen_previously should have a negative effect on the desirability of the
event, as for instance seeing a third predator is less desirable than seeing a second predator. Taking
all of this into account, the EIV for this event was defined as follows:
EIV = <desirability(health, #seen_previously)>
= <1 * health 20 * #seen_previously 100>
Only qualitative and no quantitative arguments and no were taken into account while drawing up
the formula and as a consequence the 1, 20 and 100 in the formula could for instance just as well
have been 1.5, 30 and 0.
The Normalizer is a simple feed forward network, with one hidden layer, consisting of only 2
neurons. Only 20 trainingsamples were used to train the Normalizer. The other Event Appraisers
and Normalizers were constructed in a similar way.
How to test the quality of the system is a hard question and a research project by itself. This version
was tested in an informal way. We checked whether the emotional behaviour shown by the system
corresponds to our ideas of emotions, which we also used for annotation. This was the case and
therefore we can conclude that the architecture looks promising, but a lot more experiments need to
be conducted before definite conclusions can be made.
7 Comparison with other models
Before defining our own models we studied other models. [Pfeifer 1988] gives a nice overview of
all of the AI-models of emotion until 1988. Since then, more interesting models were put forward.
[Reilly & Bates 1992] also based their model on the OCC model. In [El-Nasr & Yen 1988] fuzzy
logic was used, a very useful technique for the emotion domain we think. A system which we want
to discuss more thoroughly in this section is Cathexis [Velásquez 1997], as it resembles our model
in a couple of ways; it also has a distributed architecture and special care has been taken to allow
for maximum flexibility.
Cathexis is composed of a number of proto-specialist agents [Minsky 1986], each representing a
different emotion type. Within each proto-specialist, different sensors are monitoring both external
and internal stimuli for the existence of the appropriate conditions that would elicit the emotion
type represented by that particular proto-specialist. Input from these sensors either increases or
decreases the intensity of the emotion type. Parallel to this process, theres always a decay process
going on, which sees to it that the intensity of the emotion type eventually returns to a state of rest.
The way the sensors function and how the sensors and the decay function affect the intensity can be
completely defined by a user. Not only the sensors can influence the intensity of an emotion, but
also other proto-specialists agents can, by providing either inhibitory or excitatory input. A
schematic overview of the system is depicted in Figure 6.
Both Cathexis and our approach have distributed architectures. The difference is, that in Cathexis
the emotions are scalable and in our architecture the complexity of the domain is scalable. The
advantage of making the emotions scalable is, that it makes it easier to implement a system within a
simple domain. But, as the domain gets more complex, the approach of Cathexis will experience
scalability problems we think, as all the proto-specialists have to perform increasingly complex task
then, while no provisions have been taken to deal with this complexity.
I
I
n
h
i
b
i
to
r
y
/
E
x
c
i
t
a
to
r
y
I
n
p
u
t
I
n
h
i
b
i
t
o
r
y
/
E
x
c
i
t
a
t
o
r
y
I
n
p
u
t
I I
Inhibitory /
Excitatory
Input
Anger Fear
Distress
Figure 6 The “proto-specialist” agents for anger, fear and distress. (I:
Intensity)
In our approach its difficult to train the ESC for a system with multiple emotion types. But after
that part has been done, adding more event types is easy.
The Cathexis system allows for a large degree of flexibility. Nevertheless we dont think the
framework is powerful enough to model complex temporal emotional phenomena and phenomena
in which more than two emotions participate. Our approach does have these possibilities and
therefore is more powerful than Cathexis.
For Cathexis a bottom-up approach is used to make a system suitable for a particular domain. We
think that it will be very difficult to precisely define all the functions and parameters in such a way
that the system performs well in the case of a complex environment in which a lot of types of events
can occur. In our architecture the generalization capabilities of neural networks should make it
easier to handle complex environments and as a top-down approach is used, a full understanding of
emotional processes and the environment is not needed; the trainer only has to know in practical
cases, what the right emotional behaviour is.
8 Conclusion
An emotion theory based architecture for simulating emotions has been presented in this paper,
which can be used to develop emotional agents acting within complex environments. A first
prototype indicated that the various subparts of the distributed architecture are able to cooperate
well and that the system is able to show natural emotional behaviour.
In this first implementation, the Emotional State Calculator (the ESC) was kept very simple. For the
eventual system we have proposed a recurrent network of which the state of the network can be
seen as an implicit emotional state. A foreseen problem of our approach is training the recurrent
network, as its difficult to obtain sufficient trainingsdata. But we feel that this problem is largely
inherent to the complexity of emotions and that a nice property of our architecture is that a major
part of this complexity is concentrated within a domain-independent part of the architecture.
In the future we want to expand our prototype with more emotions and event types. Furthermore we
think that more research is needed on the ESC.
References
Damasio, A.R. (1994). Descartes Error: Emotion, Reason, and the Human Brain. New York: G.P.
Putnam.
Elman, J.L. (1990). Finding structure in time. Cognitive Science, 14, pp. 179-211.
El-Nasr, M.S., & Yen, J. (1988). Agents, Emotional Intelligence and Fuzzy Logic. Proceedings of
the IEEE NAFIPS 98, Fl.
Frijda, N.H. (1986). The Emotions. New York: Cambridge University Press.
Inoue, K., Kawabata, K., & Kobayashi, H. (1996). On a Decision Making System with Emotion.
IEEE International Workshop on Robot and Human Communication, Tokyo, Japan, pp. 461-465.
Minsky, M. (1986). The Society of Mind. New York: Simon and Schuster.
Nijholt, A. (1999). The Twente Virtual Theatre Environment: Agents and Interactions. Proceedings
of the fifteenth Twente Workshop on Language Technology, pp. 147-165.
Ortony, A., Clore, G.L., & Collins, A. (1988). The Cognitive Structure of Emotions. Cambridge,
UK: Cambridge University Press.
Pfeifer, R. (1988). Artifical Intelligence Models of Emotion. In V. Hamilton, G.H. Bower, & N.H.
Frijda (Eds.), Cognitive Perspectives on Emotion and Motivation, Netherlands: Kluwer, pp. 287-
320.
Reilly, W., & Bates, J. (1992). Building emotional Agents, Pittsburgh, PA: Carnegie Mellon
University, Technical Rep. CMU-CS-92.143.
Roseman, I.J., Jose, P.E., & Spindel, M.S. (1990). Appraisal of Emotion-Eliciting Events: Testing a
Theory of Discrete Emotions. Journal of Personality and Social Psychology, 59(5), pp. 899-915.
Velásquez, J. (1997). Modeling Emotions and Other Motivations in Synthetic Agents. In:
Proceedings of the AAAI Conference 1997, Providence, RI, pp. 10-15.
... Picard [17] proposed the hidden Ma emotion model, which can capture emotion changes exhibited by external states su changes in facial expressions, heart rate changes, voice fluctuations, and so on. Keste al. [18] established a distributed emotion state model based on neural network algori for event processing, which could convert the external emotion events into correspon emotion states by using neural network algorithms. Samani et al. [19] proposed a co site artificial emotion system, which consisted of three modules: probability-based tion matching (PLA), emotion-based artificial endocrine system (AES), and emotion transition (AST). ...
... Picard [17] proposed the hidden Markov emotion model, which can capture emotion changes exhibited by external states such as changes in facial expressions, heart rate changes, voice fluctuations, and so on. Kestern et al. [18] established a distributed emotion state model based on neural network algorithms for event processing, which could convert the external emotion events into corresponding emotion states by using neural network algorithms. Samani et al. [19] proposed a composite artificial emotion system, which consisted of three modules: probability-based emotion matching (PLA), emotion-based artificial endocrine system (AES), and emotion state transition (AST). ...
... pm pm e e e E 0 n s l n r t B = P P P (18) where ˆi pm e , ˆj pm e , and ˆk pm e represent different emotions, respectively. ...
Article
Full-text available
How to express emotions through motion behaviors of robots (mainly for robotic arms) to achieve human–robot emotion interactions is the focus of this paper. An artificial emotion expression method that accords with human emotion that can deal with external stimuli and has the capability of emotion decision-making was proposed based on the motion behaviors of robot. Firstly, a three-dimensional emotion space was established based on the motion indexes (deviation coefficient, acceleration, and interval time). Then, an artificial emotion model, which was divided into three parts (the detection and processing of external events, the generation and modification of emotion response vectors, and the discretization of emotions) was established in the three-dimensional emotion space. Then emotion patterns (love, excited, happy, anxiety, hate) and emotion intensity were calculated based on the artificial emotion model in human–robot interaction experiments. Finally, the influence of motion behaviors of humanoid robot NAO on the emotion expression of experimenters was studied through human–robot emotion interaction experiments based on the emotion patterns and emotion intensity. The positive emotion patterns (love, excited, happy) and negative emotion patterns (anxiety, hate) of the experimenters were evaluated. The experimental results showed that the personalized emotion responses could be generated autonomously for external stimuli, and the change process of human emotions could be simulated effectively according to the established artificial emotion model. Furthermore, the experimenters could recognize the emotion patterns expressed by the robot according to the motion behaviors of the robot, and whether experimenters were familiar with robots did not influence the recognition of different emotion patterns.
... A distributed architecture for a system simulating the emotional state of an agent acting in a virtual environment was presented by Aard-Jan van Kesteren, Rieks op den Akker, Mannes Poel, Anton Nijholt [3]. The system is an implementation of an event appraisal model of emotional behavior and uses neural networks to learn how the emotional state should be influenced by the occurrence of environmental and internal stimuli. ...
... The links in this layer are used to perform antecedent matching fuzzy logic rules and also the fuzzy AND operation. The link weight, w 3 i , at the third layer is also equal to unity. Layer 4: the links to this layer define consequences of the rule nodes. ...
Article
Full-text available
In a very real sense we have two minds, one that thinks (rational mind) and one that feels (emotional mind). These two fundamentally different ways of knowing interact to construct our mental life. This results in two different kinds of intelligence i.e, rational and emotional. Intellect cannot work at its best without emotional intelligence. If our aim in AI is to build systems that behave like human beings then it is necessary that we incorporate elements of both rational and emotional intelligence into the system. The main concept in this paper is to develop an Intelligent Agent that exhibits the notion of emotional intelligence. The architecture that will be considered for implementation is a Neuro-Fuzzy system with concepts of artificial intelligence incorporated into the functionality of the system. The approach chosen for the implementation is Soft Computing which is basically a synergistic integration of three computing paradigms: neural networks, fuzzy logic and probabilistic reasoning, to provide a flexible framework to construct computationally intelligent systems.
... The MIT lab is also producing an emotional multi-agent project.The model only describes basic emotions and innate reactions. A distributed architecture for a system simulating the emotional state of an agent acting in a virtual environment was presented by Aard-Jan van Kesteren, Rieks op den Akker, Mannes Poel, Anton Nijholt [3]. The system is an implementation of an event appraisal model of emotional behavior and uses neural networks to learn how the emotional state should be influenced by the occurrence of environmental and internal stimuli. ...
Conference Paper
Full-text available
Text Analytics is an interdisciplinary field which draws on data mining, machine learning, information extraction, statistics and computational linguistics.This paper deals with the development of an IntelligentAgent that exhibits the concept of emotional intelligence and which can be used for Text Analytics. Text Analytics is the most recent name given to Natural Language Understanding, Data and Text Mining. The architecture used is a Neuro-Fuzzy system which is a fuzzy system that uses a learning algorithm derived from or inspired by neural network theory to determine its parameters (fuzzy sets and fuzzy rules) by processing data samples. The input to the system is a real life event which is tokenized and the tokens are compared to a pre-defined database of emotional keywords. Based on the matching, the tokens are processed by the neuro-fuzzy controller and the corresponding emotion underlying the event is generated.The system can be used to analyze the text content in emails, blogs, tweets, forums and other forms of textual communication.
... A distributed architecture for a system simulating the emotional state of an agent acting in a virtual environment was presented by Aard-Jan Van Kesteren, Rieks Op Den Akker, Mannes Poel, Anton Nijholt [1]. Devillers et al. [5] found the most appropriate emotional state by calculating the conditional probability between the emotional keywords and emotional states. ...
Chapter
Full-text available
Emotion recognition is an important aspect of Affective Computing. This paper deals with the development of an intelligent agent for automatic recognition of emotion from text based events. The approach chosen is Soft Computing and the architecture used is a Neuro-Fuzzy system. The input (event) string is divided into tokens which are then compared to a standard corpus (i.e.,Wordnet-Affect) of emotional keywords. The computed values of emotional weight and polarity are then processed by a Neuro-Fuzzy Controller, which generates the emotion underlying the event. The system considers Ekman’s six basic emotions {Happiness, Despair, Disgust, Fear, Anger, Surprise}. The controller is trained to generate correct output through the backpropagation algorithm. The system is implemented using Java, and, Matlab is used for mathematical analysis. The performance of the system is graded based on the measures of Precision and Accuracy.
... This model is based on the study of Ard-jan van Kesteren et al. [15], and sets up a distributed model aim at behavior decision-making (Fig. 3), The whole distributed system converts specific behavior decision-making into the corresponding active field, process is divided into two stages: The first phase evaluates the emotional significance of behavior by Behavior Appraiser, for each type of behavior defined a corresponding Behavior Appraiser. When agent encounters with external behavior stimulus, first determines the behavior type and information, and then selects the relevant Behavior Appraiser to evaluate emotion and classify behavior stimulus. ...
Article
Emotional intelligence is an important branch of artificial intelligence, it is an important way to achieve the truly intelligent robot. This paper regards an Agent living in a virtual environment as the research object and proposes a fuzzy emotion model based on particle system, and combines the feedback signal of emotion with associative learning mechanism based on CMAC neural network. Finally, combined the environment, emotion model and CMAC network we propose a new Agent's behavior decision-making model based on artificial emotion, the results show that this model can improve the performance of the agent.
... studying the expressions of emotions by our talking faces and otherwise embodied agents, we are also studying computational models of emotions. An architecture for a system simulating the emotional state of an agent that acts in a virtual environment was constructed ([6]). It is an implementation of an event-appraisal model of emotional behaviour by Ortony, Clore and Collins ([11]). ...
Article
Full-text available
In this paper, we provide an overview of part of our experience in designing and implementing some of the embodied agents and talking faces that we have used for our research into human computer interaction. We focus on the techniques that were used and evaluate this with respect to the purpose that the agents and faces were to serve and the costs involved in producing and maintaining the software. We discuss the function of this research and development in relation to the educational programme of our graduate students.
Article
Aiming at the problems of lack of background knowledge and the inconsistent response of robots in the current human-computer interaction system, we proposed a human-computer interaction model based on a knowledge graph ripple network. The model simulated the natural human communication process to realize a more natural and intelligent human-computer interaction system. This study had three contributions: first, the affective friendliness of human-computer interaction was obtained by calculating the affective evaluation value and the emotional measurement of human-computer interaction. Then, the external knowledge graph was introduced as the background knowledge of the robot, and the conversation entity was embedded into the ripple network of the knowledge graph to obtain the potential entity content of interest of the participant. Finally, the robot replies based on emotional friendliness and content friendliness. The experimental results showed that, compared with the comparison models, the emotional friendliness and coherence of robots with background knowledge and emotional measurement effectively improve the response accuracy by 5.5% at least during human-computer interaction.
Article
Improving the interaction between artificial systems and their users is an important issue in artificial intelligence. In current trends in social robotics, this is accomplished by making such systems not just intelligent but also emotionally sensitive. Therefore, artificial emotional intelligence (AEI) is focused on simulating and extending natural emotion (especially human emotion) to provide robots with the capability to recognise and express emotions in human–robot interaction (HRI). In this treatise, we present an overview of the advances made in AEI by highlighting the progress recorded in the areas of emotion classification, emotional robots, and emotion space modelling in HRI. This review is aimed at providing readers with a succinct, yet adequate compendium of the progresses made in the emerging AEI sub-discipline. It is hoped this effort will stimulate further interest aimed at the pursuit of more advanced algorithms and models for available technologies and possible applications to other domains.
Article
Traditional interactive evolutionary computing approaches are usually susceptible to limited searching ability and human’s strong subjectivity. In response, by extending a traditional Belief-Desire-Intention (BDI) structure, a kind of affective learning agent which can perform affective computing and learning activities in human-computer interaction environment is explicitly introduced. In solving human-computer interactive multi-objective decision-making problems whose objectives are usually far from well structured and quantified, this kind of agent may help reduce human’s subjective fatigue as well as make decisions more objective and scientific. Specifically, a conceptual model of the agent, affective learning-BDI (AL-BDI) agent, is proposed initially, along with corresponding functional modules to learn human’s affective preference. After that, a kind of high level Petri nets, colored Petri nets are employed to realize the components and scheduler of the AL-BDI agents. To exemplify applications of the approaches, test functions are suggested to case studies, giving rise to satisfied results and showing validity of the contribution.
Article
This paper describes how personalities influence the behaviour of emotional agents. Two different emotional models are discussed, namely the OCC emotion model and the emotion-connectionist model. In both models personalities changes the behaviour on a different way. This paper also presents an implementation of a prototype to evaluate the performance of the different personalities in the emotion-connectionist model, whose testing results are discussed. Medium susceptible to all emotions showed to perform best.
Article
Full-text available
A recent theory by I. J. Roseman (1979, 1984) attempts to specify the particular appraisals of events that elicit 16 discrete emotions. This study tested hypotheses from the latest version of the theory and compared them with hypotheses derived from appraisal theories proposed by M. B. Arnold (1960) and by K. R. Scherer (1988), using procedures designed to address some prior methodological problems. Results provided empirical support for numerous hypotheses linking particular appraisals of situational state (motive-inconsistent/motive-consistent), motivational state (punishment/reward), probability (uncertain/certain), power (weak/strong), legitimacy (negative outcome deserved/positive outcome deserved), and agency (circumstances/other person/self) to particular emotions. Where hypotheses were not supported, new appraisal–emotion relationships that revise the theory were proposed. (PsycINFO Database Record (c) 2012 APA, all rights reserved)
Conference Paper
Full-text available
Emotions were proven to lead an important role in human intelligence. Intelligent agents' research produced many emotional agents. Research on human psychology had long considered the notion of an emotion (e.g., happy) to be a matter of degree; however, most existing research on emotional intelligent agents treat emotions as a black-and-white matter. We are proposing a model called FLAME-Fuzzy Logic Adaptive Model of Emotions. FLAME was modeled to produce emotions and to simulate the emotional intelligence process. FLAME was built using fuzzy rules to explore the capability of fuzzy logic in modeling the emotional process. Fuzzy logic helped us in capturing the fuzzy and complex nature of emotions. Throughout the paper we try to point out the advantages of using fuzzy modeling over conventional models to simulate a better illusion of reality
Article
Artificial Intelligence (AI), as we use the term here, is part of Cognitive Science. Cognitive Science is, by definition, the interdisciplinary study of cognition. Participating disciplines are AI, psychology (in particular, cognitive psychology), linguistics, philosophy, and neurobiology. The study of emotion not only includes cognitive aspects but also physiological and expressive ones, as well as subjective experience. Thus emotion is a field for cognitive science research par excellence. If several disciplines have to work together there must be a common language. This language is provided by AI. One of the major and most basic research strands in AI for the last 30 years has been the problem of knowledge representation. Although there is no consensus on how knowledge should be represented, there are many approaches which seem well-suited for our purposes. The concepts of goals, plans, or complex knowledge structures serve as useful metaphors for understanding emotion.
Article
Time underlies many interesting human behaviors. Thus, the question of how to represent time in connectionist models is very important. One approach is to represent time implicitly by its effects on processing rather than explicitly (as in a spatial representation). The current report develops a proposal along these lines first described by Jordan (1986) which involves the use of recurrent links in order to provide networks with a dynamic memory. In this approach, hidden unit patterns are fed back to themselves; the internal representations which develop thus reflect task demands in the context of prior internal states. A set of simulations is reported which range from relatively simple problems (temporal version of XOR) to discovering syntactic/semantic features for words. The networks are able to learn interesting internal representations which incorporate task demands with memory demands; indeed, in this approach the notion of memory is inextricably bound up with task processing. These representations reveal a rich structure, which allows them to be highly context-dependent, while also expressing generalizations across classes of items. These representations suggest a method for representing lexical categories and the type/token distinction.