Conference PaperPDF Available

Autonomous cognitive developmental models of robots-a survey

Authors:

Figures

Content may be subject to copyright.
2048978-1-5386-3524-7/17/$31.00 ©2017 IEEE
Autonomous cognitive developmental models of
robots-A Survey
Ke Huang, Xin Ma*, Guohui Tian,Yibin Li
School of Control Science and Engineering
Shandong University
Jinan, China
huangke0124@163.com, maxin@sdu.edu.cn, liyb@sdu.edu.cn
Abstract—This article intends to provide an overview of the
state of art in developmental models of cognitive robots. With the
development of artificial intelligence, robots have been able to
perform a variety of complex tasks controlled by human.
However, it is still a challenge for robots that they can explore
and develop their cognitive ability in the specific environment
like human beings. The current researches and experiments
endow robots with some cognitive abilities such as identifying
their own bodies, exploration behaviors, forming concepts etc.
We propose to analyze patterns of robotic cognitive development
and models in cutting-edge researches, and discuss advantages
and disadvantage of these models. Consequently, we can draw
some conclusions that may guide a way for future research.
Keywords—autonomy; robotics; learning; cognitive pattern;
development models
I.
I
NTRODUCTION
In recent years, the huge leap in the data processing speed
has contributed to the rise of high-tech sector, such as Big Data
and cloud computing. It also revitalizes the artificial neural
network which once fell into the bottleneck of development
due to the limited computing power. Artificial intelligence (AI)
has been booming. In the AI, the important direction is
cognitive robots which possess the ability of autonomous
cognitive development. As Aly said, cognitive robots can
perceive the environment, act, and learn from experience to
adapt their generated behaviors to interaction in an appropriate
manner [1]. The significance of cognitive robots is not only to
establish artificial cognitive system that can interact effectively
with humans, but also to help to verify the results in
developmental psychology and neuroscience [2].
At present, the relevant researches have made great
achievements. Nakajo et al. [3] let the robot accurately identify
their own body parts from the external environment. Meola et
al. [4] referred to the interplay of rhythmic and discrete
movements. And the model allowed robots to learn exploration
and develop the basic actions, such as reaching and grasping.
Gemignani et al. [5], Olier et al. [6] and Aguilar et al. [7]
established different conceptual representations for robots to
learn and identify objects. These abilities are the prerequisites
for robots’ continuous development. Robots could learn
language and implement complex behaviors in [8, 9, 10], and
even developed simple social skills in [11]. Developing the
ability of cognition requires establishing an appropriate model.
Vernon et al. [2] analyzed many studies and summarized ten
desiderata for developmental cognitive architectures to guide
for the model designing. However, the cognitive structure is
not a simple combination of several desiderata or functional
modules. It is a complex system that generates and coordinates
different cognitive functions. The current researches
established developmental models based on the theories of
psychology, neuroscience and computer science to develop
different cognitive behaviors of robots.
Thereby this paper intends to provide a deeper insight in the
developmental models for cognitive robots. We present a
subjective analysis for current researches to seek a more
appropriate modeling approach. The rest of paper is structured
as following: Section 2 presents a detailed analysis of
developmental patterns, Section 3 provides a summary of the
current methods to establish developmental model, Section 4
discusses the two parts above, and Section 5 draws conclusions
that may suggest future successful research endeavors.
II. C
OGNITIVE
D
EVELOPMENTAL
P
ATTERNS
The human cognitive development is an emergent process.
People interact with other persons and the environment to
accumulate their knowledge. Gradually, they can develop their
own actions and the ability to predict events [2]. However, the
artificial intelligence is created by human. Therefore, the
cognitive developmental pattern of robots may have many
forms. The pattern not only provide a concrete framework for
establishing developmental models, but also define explicit
assumptions of cognitive models [2].There are two patterns of
current cognitive researches.
One pattern is on the basis of pre-defined knowledge or
rules. The knowledge which robots learn is gradually
increasing in line with special structures or meanings. For
example, Aguilar et al. [7] defined the knowledge structure as a
combination of object’s characteristics, emotion, expectation,
and action. The environment representation described by
Gemignani et al. [5] has two parts. The first part is a
knowledge base. The second is defined in the form of semantic
grids. Besides, Prak et al. [8] predefined the rules for action
switching in the experiment. This approach is easier to
implement. However, the significant drawback is that it limits
the learning ability for robots. For instance, robots can only
This work is supported by the National High-tech Research and
Development Program 863, CHINA (2015AA042307), Shandong Provincial
Scientific and Technological Development Foundation, CHINA
(2014GGX103038), Shandong Provincial Independent Innovation &
Achievement Transformation Special Foundation, CHINA
(2015ZDXX0101E01), the Fundamental Research Funds of Shandong
University, CHINA (2015JC027), and Mount Tai Scholarship of Shandon
g
Province.
2049
learn in a particular situation or complete a specific task.
Therefore, robots developing in this pattern are not completely
autonomous.
The other is emergent paradigm. Olier et al. [6] thought that
the representation of the concept should be the result of the
action-environment coupling and dynamic, rather than
classification of objects defined by humans. This pattern is
closer to the way of human developing. Robots can develop
autonomously from a primitive state to a fully cognitive state
over time. The internal and external motivations drive robots to
explore and learn the world. They can perceive, act, adapt and
anticipate. Meanwhile, robots continually accumulate the
knowledge as their own experience [2]. It allows robots to
develop different cognitive behaviors without too many
restrictions.
We describe the whole process of cognitive development
for the second pattern of robots in Fig. 1. Robots develop from
a primitive state. Just like infants’ instinct, robots should be
endowed with some necessary attributes before development.
These attributes include robots’ interest, attention, affect,
knowledge structure etc. When robots begin to develop, they
firstly identify their own bodies to distinguish themselves from
the environment. Then driven by motivations, robots learn in
the environment though perceiving with multimodality and
exploring with actions. And the results of learning are used to
form or update knowledge and skills which will be stored in
memory. At the same time, these knowledge and skills can also
be reused in the learning process. After development, robots
can autonomously interact with environment. As for the way of
learning, we have summarized several models in the Section 3.
Fig. 1. Procession of cognitive development of robots.
III. M
ODELS
A
ND
A
PPROACHES
The development model endows robots with learning
capacity and determines their learning method. That is to say, it
determines the way of representing knowledge and learning
from the interaction with environment. Thus the development
model plays an important role in robots’ cognitive development.
It is established according to learning mechanisms. One sort of
cognitive ability may have different models based on different
learning mechanisms, and the same model can also develop
different cognitive abilities. Considering this difference, we
divide the models of current researches into the following types
according to learning mechanisms and Fig. 2 classifies the
articles mentioned in this paper. In the following sub-section,
we will elaborate these different types of models.
Fig. 2. The classification of learning mechanismes for development models
A. Application of the infant's cognitive development theories
Inspired by the cognitive process of infants, some
researches applied development theories into robotics. From
these theories, researchers can understand the way of infants’
building their structures of knowledge and developing their
behaviors, language and other complex skills. Then let robots
learn like human infants.
A famous theory referred by many studies is that infant
cognitive development theory, which was proposed by Piaget.
In this theory, learning process of infant has four stages [12]. Li
et al. [12] implemented the first three stages in real robot. In
the sensorimotor stage, they chose target object which has the
largest information entropy in visual field. Then they combined
color histogram, texture histogram and location into sample-
based representation. In pre-operational stage, the robot used
support vector machine (SVM) to form single symbolic
representation based on the sample representation. And latent
support vector machine (LSVM) was used to build symbolic
representation with multiple features in the concrete
operational stage. Finally robots could develop an ability to
identify object. While Aguilar et al. [7] focused on the first and
second sub-stages of the sensorimotor stage and stipulated the
interested object according to Piaget theory. The structure of
knowledge consists of many schemas which contain the sense
of object and related action. And the computational model is
Developmental Engagement-Reflection (Dev E-R). Namely
robot automatically generates knowledge in the engagement
phase and then analyzes, evaluates, processes the unknown
situations in the reflection phase. The whole experiment was
executed in a virtual scene. The robot could decide how to
2050
operate an object according to their own emotions and produce
cognitive curiosity. These two articles are both able to let robot
identify object accurately though exploration actions, but the
idea of development is different. In the first one, the
representation of knowledge has evolved from simple to
complex and from concrete to abstract during the development.
In the second article, the evolutionary method is expanding and
modifying knowledge constantly. We think that we should
combine these two ideas. In other words, the development of
knowledge should start with simple and concrete forms. After
accumulated to a certain quantity, the representation of
knowledge will realize a qualitative leap and become complex
and abstract. However, transforming these two processes
automatically by robots is still an unsolved question.
Other researches tend to develop robots’ behaviors on the
basis of infant’s action development. Luo et al. [13] researched
the development of robots’ basic abilities according to the
Corbetta’s new psychology discovery. When robot randomly
moved its arms like infant’s babbling actions, they used auto-
encoder to learn the sense of joints and arm form the original
raw data. Then a feedforward neural network (FNN) was
employed to map the sense of arm joint angles to the sense of
arm position and orientation. And the reverse direction
structure finally chose hierarchical inverse model (HIM) after
compared with other two different models. Thus robot could
imitate the three stages of infant reaching process. Nishide’s
research [14] also established the relationship between visual
information and proprioception through babbling which
inspired by the learning process of drawing. Then it used
Multiple Timescales Recurrent Neural Network (MTRNN) to
learn action sequences. The main idea of establishing
movement developmental models is mapping joint angles to
joint position. Then robots learn action primitive though
exploring in the environment by themselves or being taught by
human partners. When encountering an unknown situation,
robots can produce new actions based on accumulated
experience. According to Schillaci et al. [15], there are two
methods of generating exploration behaviors. The first one is
randomly motor babbling, the other one is goal-direction
exploration, such as curiosity. However, goal-direction
exploration behaviors are more efficient than random
exploration strategies. It is also an important direction for
researchers to find out how to generate goals to drive actions.
Indeed, there have been many studies devoted to this direction.
B. Learning from sensorimotor experience
Robots acquire knowledge of the world in the interaction
with environment. Some studies used sensorimotor experience
to establish developmental model, and predicted the next action
by historical behaviors. . For instance, MTRNN [9, 16] is
widely used in the development of behaviors and language. In
this model, sub-network with different time constants can self-
organize the characteristics with different time scales. The
network inputs historical visual information and proprioception,
and predicts the internal state and ontology perception at the
next time. The model controls the robot’s next action based on
the predicted value. Backpropagation through time (BPTT)
algorithm is used to get the optimal model though minimizing
the prediction error in training process [16]. Park et al. [8] built
a model with four time scales sub-network. ,n this model, the
fast dynamic sub-network learns to encode action primitives,
and slow sub-networks connect these action primitives to
produce complex actions. Tain [9] came up with a recurrent
neural network with parametric biases (RNNPB) and
parametric biases (PB) represents intentional variable in this
network. Top-down process generates actions on the grounds
of intention, and bottom-up process minimizes the prediction
error to update the network’s weight and intent state. Two
processes continually interact with each other and can achieve
to switch intentions of actions. What’s more, an association
study with language and behavior can be implemented by
binding the PB value in language RNNPB model and
behavioral RNNPB model. This idea can be applied into the
study of multimodal fusion. Besides, Yamada et al. [17]
utilized MTRNN to integrate language and behaviors. But
RNN is more suitable for processing linguistic sequences or
action sequences [8]. In order to make better use of visual
image information, Hwang et al. [18, 19] proposed a Visuo-
Motor Deep Dynamic Neural Network (VMDNN) combined
the advantages of different neural networks. They used
MTRNN to deal with the action sequences and control
attention. In the meanwhile, they used Multiple Spatio-
Temporal Neural Network (MSTNN) to process spatio-
temporal image information. One of the significant advantages
of this model is selecting a suitable network for different
modalities.
Another model is deep generative model which combines
the inference ability of the probability model and the
generalization ability of the neural network. Olier et al. [6]
used internal activation in MTRNN as probabilistic states and
used the prediction error as input to generate actions. Besides, a
core part of the model adopts variational recurrent neural
networks (VRNN). The VRNN adds temporal dependencies to
variational auto-encoders to address the dynamic updates of
internal states. The advantage of this model is obviously.
Robots can independently explore the environment without any
prior knowledge, and obtain a dynamic conceptual
representation in the form of an action-environment coupling.
However, during the learning process, there are no internal
demands and goals to drive robot move and interact with the
environment.
Sensorimotor experience can also allow robots to develop
an ability to identify their bodies. It needs robot forms body
representations. Visuo-motor coordination is widely used.
Nakajo et al. [3] came up with a stochastic continuous time
recurrent network (S-CTRNN). This network is arm at
predicting uncertainties of the sensory inputs and computing
the correlation coefficients between the moving joints and the
object in the view. Saegusa et al. [20] also judged the
correlation between visual motion and proprioception. The
model can identify multiple body parts and even the body
extended by tools. Schillaci et al. [15] summarized a number
studies showing other methods, such as using linguistic labels
to body postures, visual and auditory modalities, self-
organizing of multiple modalities etc.
2051
C. Simulation of brain mechanism
The brain is the organ controlling all the cognitive activities
of humans. There have been many studies about brain
mechanism. Some researches applied some achievement of
brain in cognitive robots. The computing models are mainly
based on brain’s working principle and memory respectively.
Inspired by perceptual control theory, Franchi et al. [21] put
forward the Intentional Distributed Robotic Architecture
(IDRA) to imitate the principle of three brain areas. This model
is an open network deliberative modules (DM) composed by a
working-memory (WM) and a goal-generator module (GM).
WM represents cerebral cortex, and it used to receive
information from other DMs. What’s more, WM can
implement unsupervised categorization, and return the cortex
activation in response to actual sensorial input. GM acts as
thalamus and can general new objectives though Hebbian
learning. Then instincts module (IM) performs as amygdala,
and broadcasts its signal to DMs. These modules cooperate
with each other and autonomously generate new goals to
promote robot’s cognitive behaviors development. Hwang et
al. [18, 19] utilized Prefrontal Cortex (PFC) network to
integrate MTRNN and MSTNN. Hwang et al. [18, 19] utilized
Prefrontal Cortex (PFC) network to integrate MTRNN and
MSTNN. The PFC network is similar to human prefrontal
cortex. The model uses visual information to generate action
and control to switch attentions. That is to say, it can achieve
the coordination of cognitive processes. In the aspect of brain
memory, Salgado et al. [22] analyzed the principle of long and
short time memory and then built the Multilevel Darwinist
Brain (MDB) evolutionary cognitive architecture. It has three
asynchronous time scales including execution, learning and
knowledge consolidation. In learning process, World Model of
the external perception and Internal Model of the internal
perception generate new behaviors. The system selects the
behavior that can maximize the value of the satisfaction in
Satisfaction Model. The best models are saved as experience in
long-time memory. In execution scale, the system will execute
the most appropriate behavior selected from memory and the
new generated. There is an assessment structure that helps to
make the best action strategies. A further research of Salgado et
al. [23] used a simplified MDB model added by predictive
model and the motivational model. Actions are generated under
the guidance of internal motivation.
At present, many unsolved secrets also exist in brain, and
the study on the brain mechanism is one of the hot research
fields now. The significance of this kind of model can not only
promote the cognitive development of the robot, in turn, the
robot cognitive development research will also help researchers
understand the working principle of the brain.
D. Other Approaches
The above researches have made great achievements in the
direction of robots’ cognitive development, and most of the
approaches mentioned above focused on Neural Network.
Nevertheless, a few exceptions adopted different thoughts in
parallel with or inspired by these endeavors. For example, in
robot cognition, Ramik et al. [24] considered curiosity as a
drive for knowledge acquisition on both perceptual and
semantic level. The system was based on visual saliency
principle and salient objects’ detection. It used a multi-layer
perception structure to control visual attention parameter. Then
the human gave the utterance of salient objects. They searched
an appropriate belief value through genetic algorithm to
minimize the difference between the robot’s interpretation and
the utterance given by the human. Robots combined
observations and the interpretation as the knowledge of the
world. According to the similar idea, Gemignani et al. [5]
presented a method for incremental online semantic mapping
which let robot understand the surrounding environment. They
built a four-layered environmental knowledge representation.
In this representation, the environment was expressed as grids
and the objects were marked with symbols. Robots extracted
the feature of each object by image processing algorithm. At
the same time, humans labeled objects and areas observed by
robot through natural language. However, the limitations of
these two approaches are the pre-defined knowledge, rather
than acquired by the robot completely. Taniguch et al. [25]
proposed nonparametric Bayesian spatial concept acquisition
method, which allowed the robot to obtain toponym from the
utterance of the sentence and use the acquired spatial concept
to reduce the uncertainty of the self-localization effectively.
Compared with the first two methods, it is more autonomous
for robots. Besides, Best et al. [11] described a learning
algorithm to cluster social cue observations and defined an “N-
Most likely State” for each cluster. Consequently, it can realize
the development of the robot’s social awareness and identify
human emotions.
IV. D
ISCUSSIO N
Cognitive developmental patterns and models were
discussed in Sections 2 and 3. When we set about doing
research, developmental pattern is the first thing we have to
consider, as it determines the thought of development. The first
pattern is easier to implement, due to the pre-defined
knowledge and rules. Robots will understand them with
specific meanings, whereas they may have different meanings
in other situations. Thus robots may only work in a particular
area or perform special tasks, which greatly limits the cognitive
ability of robots. The second pattern allows robots to explore
and learn in different environment for acquiring knowledge, or
utilize their own experience to deal with new situations. The
meanings of knowledge and actions are built by the robot itself.
Therefore, robots can be more independent.
Table 1 summarizes the studies that are cited in this work,
and for each of them it points out developed skills, model and
basis. We can see that models with different learning
mechanisms can develop the same cognitive ability, such as the
basic exploration actions. But each mechanism has its own
development emphasis, and will give us different inspirations.
According to the emergent paradigm, the cognitive
development of robots starts from an original state. The
infant’s cognitive development theory provides an explicit
thought and an initial sate. Most studies referred to infant’s
phased development process, and they built models to imitate
the special development task in each stage. Another significant
mechanism is curiosity. Curiosity can drive humans to explore
new knowledge and it belongs to internal goals. Besides,
external goals generated in interaction dive humans execute
2052
tasks. Internal and external goals are both needed for robot
development. While the learning from sensorimotor experience
focuses on develop concrete abilities, such as learning
behaviors and language. This type of model can integrate a
variety of modalities including vision, audition and actions, so
that robots can develop more complex cognitive abilities. As to
brain mechanism, studies mainly utilized the coordination of
the brain. Robots can also use a brain-like network to
coordinate a variety of cognitive behaviors to achieve the
smooth switching and cooperating. In addition, this
coordination ability may be a solution for robots to achieve the
autonomous transition from one stage to another. Other
methods also have a lot of innovations, and can give us some
guidance in specific development tasks.
TABLE I. S
UMMARY OF
T
HE
S
YUDIES
D
ISCUSSED
I
N
T
HIS
W
ORK
Study Abilities developed Pattern Model Learning mechanism
Nakajo et al. [3] Identifying body Emergent paradigm S-CTRNN Predictive learning
Meola et al. [4] Primitive actions Emergent paradigm Reinforcement learning Rhythmic and discrete movement
of infants
Gemignani et al. [5] Knowledge representation Pre-defined knowledge Semantic mapping Other Approaches
Olier et al. [6] Concept representation Emergent paradigm VRNN Learning from sensorimotor
experience
Aguilar et al. [7] Building new behaviors Pre-defined knowledge Dev E-R Piaget’s theory
Park et al. [8] Identifying humans gestures and
Imitating actions Pre-defined rules MTRNN Learning from sensorimotor
experience
Tani[9]
Shifts of action intention,
language-action associations, and
learning of goal-directed actions
Emergent paradigm RNNPB, MTRNN Learning from sensorimotor
experience
Maniadakis et al. [10] Rule switching and confidence Pre-defined rules CTRNN Brain mechanism
Best et al. [11] Social signal detection Pre-defined knowledge N-Most Likely States Other Approaches
Li et al. [12] Recgnization of objects Emergent paradigm SVM and LSVM Piaget’s theory
Luo et al. [13] Reaching Emergent paradigm Auto-encoder, FNN,
and HIM Corbetta’s psychology discovery
Nishide et al. [14] Drawing imitation Emergent paradigm MTRNN Human’s drawing skills
Yamada et al. [17] Integration of language and
behavior Pre-defined knowledge MTRNN Learning from sensorimotor
experience
Hwang et al. [18, 19] Seamless integration of cognitive
skills Emergent paradigm VMDNN Visuo-motor coordination
Saegusa et al. [20] Perception of the self and
primitive actions Emergent paradigm Multi-layer perceptron Visuo-motor correlation
Franchi et al. [21] Learning new goals and
composing new behaviors Emergent paradigm IDRA Brain mechanism
Salgado et al. [22, 23] Learning procedural
representations Emergent paradigm MDB Brain mechanism
Ramík et al. [24] discover autonomously objects
and learning new knowledge Pre-defined knowledge Multi-layer perceptron
and genetic algorithm Other Approaches
Taniguchi et al. [25] Spatial concept acquisition Pre-defined knowledge
Nonparametric
Bayesian spatial
concept acquisition
method
Other Approaches
Baxter et al. [26] Human–robot interaction Pre-defined knowledge Distributed Associative
Interactive Memory Other Approaches
Pointeau et al. [27] Accumulating and consolidating
experience Pre-defined knowledge Autobiographical
memory Brain mechanism
Cervantes et al. [29] Planning and decision-making Emergent paradigm Bio-inspired
computional model Brain mechanism
Reder et al. [30] Learning behaviors Pre-defined knowledge Case Based Reasoning Brain mechanism
V. C
ONCLUSION
The present study presented a constrained review on
autonomous cognitive developmental models of robots in
current literature. It analyzed two developmental patterns, and
summarized prevalent modeling methods. Based on this
information, we can find some inspiration for our future
research.
In our opinion, emergent paradigm is a more promising
research direction for cognitive development. Robot should
have the ability to develop independently. Pre-defined
knowledge or rules maybe limit this ability. In the future
research, we will follow the idea in Fig. 1, and firstly consider
how to build the primitive state of robots. We should figure out
the optimal curiosity approach for robots’ interest and attention.
What’s more, we are going to add emotional factors in
primitive state, for affect can help robots make decision in
learning process. Then, when robots start developing, we also
need to find a method of identifying robot’s body. As for
learning models, each approach will give us some
enlightenment. We plan to combine the advantages of each
model. Specifically, developmental framework of robots can be
established according to the process of infant’s development. In
the concrete realization of each stage’s cognition, we can build
2053
artificial neural network to learn knowledge and behaviors
from sensorimotor experience. Especially, deep learning has
achieved good results in solving the dynamic time sequence
problem. To coordinate various cognitive behaviors, we need a
structure similar to the cerebral cortex. In the meanwhile, we
also require a memory to save knowledge. Some articles even
proposed to let robots have the ability to dream. One way is to
consolidate what have learned [27], and another way is to
create new knowledge in dream based on existing knowledge
[28].Additionally, most researches focused on combining
visual and proprioceptive information, or associating language
and actions. However, there are few work consider to integrate
all three modalities including vision, auditory, actions.
Referring to Hwang’s model, we plan to add a MTRNN to
process auditory information and generate speech signals.
A
CKNOWLEDGMENT
The authors would like to thank the reviewers for their
comments, corrections, and suggestions which improved the
quality of this text immensely.
R
EFERENCES
[1] A. Aly, S. Griffiths, and F. Stramandinoli, “Metrics and benchmarks in
human-robot interaction: Recent advances in cognitive robotics,”
Cognitive Systems Research. vol. 43, pp. 313-323, June 2017.
[2] D. Vernon, C. Hofsten, and L. Fadiga, “Desiderata for developmental
cognitive architectures,” Biologically Inspired Cognitive Architectures.
vol. 18, pp.116-127, October 2016.
[3] R. Nakajo, M. Takahashi, S. Murata, H.Arie,and T. Ogata,Self and
non-self discrimination mechanism based on predictive learning with
estimation of uncertainty,” ICONIP 2016, Part IV, LNCS 9950, pp. 228-
235, 2016.
[4] V. C. Meola, D. Caligiore, V. Sperati, L. Zollo, A. L. Ciancio, F. Taffoni,
and et al, “Interplay of rhythmic and discrete manipulation movements
during development: a policy-search reinforcement-learning robot
model,” IEEE Ttans. Cognitive and Developmental Systems. vol. 8, NO.
3, September 2016.
[5] G. Gemignani, R. Capobianco, E. Bastianelli, D. D. Bloisi, L. Iocchi,
and D. Nardi, “Living with robots: Interactive environmental knowledge
acquisition,” Robotics and Autonomous Systems. vol. 78, pp. 1–16,
April 2016.
[6] J. S. Olier, E. Barakova, C. Regazzoni, M. Rauterberg, “Re-framing the
characteristics of concepts and their relation to learning and cognition in
artificial agents,” Cognitive Systems Research. vol. 44, pp.50-68,
August 2017.
[7] W. Aguilar, R. P. Pérez, “Dev E-R: A computational model of early
cognitive development as a creative process,” Cognitive Systems
Research. vol. 33, pp.17-41, March 2015.
[8] G. Park, J. Tani, “Development of compositional and contextual
communicable congruence in robots by using dynamic neural network
models,” Neural Networks. vol. 72, pp.109-122, December 2015.
[9] J. Tani, “Self-organization and compositionality in cognitive brains: A
neurorobotics study,” Proceedings of the IEEE. vol. 102,No.4,April
2014.
[10] M. Maniadakis, P. Trahanias, J. Tani, “Self-organizing high-order
cognitive functions in artificial agents: Implications for possible
prefrontal cortex mechanisms,” Neural Networks. vol. 33, pp.76-87,
September 2012.
[11] A. Best, K. A. Kapalo, S. F. Warta, S. M. Fiore, “Clustering social cues
to determine social signals: Developing learning algorithms using the
"N-Most Likely States" approach,” Proc. of SPIE .vol. 9837, 98370L,
April 2016.
[12] K. Li, Max Q.-H. Meng, “Learn like infants: A strategy for
developmental learning of symbolic skills using humanoid robots,”
International Journal of Social Robotics, vol. 7, issue 4, pp.439–450,
August 2015.
[13] D. Luo, F. Hu, Y. Deng, W. Liu, X. Wu, “An infant-inspired model for
robot developing its reaching ability,” ICDL-EpiRob Cergy-Pontoise,
Paris, France,Sept 19-22, 2016.
[14] S. Nishide, K. Mochizuki, H. G. Okuno , and T. Ogata, “Insertion of
pause in drawing from babbling for robot’s developmental imitation
learning,”ICRA, Hong Kong, China, May 31 - June 7, 2014.
[15] G. Schillaci, V.V. Hafner, and B. Lara, “Exploration behaviors, body
representations, and simulation processes for the development of
cognition in artificial agents,” Front. Robot. AI. 3:39, June 2016.
[16] Y. Yamashita, J. Tani, “Emergence of functional hierarchy in a multiple
timescale neural network model: A humanoid robot experiment,” PLoS
Comput Biol. vol. 4, issue.11, e1000220, November 2008.
[17] T. Yamada, S. Murata, H. Arie and T. Ogata, “Dynamical Integration of
Language and Behavior in a Recurrent Neural Network for Human–
Robot Interaction” Front. Neurorobot. vol. 10, article. 5, July 2016.
[18] J. Hwang, M. Jung, N. Madapana, J. Kim, M. Choi, and J. Tani,
“Achieving “synergy” in cognitive behavior of humanoids via deep
learning of dynamic visuo-motor-attentional coordination,” IEEE-RAS
15th International Conference on Humanoid Robots, November 2015.
[19] J. Hwang, M. Jung, J. Kim and J. Tani, “A deep learning approach for
seamless integration of cognitive skills for humanoid robots,” ICDL-
EpiRob. Cergy-Pontoise, France, 19-22 Sept. 2016.
[20] R. Saegusa, G. Metta, G.Sandini, and L. Natale, “Developmental
perception of the self and action,” IEEE Trans. Neural Networks and
Learning System. vol. 25, NO. 1, January 2014.
[21] A.M. Franchi, F. Mutti, and G. Gini,“From learning to new goal
generation in a bioinspired robotic setup,” Advanced Robotics, vol. 30,
NOS. 11-12, 795-805, April 2016.
[22] R. Salgado, F. Bellas, and R. J. Duro, “Autonomous learning of
procedural knowledge in an evolutionary cognitive architecture for
robots,” Mora A., Squillero G. (eds) Applications of Evolutionary
Computation. EvoApplications 2015. Lecture Notes in Computer
Science, vol 9028. Springer,Cham.2015.
[23] R. Salgado, A. Prieto, F. Bellas, L. Calvo-Varela, and R.J. Duro,
“Motivational engine with autonomous subgoal identification for the
Multilevel Darwinist Brain,” Biologically Inspired Cognitive
Architectures. vol. 17, pp.1-11, July 2016.
[24] D. M. Ramík, K. Madani, C. Sabourin, “A soft-computing basis for
robots’ cognitive autonomous learning,” Soft Comput.vol.19,issue.9,
pp.2407–2421, September 2015.
[25] A. Taniguchi, T. Taniguchi, and T. Inamura, “Spatial concept
acquisition for a mobile robot that integrates self-localization and
unsupervised word discovery from spoken sentences,” IEEE Trans.
Cognitive and Developmental Systems.vol. 8, issue.4, December 2016.
[26] P. E. Baxter, J. Greeff, and T. Belpaeme, “Cognitive architecture for
human–robot interaction: Towards behavioural alignment,” Biologically
Inspired Cognitive Architectures. vol. 6, pp. 30-39, October 2013.
[27] G.Pointeau, M. Petit, and P. F. Dominey, “Successive developmental
levels of autobiographical memory for learning through social
interaction,” IEEE Ttans. Autonomous Mental Development. vol. 6, NO.
3, September 2014.
[28] G. Marzanoa, A. Novembreb, “Machines that dream: a new challenge in
behavioral-basic robotics,” ICTE 2016, Riga, Latvia, December 2016.
[29] J. Cervantes, J. Rosales, S. Lopez, F. Ramos, M. Ramos, “Integrating a
cognitive computational model of planning and decision-making
considering affective information,” Cognitive Systems Research. vol. 44,
pp. 10-39, August 2017.
[30] I. H. Reder, C. Urdiales, J. M. Peula, F. Sandoval, “CBR based reactive
behavior learning for the memory-prediction framework,”
Neurocomputing. Vol. 250, pp. 18-27, August 2017.
... This is a multifaceted and interdisciplinary research area at the intersection of robotics, developmental psychology and developmental neuroscience. It is more than twenty years old (Sandini et al., 1998;Asada et al., 2001;Lungarella et al., 2003;Schmidhuber, 2006;Cangelosi et al., 2015;Min et al., 2016;Huang et al., 2017) and it clearly represents a brain-inspired approach to the design of robots: in contrast with industrial robots that perform repetitive predefined tasks in a predefined environment: these robots are supposed to be dynamic models of how humans develop, explore, and quickly adapt in an open-ended manner to a changing environment through lifelong learning, in order to cope with unpredictable challenges. Thus, the biomimetic goal of developmental robotics that is pursued is not to imitate the brain per se, namely the performing brain of trained adults, but the process of progressive knowledge acquisition, leading to autonomous decision-making ability by interaction with the physical and social environment. ...
... On the other hand, in spite of the intense and wide research activities on developmental robotics in the last two decades (Min et al., 2016;Huang et al., 2017;Wieser and Cheng, 2018;Naya-Varela et al., 2021;Rayyes et al., 2022), we are still far away from a level of understanding of the global implementation issues that may allow testing and evaluation in a sufficiently general manner. Certainly, there are still limitations regarding the "bodyware" capabilities of robots used to implement a developmental process inspired by humans, including the two most advanced child robots, namely iCub (Sandini et al., 2004) and CB2 (Minato et al., 2008). ...
Article
Full-text available
The trend in industrial/service robotics is to develop robots that can cooperate with people, interacting with them in an autonomous, safe and purposive way. These are the fundamental elements characterizing the fourth and the fifth industrial revolutions (4IR, 5IR): the crucial innovation is the adoption of intelligent technologies that can allow the development of cyber-physical systems , similar if not superior to humans. The common wisdom is that intelligence might be provided by AI (Artificial Intelligence), a claim that is supported more by media coverage and commercial interests than by solid scientific evidence. AI is currently conceived in a quite broad sense, encompassing LLMs and a lot of other things, without any unifying principle, but self-motivating for the success in various areas. The current view of AI robotics mostly follows a purely disembodied approach that is consistent with the old-fashioned, Cartesian mind-body dualism, reflected in the software-hardware distinction inherent to the von Neumann computing architecture. The working hypothesis of this position paper is that the road to the next generation of autonomous robotic agents with cognitive capabilities requires a fully brain-inspired, embodied cognitive approach that avoids the trap of mind-body dualism and aims at the full integration of Bodyware and Cogniware. We name this approach Artificial Cognition (ACo) and ground it in Cognitive Neuroscience. It is specifically focused on proactive knowledge acquisition based on bidirectional human-robot interaction: the practical advantage is to enhance generalization and explainability. Moreover, we believe that a brain-inspired network of interactions is necessary for allowing humans to cooperate with artificial cognitive agents, building a growing level of personal trust and reciprocal accountability: this is clearly missing, although actively sought, in current AI. The ACo approach is a work in progress that can take advantage of a number of research threads, some of them antecedent the early attempts to define AI concepts and methods. In the rest of the paper we will consider some of the building blocks that need to be re-visited in a unitary framework: the principles of developmental robotics, the methods of action representation with prospection capabilities, and the crucial role of social interaction.
... Huang, Ma, Tian, and Li [22] surveyed the cognitive development patterns in robots based on emergent and predefined knowledge paradigms and reported the superiority of cognitive development schemes based on emergent knowledge over the ones based on predefined knowledge. In this analysis, the difference between emergent and predefined knowledge is mainly characterized by the perspective and structure of knowledge acquisition. ...
Conference Paper
This article addresses the fundamental questions on machine learning: what does it mean for machines to learn from experience, and what does it mean by machines in machine learning? Despite recent popularity and growth, significant challenges remain as the industry rapidly advances toward autonomous machines. In the context of autonomy, there is more to learning from experience than training machines to approximate big data. This brings the fundamental questions to the forefront of automation science and engineering as a critical area of exploration. This article examines the precise notion of autonomy in the context of machine learning and provides a general framework for cyber-physical systems to become fully autonomous by learning from experience. The framework is derived from the principles of developmental autonomous behavior, which encapsulates broad classes of learning mechanisms. It offers a novel use case of machine learning where sensorimotor systems build inference engines internally on their own by their own initiatives to develop new skills and behavior. The key contributions of this article are threefold. First on knowledge: it provides precise definitions of emergent and predefined knowledge and their roles in cognitive development of machines. Second on autonomy: it clarifies what a fully autonomous machine means by providing the precise definitions of autonomy and emergent behavior. Third on machine learning: it unifies machine learning as the ultimate-proximate causal drivers of emergent behavior. Ultimately, this article logically explains why and how a fully autonomous machine is possible by directly answering the fundamental questions.
... Inspired by infants' cognitive process, some researchers applied development theories into robotics. From these theories, researchers can understand the way infants build their structures of knowledge and develop their behaviors, language, and other complex skills [115]. Then let robots learn like human infants. ...
Article
Full-text available
Research studies on social robotics and human-robot interaction have gained insights into factors that influence people’s perceptions and behaviors towards robots. However, adults’ perceptions of robots may differ significantly from those of infants. Consequently, extending this knowledge also to infants’ attitudes toward robots is a growing field of research. Indeed, infant-robot interaction (IRI) is emerging as a critical and necessary area of research as robots are increasingly used in social environments, such as caring for infants with all types of disabilities, companionship, and education. Although studies have been conducted on the ability of robots to positively engage infants, little is known about the infants’ affective state when interacting with a robot. In this systematic review, technologies for infant affective state recognition relevant to IRI applications are presented and surveyed. Indeed, adapting techniques currently employed for infant’s emotion recognition to the field of IRI results to be a complex task, since it requires timely response while not interfering with the infant’s behavior. Those aspects have a crucial impact on the selection of the emotion recognition techniques and the related metrics to be used for this purpose. Therefore, this review is intended to shed light on the advantages and the current research challenges of the infants’ affective state recognition approaches in the IRI field, elucidates a roadmap for their use in forthcoming studies as well as potentially provide support to future developments of emotion-aware robots.
Conference Paper
Full-text available
In robotics community, inspiration from human development theory offers a promising way for robots to efficiently achieve many abilities. Humanoid robot reaching, as an essential ability which forms the foundation of grasping and manipulation, is also possible to take advantages from human. For human infants, reaching is emerging around 3–5 months of age, which is also a fundamental skill for the development and refinement of future high level motion or cognitive behaviors. The issue of how infant reaching is developed has been investigated for decades and several views were established and discussed. Recently, Corbetta et al. proposed that the emergence of reaching is the product of a deeply embodied process, in which infants first learn how to direct their movement in space using proprioceptive and haptic feedback and then map the visual attention onto these bodily centered experiences. With this new sight, this paper addresses the problem of how a robot develops its reaching ability autonomously, and a novel infant-inspired model is proposed. The model is composed of several blocks so that to closely capture the inherent mechanisms of infants in the process of developing the reaching skill. To evaluate the proposed model, a child-sized physical humanoid robot PKU-HR6.0 is employed. Just like an early infant, the robot is assumed without any other abilities. It is only equipped with the proposed model with random initialized parameters. The reaching ability is expected to be developed all by the robot itself. Through iteratively babbling arm motions in its workspace, the robot firstly acquire a reliable sense of its body and movement in space, and then map the sense of the vision onto the proprioception when contingencies happen. Experimental results show that the robot equipped with the proposed model achieves the reaching ability effectively and successfully in a completely autonomous style. The robot is able to reach the objects in different real environments with a high performance, and can even grasp the objects with an average successful rate of 82.50%. Our experimental results also verify that the movement of one joint may induce another, and some sepecific joint may play dominant role in certain category of movements.
Article
Full-text available
The digital revolution is transforming contemporary society. Connective intelligence is an emerging property deriving from the embedding of intelligence into the connected data, concepts, applications, and people. Furthermore, the progress in behavioral-basic robotics opens new fields of innovative investigation.
Article
Full-text available
This paper complements Ron Sun’s influential Desiderata for Cognitive Architectures by focussing on the desirable attributes of a biologically-inspired cognitive architecture for an agent with a capacity for autonomous development. Ten desiderata are identified, dealing with value systems & motives, embodiment, sensorimotor contingencies, perception, attention, prospective action, memory, learning, internal simulation, and constitutive autonomy. These desiderata are motivated by studies in developmental psychology, cognitive neuroscience, and enactive cognitive science. All ten focus on the ultimate aspects of cognitive development — why a feature is necessary and what it enables — rather on than the proximate mechanisms by which it can be realized. As such, the desiderata are for the most part neutral regarding the paradigm of cognitive science — cognitivist or emergent — that is adopted when designing a cognitive architecture. Where some element of a desideratum is specific to a particular paradigm, this is noted.
Article
This study investigates how adequate coordination among the different cognitive processes of a humanoid robot can be developed through end-to-end learning of direct perception of visuomotor stream. We propose a deep dynamic neural network model built on a dynamic vision network, a motor generation network, and a higher-level network. The proposed model was designed to process and to integrate direct perception of dynamic visuomotor patterns in a hierarchical model characterized by different spatial and temporal constraints imposed on each level. We conducted synthetic robotic experiments in which a robot learned to read human's intention through observing the gestures and then to generate the corresponding goal-directed actions. Results verify that the proposed model is able to learn the tutored skills and to generalize them to novel situations. The model showed synergic coordination of perception, action and decision making, and it integrated and coordinated a set of cognitive skills including visual perception, intention reading, attention switching, working memory, action preparation and execution in a seamless manner. Analysis reveals that coherent internal representations emerged at each level of the hierarchy. Higher-level representation reflecting actional intention developed by means of continuous integration of the lower-level visuo-proprioceptive stream.
Article
In this work, the problems of knowledge acquisition and information processing are explored in relation to the definitions of concepts and conceptual processing, and their implications for artificial agents. The discussion focuses on views of cognition as a dynamic property in which the world is actively represented in grounded mental states which only have meaning in the action context. Reasoning is understood as an emerging property consequence of actions-environment couplings achieved through experience, and concepts as situated and dynamic phenomena enabling behaviours. Re-framing the characteristics of concepts is considered crucial to overcoming settled beliefs and reinterpreting new understandings in artificial systems. The first part presents a review of concepts from cognitive sciences. Support is found for views on grounded and embodied cognition, describing concepts as dynamic, flexible, context-dependent, and distributedly coded.That is argued to contrast with many technical implementations assuming concepts as categories, whilst explains limitations when grounding amodal symbols, or in unifying learning, perception and reasoning. The characteristics of concepts are linked to methods of active inference, self-organization, and deep learning to address challenges posed and to reinterpret emerging techniques. In a second part, an architecture based on deep generative models is presented to illustrate arguments elaborated. It is evaluated in a navigation task, showing that sufficient representations are created regarding situated behaviours with no semantics imposed on data. Moreover, adequate behaviours are achieved through a dynamic integration of perception and action in a single representational domain and process.
Conference Paper
This study investigates the seamless integration of cognitive skills, such as visual recognition, attention switching, action preparation and generation for a humanoid robot. In our preliminary study [1], the deep dynamic neural network model was introduced to process spatio-temporal visuomotor patterns. In the current study, we extended the previous model further to enhance its capability of handling sequential visuomotor information as well as forming visuomotor representation. We conducted synthetic robotic experiments in which a robot learned goal-directed actions of reaching to grasp objects under two different experimental settings. In the first experiment, a task of reaching to grasp objects was conducted under parameterized visual occlusion condition for the purpose of examining the memory capability in the model. In the second experiment, the action of reaching to grasp objects was incorporated with visual recognition of human gesture patterns with using the working memory. The experimental results revealed that the proposed model was able to generalize its reaching and grasping skills in the novel situations. Furthermore, the analysis using the dimensionality reduction technique on neuron activation verified that the proposed model was capable of manipulating high dimensional spatio-temporal visuomotor patterns by forming their dynamic link to the actional intention developed in the higher level of the model via iterative learning.
Article
Some approaches to intelligence state that the brain works as a memory system which stores experiences to reflect the structure of the world in a hierarchical, organized way. Case Based Reasoning (CBR) is well suited to test this view. In this work we propose a CBR based learning methodology to build a set of nested behaviors in a bottom up architecture. To cope with complexity-related CBR scalability problems, we propose a new 2-stage retrieval process. We have tested our framework by training a set of cooperative/competitive reactive behaviors for Aibo robots in a RoboCup environment.
Conference Paper
In this paper, we propose a model that can explain the mechanism of self and non-self discrimination. Infants gradually develop their abilities for self–other cognition through interaction with the environment. Predictive learning has been widely used to explain the mechanism of infants’ development. We hypothesized that infants’ cognitive abilities are developed through predictive learning and the uncertainty estimation of their sensory-motor inputs. We chose a stochastic continuous time recurrent neural network, which is a dynamical neural network model, to predict uncertainties as variances. From the perspective of cognitive developmental robotics, a predictive learning experiment with a robot was performed. The results indicate that training made the robot predict the regions related to its body more easily. We confirmed that self and non-self cognitive abilities might be acquired through predictive learning with uncertainty estimation.