This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Leading or Following? Dyadic Robot Imitative Interaction Using the Active Inference Framework

Nadine Wirkuttis1 and Jun Tani2,∗ This work was sponsored by the Okinawa Institute of Science and Technology Graduate University, Okinawa, Japan 904-0302.1 The authors are with the Cognitive Neurorobotics Research Unit, Okinawa Institute of Science and Technology Graduate University, Okinawa, Japan 904-0302. {nadine.wirkuttis,jun.tani}@oist.jp*Corresponding author.
Abstract

This study investigated how social interaction among robotic agents changes dynamically depending on the individual belief of action intention. In a set of simulation studies, we examine dyadic imitative interactions of robots using a variational recurrent neural network model. The model is based on the free energy principle such that a pair of interacting robots find themselves in a loop, attempting to predict and infer each other’s actions using active inference. We examined how regulating the complexity term to minimize free energy determines the dynamic characteristics of networks and interactions. When one robot trained with tighter regulation and another trained with looser regulation interact, the latter tends to lead the interaction by exerting stronger action intention, while the former tends to follow by adapting to its observations. The study confirms that the dyadic imitative interaction becomes successful by achieving a high synchronization rate when a leader and a follower are determined by developing action intentions with strong belief and weak belief, respectively.

I INTRODUCTION

Social interaction is considered an essential cognitive behavior. In both empirical studies and synthetic modeling, researchers have investigated underlying cognitive, psychological, and neuronal mechanisms accounting for various aspects of social cognitive behaviors. This study investigates mechanisms underlying synchronized imitation as a representative social cognitive act, by formulating the problem using the free energy principle (FEP) [1, 2]. In simulation experiments of dyadic robot imitative interaction, we examine how a leader and follower can be determined in conflicting situations by investigating the underlying network dynamics.

Numerous robotic studies have investigated imitative interaction. In the 90s, imitation was identified as an indispensable human competency required in early development of cognitive behaviors [3, 4, 5, 6]. Rizzolatti and colleagues [7] showed that the mirror neuron system uses observations of an action to generate the same action. Arbib and Oztop [8, 9] indicated that mirror neurons may participate in imitative behaviors. Upon this development, several research groups proposed computational mirror neuron models for imitation using Hidden Markov Models [10] and neural network models [11, 12, 13, 14].

An essential unsolved question in modeling of imitative interaction is how a leader, who initiates an action, and a follower, who imitates this action, can be determined when multiple choices of actions are possible among a set of well-habituated ones.

Recent theories on predictive coding (PC) and active inference (AIF) based on the free energy principle (FEP) [2, 15] show that ”action intention” and its ”belief” can be formulated as a predictive model. ”Action intention” is considered a top-down prediction of action outcomes and ”belief” as an estimated precision of this prediction or the strength of intention (as described in [2, 15]). Analogous to PC and AIF, Ito and Tani showed that imitative interaction can be performed using an RNN model by minimizing the prediction error instead of free energy in order to update deterministic latent variables [13]. However, this deterministic model does not account for the belief of action intention because the precision of prediction cannot be estimated. On a related topic, Ahmadi and Tani developed predictive-coding inspired variational RNN (PV-RNN) [16]. Their model was used to investigate how the strength of top-down intention in predicting fluctuating temporal patterns was modulated, depending on learning conditions in the model. In the learning process, free energy represented by the weighted sum of the accuracy term and the complexity term is minimized. Ahmadi and Tani found that softer regulation of the complexity term during network training develops strong top-down intention. Predictions are more deterministic by self-organizing deterministic dynamics with the initial sensitivity characteristics in the network. Likewise, tighter regulation of the complexity term results in weaker intention and increased stochasticity. Compared to other neural network models based on the FEP [17, 18, 19, 20], PV-RNN has advantages when applied to problems in robotics. It can cope with temporal structure by using recurrence associated with stochastic latent variables and by hierarchical abstraction through a multiple timescale structure [21].

Our research group further investigated human-robot imitative interaction using PV-RNN. Chame and Tani [22] showed that a humanoid robot with force feedback control tends to lead or follow the human counterpart in imitative interaction when its PV-RNN is set to softer or tighter regulation, respectively. However, the result is preliminary, merely showing a one-shot experimental result without any quantitative analysis. In a similar experimental setup, Ohata and Tani [23] showed that this tendency can be also observed when regulation of the complexity term is modulated during the interaction phase, rather than during the prior learning phase. The study investigated pseudo-imitative interaction between a simulated robot and a human counterpart. This study, however, lacks genuine interaction between the simulated robot and the human counterpart because the outputs of the counterpart were replaced with static output sequences prepared in advance.

The main contribution of the current study is to clarify the underlying mechanism of how a leader and a follower can be determined in dyadic synchronized imitative interaction using the framework of AIF. This study is distinct from the author’s aforementioned past studies because genuine interaction between two robots using the same model is examined and results are analysed both quantitatively and qualitatively. An advantage of performing a robot-robot interaction experiment is that the internal dynamics can be analyzed in a comparative way between the two robots.

The interaction experiment considers two robotic agents that are trained to generate a set of movement primitive sequences. When movements are generated by following a probabilistic finite state machine, the transition probability differs, depending on each of the two robots. After each robot learns the given probabilistic transition structure for a sequence, the experimental design allows us to investigate how two robots generate movement primitives in the synchronised imitative interaction. In particular, we examine conflicting situations in which each robot prefers to generate different movement patterns, depending on its learned experience. Do they synchronize to generate the same movement pattern with one robot following the other or leading by adapting the intention? Or do they desynchronize by generating different movement patterns, ignoring their counterparts by following their own action intentions? The current study hypothesizes that these dyadic interaction outcomes depend on the relative strength of the intention between the robots as a result of regulating FEP complexity.

II Model

II-A Predictive Coding and Active Inference

The current study applies the concepts of PC and AIF based on FEP [1]. PC considers perception as the interplay between a prior expectation of a sensation and a posterior inference about a sensory outcome. Expectation of the sensation can be modeled by a generative model that maps the prior of the latent state to the expectation of sensation. The posterior inference of the observed sensation can be achieved by jointly minimizing the error computed between the expected sensation and its outcome, i.e. the accuracy, plus the Kullback-Leibler (KL) Divergence between the posterior and the prior distributions, i.e. the complexity. Posterior and prior are both represented by Gaussian probability distributions using means and variances. This is to minimize free energy or to maximize the lower bound of the logarithm of marginal likelihood:

logpθ(𝑿¯)\displaystyle\footnotesize\log p_{\theta}(\bar{\boldsymbol{X}}) qϕ(𝒛|𝑿¯)logpθ(𝑿¯,𝒛)qϕ(𝒛|𝑿¯)d𝒛Evidencelowerbound\displaystyle\geq\underbrace{\int q_{\phi}(\boldsymbol{z}|\bar{\boldsymbol{X}})\log\frac{p_{\theta}(\bar{\boldsymbol{X}},\boldsymbol{z})}{q_{\phi}(\boldsymbol{z}|\bar{\boldsymbol{X}})}d\boldsymbol{z}}_{\rm Evidence\ lower\ bound} (1)
=𝔼qϕ(𝒛|𝑿¯)[logpθ(𝑿¯|𝒛)]AccuracyDKL[qϕ(𝒛|𝑿¯)pθ(𝒛)]Complexity\displaystyle=\underbrace{\mathbb{E}_{q_{\phi}(\boldsymbol{z|\bar{\boldsymbol{X}}})}[\log p_{\theta}(\bar{\boldsymbol{X}}|\boldsymbol{z})]}_{\rm Accuracy}-\underbrace{D_{\rm KL}[q_{\phi}(\boldsymbol{z}|\bar{\boldsymbol{X}})\|p_{\theta}(\boldsymbol{z})]}_{\rm Complexity}

𝒛\boldsymbol{z}, 𝑿¯\bar{\boldsymbol{X}}, pθp_{\theta}, and qϕq_{\phi} denote the latent state, the observation, the prior distribution, and the approximate posterior, respectively. θ\theta and ϕ\phi are the parameters of the generative and inference model. In maximizing the lower bound, the interplay between accuracy and complexity characterizes the model performance in learning, prediction, and inference.

Consistent with the AIF, actions are generated so that the error between the expected action outcome and the actual outcome is minimized. In robotic applications, this is equivalent to determining how expected proprioception in terms of robot joint angles can be achieved by generating adequate motor torque. A simple solution is to use a PID controller, in which adequate motor torque to minimize errors between expected joint angles and actual angles can be obtained by means of error feedback schemes. Finally, perception by predictive coding and action generation by active inference are deployed simultaneously, thereby closing the loop of action and perception.

II-B Overview of PV-RNN

The PV-RNN model is designed to predict future sensation by means of prior generation, while reflecting the past by means of posterior inference based on learning (see Fig. 1). One essential element of the model is the introduction of a parameter 𝐰\mathbf{w}, the so-called meta-prior, which regulates the complexity term in free energy. Different 𝐰\mathbf{w} settings results in alternation of the estimated precision in predicting the sensation, as described later as prior generation (see section III-C). The model is also characterised by employing an architecture of multiple timescale RNN (MTRNN) [21]. The whole network comprises multiple layers of RNNs wherein the dynamics of each layer are governed by different time constant parameters τ\tau. This scheme supports development of hierarchical information processing by adequately setting the timescale of each layer [21, 14]. This approach is considered as analogous to [24, 25].

The following briefly describes the two essential parts, a generative model which is used for prior generation to make future predictions, and an inference model, which is used for posterior inference about the past. For further details, refer to [16, 23].

II-B1 Generative Model

The stochastic generative model is used for prior generation, as illustrated in the future prediction part (after time step 4) in Fig. 1.

\includegraphics

[width=0.45]figures/architecture.pdf

Figure 1: Graphical representation of a hierarchical 3-layer PV-RNN architecture. Layers are indicated on the left. Time increases from left to right and is indicated as a subscript. The representation shows the network in t=3t=3 with a two-time-step posterior inference window [2,3][2,3] and prior generation for t=[4,5]t=[4,5]. In the posterior inference window, prediction error 𝐞x\mathbf{e}^{x} and the 𝐰\mathbf{w} weighted KL Divergence 𝐞z\mathbf{e}^{z} are minimized.

PV-RNN is comprised of deterministic variables 𝐝\mathbf{d} and random variables 𝐳\mathbf{z}. An approximate posterior distribution qϕq_{\phi} is inferred based on the prior distribution pθp_{\theta} by means of error minimization on the generated prediction 𝐗\mathbf{X}. The generative model can be factorized as:

pθ(𝑿¯1:T,𝐝1:T,𝒛1:T|𝒅0)=\displaystyle\footnotesize p_{\theta}(\bar{\bm{X}}_{1:T},{\mathbf{d}}_{1:T},\bm{z}_{1:T}|\bm{d}_{0})= (2)
t=1Tpθ𝑿¯(𝑿¯t|𝐝t)pθd(𝐝t|𝐝t1,𝒛t)pθz(𝒛t|𝐝t1)\displaystyle\prod_{t=1}^{T}p_{\theta_{\bar{\bm{X}}}}(\bar{\bm{X}}_{t}|{\mathbf{d}}_{t})p_{\theta_{{d}}}({\mathbf{d}}_{t}|{\mathbf{d}}_{t-1},\bm{z}_{t})p_{\theta_{z}}(\bm{z}_{t}|{\mathbf{d}}_{t-1})

Although 𝐝\mathbf{d} is a deterministic variable, it can be considered to have a Dirac delta distribution centered on 𝒅~\tilde{\bm{d}} as 𝝈(𝒅𝒅~)\bm{\sigma}(\bm{d}-{\tilde{\bm{d}}}). 𝑿¯\bar{\bm{X}} is conditioned directly on 𝒛\bm{z} through 𝒅~\tilde{\bm{d}}. At the initial time step, 𝒅~\tilde{\bm{d}} is set to 0. Otherwise, 𝒅~\tilde{\bm{d}} is recursively computed, for which the internal state before activation is denoted by 𝒉\bm{h}. This internal state 𝒉\bm{h} is a vector, calculated as the sum of the internal states of the current level ll and its connecting layers of the previous time step t1t-1 plus the latent 𝒛\bm{z} in the same layer of the current time step tt:

𝒅~tl\displaystyle\bm{\tilde{d}}^{l}_{t} =tanh(𝒉tl)\displaystyle=\text{tanh}(\bm{h}^{l}_{t}) (3)
𝒉tl\displaystyle\bm{h}^{l}_{t} =(11τl)𝒉t1l+\displaystyle=\left(1-\frac{1}{\tau^{l}}\right)\bm{h}^{l}_{t-1}+
1τl(𝑾ddll𝒅~t1l+𝑾zdll𝒛tl+𝑾ddll+1𝒅~t1l+1+𝑾ddll1𝒅~t1l1)\displaystyle\frac{1}{\tau^{l}}\left(\bm{W}^{ll}_{dd}\bm{\tilde{d}}^{l}_{t-1}+\bm{W}^{ll}_{zd}\bm{z}^{l}_{t}+\bm{W}^{ll+1}_{dd}\bm{\tilde{d}}^{l+1}_{t-1}+\bm{W}^{ll-1}_{dd}\bm{\tilde{d}}^{l-1}_{t-1}\right)

τl\tau^{l} denotes the layer-specific time constant. With larger value for τl\tau^{l}, slower timescale dynamics develop, whereas with a smaller value set, faster timescale dynamics dominate. 𝑾\bm{W} represents connectivity weight matrices between layers and their deterministic and stochastic units. The output with size NxN_{x} is computed as mapping from 𝒅~1\bm{\tilde{d}}^{1} as:

𝐗t=𝑾Xdll𝒅~t1+𝒃X\displaystyle\mathbf{X}_{t}=\bm{W}^{ll}_{Xd}\bm{\tilde{d}}_{t}^{1}+\bm{b}_{X} (4)

The prior distribution pθ(𝒛t)p_{\theta}(\bm{z}_{t}) is a Gaussian distribution represented with mean 𝝁tp\bm{\mu}_{t}^{p} and standard deviation 𝝈tp\bm{\sigma}_{t}^{p}. The prior depends on 𝒅~t1\bm{\tilde{d}}_{t-1} by following the idea of a sequence prior [26], except at t=1t=1 where it follows a unit Gaussian distribution.

pθ(𝒛1)\displaystyle p_{\theta}(\bm{z}_{1}) =𝒩(0,I)\displaystyle=\mathcal{N}(0,I) (5)
pθ(𝒛t|𝒅~t1)\displaystyle p_{\theta}(\bm{z}_{t}|\tilde{\bm{d}}_{t-1}) =𝒩(𝝁tp,(𝝈tp)2) where t>1\displaystyle=\mathcal{N}(\bm{\mu}^{p}_{t},(\bm{\sigma}^{p}_{t})^{2})\text{ where $t>1$}
𝝁tp\displaystyle\bm{\mu}^{p}_{t} =tanh(𝑾dμpll𝒅~t1)\displaystyle=\text{tanh}(\bm{W}^{ll}_{d\mu^{p}}\bm{\tilde{d}}_{t-1})
𝝈tp\displaystyle\bm{\sigma}^{p}_{t} =exp(𝑾dσpll𝒅~t1)\displaystyle=\exp(\bm{W}^{ll}_{d\sigma^{p}}\bm{\tilde{d}}_{t-1})

Based on the work on variational autoencoders, we use the reparameterization trick to formulate the latent prior of 𝒛t\bm{z}_{t} as mean 𝝁tp\bm{\mu}^{p}_{t} and standard deviation 𝝈tp\bm{\sigma}^{p}_{t}. The reparameterization trick was introduced by Kingma and Welling [27] to make random variables differentiable for backpropagating errors through the network for learning. The same consideration is taken for the posterior of 𝒛t\bm{z}_{t} in the inference model as well (cf. below Eq. 7).

II-B2 Inference Model

Posterior inference is performed during learning and afterward, during action and perception. Fig. 1 illustrates information flow in the posterior inference in a time window from time step 2 to time step 3. The inference model for the posterior is described as:

Qϕ(𝒛t|𝒅~t1,𝒆t:T)=𝒩(𝒛t;𝝁tq,𝝈tq)\displaystyle\footnotesize Q_{\phi}(\bm{z}_{t}|\bm{\tilde{d}}_{t-1},\bm{e}_{t:T})=\mathcal{N}(\bm{z}_{t};\bm{\mu}^{q}_{t},\bm{\sigma}^{q}_{t}) (6)

where 𝒆t\bm{e}_{t} denotes the error between the target 𝐗¯t\bar{\mathbf{X}}_{t} and the predicted output 𝐗t\mathbf{X}_{t}. Like the prior pθp_{\theta}, the posterior qϕq_{\phi} is also a Gaussian distribution with mean 𝝁tq\bm{\mu}_{t}^{q} and standard deviation 𝝈tq\bm{\sigma}_{t}^{q}. For 𝒛1:T\bm{z}_{1:T} it is defined as:

qϕ(𝒛t|𝒆t:T)\displaystyle q_{\phi}(\bm{z}_{t}|\bm{e}_{t:T}) =𝒩(𝝁tq,𝝈tq)\displaystyle=\mathcal{N}(\bm{\mu}^{q}_{t},\bm{\sigma}^{q}_{t}) (7)
𝝁tq\displaystyle\bm{\mu}^{q}_{t} =tanh(𝑨tμ)\displaystyle=\text{tanh}(\bm{A}^{\mu}_{t})
𝝈tq\displaystyle\bm{\sigma}^{q}_{t} =exp(𝑨tσ)\displaystyle=\exp(\bm{A}^{\sigma}_{t})

Since computing the true posterior is intractable, an approximate posterior qϕq_{\phi} is inferred by maximizing the lower bound, analogous to Eq. (1). Here, the adaptation variable 𝐀1:T\mathbf{A}_{1:T} forces the parameters ϕ\phi of the inference model to represent meaningful information. The lower bound of PV-RNN can be derived as:

L(θ,ϕ)\displaystyle\footnotesize L(\theta,\phi) =t=1T(1NXEqϕ(𝒛t|𝒅~t1,𝒆t:T)[logpθX¯(𝑿¯t|𝒅~t)]\displaystyle=\sum_{t=1}^{T}(\frac{1}{N_{X}}E_{q_{\phi}(\bm{z}_{t}|\bm{\tilde{d}}_{t-1},\bm{e}_{t:T})}\big{[}\log p_{\theta_{\bar{X}}}(\bar{\bm{X}}_{t}|\bm{\tilde{d}}_{t})\big{]}- (8)
lL𝐰lNzlDKL[qϕ(𝒛t|𝒅~t1,𝒆t:T)||pθz(𝒛t|𝒅~t1)])\displaystyle\sum_{l}^{L}\frac{\mathbf{w}^{l}}{N_{z}^{l}}D_{KL}\big{[}q_{\phi}(\bm{z}_{t}|\bm{\tilde{d}}_{t-1},\bm{e}_{t:T})||p_{\theta_{z}}(\bm{z}_{t}|\bm{\tilde{d}}_{t-1})\big{]})

where the first term is the accuracy and the second term is the complexity (for details referred to [16]). NxN_{x} and NzlN_{z}^{l} are the number of sensory dimensions and the number of the latent random variables at the lthl^{th} layer, respectively. 𝐰l\mathbf{w}^{l} serves as a weighting parameter for the complexity term in layer ll and is referred to as the meta-prior [16]. The meta-prior represents the strength for regulating the closeness between the posterior and the prior distributions. In t=1t=1, 𝐰1l\mathbf{w}^{l}_{1} is set with 1.0. 𝐰2:Tl\mathbf{w}^{l}_{2:T} is set to a specific value when the sequence prior [26] is used after time step 1. In the posterior inference, all learning-related network parameters of θ\theta, ϕ\phi, and the adaptive variable 𝐀\mathbf{A} are updated to maximize the lower bound by back-propagating the error from time step TT back to t1t_{1} [28].

II-B3 PV-RNN in Dyadic Robot Interaction

Two robots equipped with the PV-RNN model interact during synchronized imitation. In the interaction, the robots predict proprioception 𝐗t+1pr\mathbf{X}^{pr}_{t+1} and exteroception 𝐗t+1ex\mathbf{X}^{ex}_{t+1} for the next time step. The predicted 𝐗t+1pr\mathbf{X}^{pr}_{t+1} regulates joint angle movements of a robot by considering a PID controller. This movement 𝐗t+1pr\mathbf{X}^{pr}_{t+1} can then be sensed by the other robot in terms of exteroception 𝐗¯t+1ex\bar{\mathbf{X}}^{ex}_{t+1}. This is provided through the kinematic transformation of joint angles 𝐗t+1pr\mathbf{X}^{pr}_{t+1} (cf. Fig. 2).

\includegraphics

[trim=0 20 0 -10mm, width=0.40]figures/interaction.pdf

Figure 2: Schematic of dyadic robot interaction where robots are equipped with the PV-RNN model.

While in the training phase, the error signal is taken from the proprioceptive 𝐗¯pr\bar{\mathbf{X}}^{pr} as well as the exteroceptive 𝐗¯ex\bar{\mathbf{X}}^{ex} target sequences, in the interaction phase the error signal for each robot is taken only from 𝐗¯ex\bar{\mathbf{X}}^{ex}111Considering only 𝐗¯ex\bar{\mathbf{X}}^{ex} for the error term in interaction settings assumes that the PID controller generates only negligible position errors for the joints.. During interactions, prediction errors 𝐞x{\mathbf{e}}^{x} are generated and propagated bottom-up throughout all layers, as well as time steps in the posterior inference window, in terms of the latent error 𝐞z{\mathbf{e}}^{z}. This updates posterior distributions in the network and minimizes the variational free energy. In this phase, only 𝐀1:T\mathbf{A}_{1:T} is updated without updating network parameters θ\theta and ϕ\phi.

III Robot Experiments

To investigate how the interaction of two robots changes with tighter and looser regulation of complexity, each robot was trained and tested individually, as described in III-B and in III-C, respectively. Finally, two robots were examined during a dyadic interaction (III-D).

III-A Task Design

Robotic agents are trained with three movement primitives A, B, and C (Fig. 3 (a)). Each primitive is 40 time steps in length. A human experimenter generated the primitive data via a master control of a humanoid OP2222Humanoid OP2 and its master controller are developed by Robotis: www.robotis.us/robotis-op2-us/.. The experimenter controlled six joints in the upper body of one humanoid 𝐗¯pr\bar{{\mathbf{X}}}^{pr}. The exteroceptive trajectory 𝐗¯ex\bar{{\mathbf{X}}}^{ex} is generated by mirroring its own movement 𝐗¯pr\bar{{\mathbf{X}}}^{pr} and transformed into 𝐗¯ex\bar{{\mathbf{X}}}^{ex} xy-coordinate positions of the left hand and right hand tips of the robot (Fig. 3(b)). 𝐗¯pr\bar{{\mathbf{X}}}^{pr} and 𝐗¯ex\bar{{\mathbf{X}}}^{ex} are six and four dimensions, respectively. Individual movement primitives are sampled and combined to form a continuous pattern of 400 time steps that follows a probabilistic sequence (analogous to [23]). Two probabilistic patterns were generated, A20%B80%C and A80%B20%C as shown in the form of a probabilistic finite state machine (P-FSM) (Fig. 3 (c)). The difference between these two probabilistic patterns is that C is biased and comes more often (80%) than B (20%) after A in the former, and vice versa for the latter.

A point of interest is the interaction phase after the learning phase. It is expected that both robots can generate A synchronously, since it is a deterministic state. This could be different from generating B or C as two robots learned different preferences in terms of transition probabilities. One robot may lead so as to generate B or C while the other may just follow it. However, both robots may generate their own biased movements and, thus, desynchronize their behavior. The current study hypothesizes that whether B or C is generated synchronized or desynchronized between the two robots depends on the complexity regulation of each robot.

\includegraphics

[trim=0 0 0 -5mm, width=]figures/dataset_a.pdf

(a) robot movements
\includegraphics

[width=]figures/dataset_b.pdf

(b) movement trajectories
\includegraphics

[width=]figures/dataset_c.pdf

(c) probabilistic transition
Figure 3: Task design. Robot movement primitives A, B, and C of the training dataset (a). Proprioceptive trajectories 𝐗¯pr\bar{\mathbf{X}}^{pr} plus exteroceptive trajectories 𝐗¯rex\bar{\mathbf{X}}^{ex}_{r} and 𝐗¯lex\bar{\mathbf{X}}^{ex}_{l}. Colors represent six dimensions of joint angles for 𝐗¯pr\bar{\mathbf{X}}^{pr} and xy-coordinate positions of the right and left end effector 𝐗¯rex\bar{\mathbf{X}}^{ex}_{r} and 𝐗¯lex\bar{\mathbf{X}}^{ex}_{l} (b). Two P-FSMs representing different movement primitive transition patterns of A20%B80%C and A80%B20%C (c).

III-B Robot Training

The PV-RNN was trained with 20 data samples on a set of different parameters (TABLE I). All network specific parameters were fixed during training. To explore the influence of the meta-prior, w, only this parameter changed for different networks and was repeated with different random seeds to ensure reproducibility.

TABLE I: Network parameters for training PV-RNNs.
𝐝\mathbf{d} 𝐳\mathbf{z} τ\tau 𝐰1\mathbf{w}_{1} 𝐰2:T\mathbf{w}_{2:T}
layer 1 40 4 22 11 𝐰1=[0.0,0.001,4.999,5.0]\mathbf{w}^{1}=[0.0,0.001,...4.999,5.0]
layer 2 20 2 44 11 𝐰2=w1×10\mathbf{w}^{2}={w}^{1}\times 10
layer 3 10 1 88 11 𝐰3=w1×100\mathbf{w}^{3}={w}^{1}\times 100

Networks were trained for 80,000 epochs, using Adam Optimizing and back-propagation through time (BPTT) [28] with learning rate 0.001. After training, network performance was first analysed in stand-alone robot experiments (subsection III-C). Thereafter, dyadic robot interaction was studied using networks trained with 𝐰\mathbf{w} set for the two representatives of tight and loose regulation of FEP complexity (subsection III-D).

III-C Preparatory Analysis of Training Results

TABLE II: Training performance of representative meta-prior 𝐰\mathbf{w}. The mean ±\pm standard deviation represent three random seeds and 20 repetitions of prior generation for each 𝐰\mathbf{w}.
𝐰\mathbf{w} training sequence A20%B80%C BC-ratio divergence step 𝐭\mathbf{t}
A B C not classified
0.0050.005 34±134\pm 1 11±311\pm 3 40±240\pm 2 15±115\pm 1 22±622\pm 6B 78±578\pm 5C 4343
0.010.01 35±235\pm 2 13±0.213\pm 0.2 30±230\pm 2 22±422\pm 4 30±530\pm 5B 70±570\pm 5C 5050
1.01.0 36±136\pm 1 11±211\pm 2 40±240\pm 2 13±0.213\pm 0.2 22±422\pm 4B 78±478\pm 4C 9191
22 41±0.741\pm 0.7 10±0.510\pm 0.5 39±0.739\pm 0.7 10±0.310\pm 0.3 21±821\pm 8B 79±879\pm 8C 120120
3.43.4 45±145\pm 1 11±111\pm 1 38±238\pm 2 6±0.56\pm 0.5 24±224\pm 2B 76±276\pm 2C 139139

To investigate how the model learns the probabilistic structure of the training data, we conducted a first analysis in the form of prior regeneration. For prior regeneration we choose one training sample and use two time steps of the adaption variable 𝐀1:2𝐗¯\mathbf{A}_{1:2}^{\bar{\mathbf{X}}} to initialize the prior distribution p(𝐳1:2)p(\mathbf{z}_{1:2}) in the PV-RNN. Thereafter the future prediction 𝐗3:400{\mathbf{X}}_{3:400} for the remaining training sample length can be calculated (cf. prior generation in Fig. 1). Using this scheme, we generated 20 sequences for each meta-prior 𝐰\mathbf{w}. This was repeated for each network that was trained for that parameter for all random seeds. For brevity, training analysis is reported only for the network that was trained on the probabilistic sequence A20%B80%C. Training of A80%B20%C showed comparable results. An Echo State Network for multivariate time series classification [29] with reservoir size N=45N=45, 25%25\% connectivity and leakage 60%60\% was used for classification of movement primitives. Movement patterns were identified as not classified below a 55%55\% threshold.

III-C1 Analysis of Probabilistic Transition

A robot that is trained with A20%20\%B80%80\%C will first generate an A movement, and then transition to B with 20 percent probability and to C with 80 percentage probability. We found that smaller 𝐰\mathbf{w} settings are less stable in reproducing the probabilistic structure of the training data. The BC-ratio was either greater or less than 20%20\% for B or greater or less than 80%80\% for C. Networks trained with larger meta-priors become more reliable in regenerating the probabilistic training sequence (BC-ratio in Table II). In addition to the capacity of learning the probability distribution of the training data, we found that smaller meta-priors show noisier pattern generation. Non-classified movements were as high as 22%±422\%\pm 4 with 𝐰=0.01\mathbf{w}=0.01 and decreased to 6%±0.66\%\pm 0.6 with 𝐰=3.4\mathbf{w}=3.4.

III-C2 Divergence Analysis

Repeatability in generating sequences in prior generation was examined by conducting a divergence analysis. Sequences are considered diverged when a comparison per time step of 𝐗pr\mathbf{X}^{pr} exceeds a threshold333As the threshold for the divergence analysis, we use the mean squared error of joint angle data [180,180][-180,180] of 𝐗pr\mathbf{X}^{pr}. The threshold is set to 5555.. Out of 20 regeneration sequences, we randomly select one as a reference and calculate the average divergence step of the other sequences to this the reference. Out of 400 time steps of prior generation, sequences diverged from the reference around time step 4343 for networks trained with smaller 𝐰\mathbf{w}. With increasing 𝐰\mathbf{w}, repeatability of the trajectories increased. Here the divergence step was around 139139 (cf. divergence step tt in Table II).

III-C3 Summary of Preparatory Analysis

Loose regulation of the complexity term results in noisier, less repeatable prior generation performance. Also the learned probability for transition to either B or C is not accurate. This observation changes with increasing meta-prior. The larger 𝐰\mathbf{w}, the more accurate the learned transition probability becomes. Also, prior generation becomes more repeatable by developing more deterministic dynamics with the initial sensitivity characteristics (i.e., the sequence is generated solely depending on the latent state in the initial time step). For subsequent dyadic robot interaction experiments, we empirically select the meta-prior setting 𝐰=0.005\mathbf{w}=0.005 and 𝐰=3.4\mathbf{w}=3.4 as two representatives of tight and loose regulation of the FEP complexity.

III-D Dyadic Robot Interaction Experiments

In the following experiments, robots are either trained with 𝐰=0.005\mathbf{w}=0.005 or 𝐰=3.4\mathbf{w}=3.4. For readability, we will consider 𝐑w1\mathbf{R}^{1}_{w} and 𝐑w2\mathbf{R}^{2}_{w} with subscripts of the respective meta-priors 𝐰\mathbf{w}. In the dyadic interaction, we present the network of each robot with observations of movements of the counterpart robot 𝐗¯ex\bar{\mathbf{X}}^{ex} as the target and perform posterior inference in a regression window with size winsize=70win_{size}=70. Inference is performed from the current time step tt back to twinsizet-win_{size}, or t1t_{1} in case twinsize1t-win_{size}\leq 1. After 200 epochs of iteration to maximize the lower bound, the time window is shifted one time step forward. Note, all experiments were conducted in simulation due to the difficulty of real-time posterior inference computation.

We investigated how two robots interact in three different dyadic conditions (TABLE III). We then analysed whether the robots trained with A80%B20%C maintained the learned preference between B and C or adapted to their counterparts that were trained with A20%B80%C. We also calculated the so-called BC-synchronization rate during the interaction. If at any time step tt, one of the robots generated B or C and the other robot generated the same movement primitive, the interaction was considered synchronized. Note that time steps in which movement patterns were identified as not classified by the Echo State Network (cf. subsection III-C) were excluded from the computation.

TABLE III: Interaction performance of three experimental settings.
Experiment BC-ratio BC-sync
ID robots stand-alone interaction
B C B C
1 𝐑0.0051\mathbf{R}^{1}_{0.005} 22±622\pm 6 78±578\pm 5 70±1170\pm 11 30±1030\pm 10 56±2356\pm 23
𝐑3.42\mathbf{R}^{2}_{3.4} 75±275\pm 2 25±325\pm 3 73±1073\pm 10 27±1027\pm 10
2 𝐑3.41\mathbf{R}^{1}_{3.4} 24±224\pm 2 76±276\pm 2 17±1217\pm 12 83±1883\pm 18 31±2431\pm 24
𝐑3.42\mathbf{R}^{2}_{3.4} 75±275\pm 2 25±325\pm 3 61±1261\pm 12 39±1239\pm 12
3 𝐑0.0051\mathbf{R}^{1}_{0.005} 22±622\pm 6 78±578\pm 5 52±1352\pm 13 48±948\pm 9 42±2042\pm 20
𝐑0.0052\mathbf{R}^{2}_{0.005} 44±1144\pm 11 56±1156\pm 11 20±720\pm 7 80±880\pm 8

TABLE III shows the summary of the analysis for all three experiments. To better understand effects of loose and tight regulation of FEP complexity, exemplar plots of robot movement patterns, as well as corresponding network dynamics, are shown (cf. Fig. 4 and Fig. 5). We provide supplementary movies of the experiments showing humanoid robot interaction and network dynamics here: https://doi.org/10.6084/m9.figshare.14099537.

\includegraphics

[trim=0 0 0 -5mm,width=]figures/interaction_weak-strong.pdf

(a) Experiment 1: 𝐑0.0051\mathbf{R}^{1}_{0.005} vs 𝐑3.42\mathbf{R}^{2}_{3.4}
\includegraphics

[width=]figures/interaction_strong-strong.pdf

(b) Experiment 2: 𝐑3.41\mathbf{R}^{1}_{3.4} vs 𝐑3.42\mathbf{R}^{2}_{3.4}
\includegraphics

[width=]figures/interaction_weak-weak.pdf

(c) Experiment 3: 𝐑0.0051\mathbf{R}^{1}_{0.005} vs 𝐑0.0052\mathbf{R}^{2}_{0.005}
Figure 4: Movement trajectories and network dynamics of dyadic robot interaction for Experiment 1 (a), Experiment 2 (b) and Experiment 3 (c). Time steps t=[100,300]t=[100,300] of movements and selected neurons in layer 1 and 3 are shown.

III-D1 Experiment 1: 𝐑0.0051\mathbf{R}^{1}_{0.005} vs. 𝐑3.42\mathbf{R}^{2}_{3.4}

In Experiment 1, 𝐑0.0051\mathbf{R}^{1}_{0.005} adapts to the probabilistic transition of 𝐑3.42\mathbf{R}^{2}_{3.4} by increasing the probability of performing B from 22%22\% in the stand-alone condition to 70%70\% in the dyad (Table III Experiment 1). Both robots are performing more B than C with a BC-synchronization of 56±23%56\pm 23\% which is significantly higher than the chance rate of 32%32\%444We assume that B and C are independent probabilistic events. Then we can consider the probabilities for a robot RR to perform either a B movement as 𝐏R(B)\mathbf{P}^{R}(B) or a C movement as 𝐏R(C)\mathbf{P}^{R}(C). The actual BC-synchronization chance level can then be calculated as: 𝐏1(B)×𝐏2(B)+𝐏1(C)×𝐏2(C)=0.8×0.2+0.8×0.2=0.16+0.16=0.32\mathbf{P}^{1}(B)\times\mathbf{P}^{2}(B)+\mathbf{P}^{1}(C)\times\mathbf{P}^{2}(C)=0.8\times 0.2+0.8\times 0.2=0.16+0.16=0.32..

Fig. 5 shows an example of how prediction of the future and posterior inference of the past proceed as time passes from time step 199199, 229229, to 259259 for both robots. We observe that the intended future behavior (the prior generation) of 𝐑0.0051\mathbf{R}^{1}_{0.005} is not consistent with the actually performed actions after posterior inference. On the other hand, in the case of 𝐑3.42\mathbf{R}^{2}_{3.4}, the performed action complies with its prediction.

\includegraphics

[trim=0 10 0 -5mm,width=.48]figures/interaction_weak-strong_prediction.pdf

Figure 5: Posterior inference and prior generation in Experiment 1. Interaction of 𝐑0.0051\mathbf{R}^{1}_{0.005} (upper) and 𝐑3.42\mathbf{R}^{2}_{3.4} (lower) in terms of 𝐗pr\mathbf{X}^{pr}. The first, the second, and the third row show 𝐗pr\mathbf{X}^{pr} after the posterior inference in the inference window with size winsize=70win_{size}=70, as well as its future prior generation with current time steps of 199, 229, and 259, respectively.

This behavior can be explained by looking at exemplar priors μilp\mu_{i}^{lp} and posteriors μilq\mu_{i}^{lq} for layer ll and neuron ii between two robots. In layer 11, selected posterior network dynamics μ11q\mu_{1}^{1q} and μ21q\mu_{2}^{1q} are deviating from prior dynamics μ11p\mu_{1}^{1p} and μ21p\mu_{2}^{1p} for robot 𝐑0.0051\mathbf{R}^{1}_{0.005}. Whereas the dynamics of 𝐑3.42\mathbf{R}^{2}_{3.4} are mostly overlapping (cf. Fig. 4(a) and supplementary movie). More specifically, the average KL Divergence 𝐞z\mathbf{e}^{z} of 𝐑0.0051\mathbf{R}^{1}_{0.005} is larger for all layers ((𝐞z,1,𝐞z,2,𝐞z,3)=(109.1,1.4,0.06)({\mathbf{e}}^{z,1},{\mathbf{e}}^{z,2},{\mathbf{e}}^{z,3})=(109.1,1.4,0.06)) than for 𝐑3.42\mathbf{R}^{2}_{3.4} ((𝐞z,1,𝐞z,2,𝐞z,3)=(0.4,0.0003,0.00001)({\mathbf{e}}^{z,1},{\mathbf{e}}^{z,2},{\mathbf{e}}^{z,3})=(0.4,0.0003,0.00001)). This means that 𝐑3.42\mathbf{R}^{2}_{3.4} tends to behave as intended because the posterior is attracted by the prior. On the other hand, 𝐑0.0051\mathbf{R}^{1}_{0.005} tends to adapt to 𝐑3.42\mathbf{R}^{2}_{3.4} since the posterior is rather attracted by the observation than by the weaker prior belief.

Note that μ13q\mu_{1}^{3q} and μ13p\mu_{1}^{3p} in layer 3 change only slowly with time. This indicates that these latent variables represent how movement primitives transit from deterministic states to non-deterministic states using their slower timescale properties characterized by τ3\tau^{3}.

III-D2 Experiment 2: 𝐑3.41\mathbf{R}^{1}_{3.4} vs. 𝐑3.42\mathbf{R}^{2}_{3.4}

When two robots with loose complexity regulation interact, both robots maintain their learned preferences in terms of probability in generating either B or C. 𝐑3.41\mathbf{R}^{1}_{3.4}, which learns a 76%76\% transition to C in a stand-alone situation, shows its preference to C in the dyad with probability of 83%83\%. 𝐑3.42\mathbf{R}^{2}_{3.4}, which in a stand-alone condition would maintain its preference to B with a probability of 75%75\%, shows 61%61\% percentage transition to B in the interaction. BC-synchronization rate turns out to be low as 31±24%31\pm 24\%, which is almost equal to the chance rate. Examining the network dynamics of the prior and posterior distributions shows that the robots executed movements based upon their prior action intention without adapting their posteriors to observations of the other robot’s movement (cf. Fig. 4(b) and supplementary movie).

III-D3 Experiment 3: 𝐑0.0051\mathbf{R}^{1}_{0.005} vs. 𝐑0.0052\mathbf{R}^{2}_{0.005}

When two robots with tight regulation of complexity interact, both try to adapt their own action to the one demonstrated by the other. Indeed, Fig. 4(c) shows that the prior and posterior do not comply, but deviate. Whether trained with the probabilistic transition of A20%B80%C or A80%B20%C, both robots significantly reduce the tendency to perform their own intended behavior C or B, respectively. This is evidenced by changes of the BC-ratio from stand-alone compared to the dyadic setting (TABLE III Experiment 3). BC-synchronization rate is 42±20%42\pm 20\% which is higher than the chance rate but not significantly. The interaction becomes noisier, compared to results of Experiments 1 and 2 (cf. Fig. 4(c) and supplementary movie), which indicate that tight regulation makes robots more sensitive to temporal fluctuations in observations of their counterparts.

IV Discussion

The current study examined how social interaction in robotic agents dynamically changes depending on how the complexity in the free energy is regulated. For this purpose, we conducted simulation experiments on dyadic imitative interactions using humanoid robots equipped with PV-RNN architectures. PV-RNN is a hierarchically organized variational RNN model that employs a framework of predictive coding and active inference based on the free energy principle. In a preparatory analysis we showed that PV-RNNs trained with looser regulation of complexity develop stronger action intentions by self-organizing more deterministic dynamics with strong initial sensitivity. Networks trained with tighter regulation develop weaker intentions by self-organizing more stochastic dynamics.

Our experiments revealed different types of interactions between robots. In the experiment where a robot having looser regulation interacts with a robot with tighter regulation, the former tends to lead the interaction by exerting action intention with stronger belief, while the latter tends to follow the other. The following robot adapts its posterior to its observations of the leading robot. In this setting, the synchronization of movement B and C (BC-synchronization rate) between the two robots was significantly higher than the chance rate. When two robots with looser regulation, i.e. intentions with stronger belief, interact, each robot tends to generate its own intended movements. Finally, in case both robots have tighter regulation, a fluctuating dyadic interaction develops where each robot attempts to adapt to the counterpart with an intention with weaker belief. In the last two cases, the BC-synchronization rate was not significantly higher than the chance rate. It can be summarized that the dyadic imitative interaction, including situations where the other’s movements are unpredictable, tends to be synchronized successfully when a dedicated leader and follower are determined; a leader develops action intentions with strong belief whereas a follower develops action intentions with weak belief.

The readers may ask why tighter or looser regulation of the complexity term results in development of weaker or stronger belief of action intention for each robot. Let us consider a situation in which the PV-RNN learns to predict probabilistic sequences 𝑿¯1:T\bm{\bar{X}}_{1:T} with meta-prior 𝐰\mathbf{w} set either with a large value (loose regulation) or a smaller one (tight regulation). The learning process infers the posterior mean 𝝁tq\bm{\mu}^{q}_{t} and standard deviation 𝝈tq\bm{\sigma}^{q}_{t} at each time step tt. In order to minimize the error 𝐞\mathbf{e} in the accuracy term, 𝝁tq\bm{\mu}^{q}_{t} is fitted with an arbitrary value, where 𝝈tq\bm{\sigma}^{q}_{t} will be minimized, in both cases. Notably, when the data 𝑿¯t{\bm{\bar{X}}_{t}} is observed as random, the corresponding posterior 𝝁tq\bm{\mu}^{q}_{t} also becomes random.

Let us consider the two cases when the meta-prior 𝐰\mathbf{w} is either set large or small. In case 𝐰\mathbf{w} is set large, the KL Divergence between the posterior and the prior is strongly minimized. Thus, 𝝁tp\bm{\mu}^{p}_{t} and 𝝈tp\bm{\sigma}^{p}_{t} of the prior latent state become close to 𝝁tq\bm{\mu}^{q}_{t} and 𝝈tq\bm{\sigma}^{q}_{t} of the posterior. By this, 𝝈tp\bm{\sigma}^{p}_{t} in the prior is forced to take a minimal value close to 0; therefore, the prior generation becomes deterministic. Since 𝝁1:Tp\bm{\mu}^{p}_{1:T} should be reconstructed as close to the sequence 𝝁1:Tq\bm{\mu}^{q}_{1:T} inferred with randomness, the prior generative model is forced to develop strongly nonlinear deterministic dynamics with the initial sensitivity through learning. On the other hand, if 𝐰\mathbf{w} is set with a small value, the KL Divergence is only weakly minimized. In this case, prior 𝝁tp\bm{\mu}^{p}_{t} and 𝝈tp\bm{\sigma}^{p}_{t} can diverge from the posterior ones; therefore, the learning becomes ”relaxed”. As a result, the prior generative model develops stochastic dynamics with only weak non-linearity, wherein 𝝁tp\bm{\mu}^{p}_{t} takes an average of 𝝁tq\bm{\mu}^{q}_{t} over all occurrences and 𝝈tp\bm{\sigma}^{p}_{t} takes their distribution at each time step. Consequently, with larger 𝐰\mathbf{w}, the generative model develops action intention with stronger belief (i.e. smaller 𝝈p\bm{\sigma}^{p}) whereas in the case of tighter regulation using a smaller 𝐰\mathbf{w}, the generative model develops action intention with weaker belief (i.e. larger 𝝈p\bm{\sigma}^{p}).

The current experiments consider a fixed meta-prior setting only. Since the meta-prior is the essential network parameter to guide the strength of action intention in the proposed framework, future studies should target meta-learning of the meta-prior in developmental processes or through autonomous adaption within dyadic contexts. This could provide further understanding of more complex social interaction phenomena, including turn-taking in the context of adaptive regulation of the complexity term in free energy.

ACKNOWLEDGMENT

We thank all members of the Cognitive Neurorobotics Research Unit. Special thanks goes to Fabien Benuerau and Jeffrey Queisser, for fruitful discussions about developing the computational model. The authors also acknowledge the support of the Scientific Computation and Data Analysis section of OIST to carry out the research presented here.

References

  • [1] K. Friston, “A theory of cortical responses,” Philosophical transactions of the Royal Society B: Biological sciences, vol. 360, no. 1456, pp. 815–836, 2005.
  • [2] K. Friston, J. Mattout, and J. Kilner, “Action understanding and active inference,” Biological cybernetics, vol. 104, pp. 137–160, 2011.
  • [3] Y. Kuniyoshi, M. Inaba, and H. Inoue, “Learning by watching: extracting reusable task knowledge from visual observation of human performance,” IEEE Transactions on Robotics and Automation, vol. 10, no. 6, pp. 799–822, 1994.
  • [4] G. Hayes and Y. Demiris, “A robot controller using learning by imitation,” Citeseer, vol. 676, 1995.
  • [5] K. Dautenhahn, “Getting to know each other - artificial social intelligence for autonomous robots.” Robotics and Autonomous System, vol. 16, no. 2–4, pp. 333–356, 1995.
  • [6] P. Gaussier, S. Moga, M. Quoy, and J. P. Banquet, “From perception-action loops to imitation processes: A bottom-up approach of learning by imitation,” Applied Artificial Intelligence, vol. 12, no. 7-8, pp. 701–727, 1998.
  • [7] G. Di Pellegrino, L. Fadiga, L. Fogassi, V. Gallese, and G. Rizzolatti, “Understanding motor events: a neurophysiological study,” Experimental brain research, vol. 91, no. 1, pp. 176–180, 1992.
  • [8] M. Arbib, “The mirror system, imitation, and the evolution of language.” Imitation in animals and artefacts, pp. 229–280, 2002.
  • [9] E. Oztop, M. Kawato, and M. Arbib, “Mirror neurons and imitation: A computationally guided review,” Neural networks, vol. 19, no. 3, pp. 254–271, 2006.
  • [10] T. Inamura, Y. Nakamura, H. Ezaki, and I. Toshima, “Imitation and primitive symbol acquisition of humanoids by the integrated mimesis loop,” in Proceedings of 2001 IEEE International Conference on Robotics and Automation, vol. 4, 2001, pp. 4208–4213.
  • [11] A. Billard and M. Mataric, “Learning human arm movements by imitation: Evaluation of a biologically-inspired connectionist architecture,” Robotics and Autonomous Systems, vol. 941, pp. 1–16, 2001.
  • [12] E. Oztop, T. Chaminade, G. Cheng, and M. Kawato, “Imitation bootstrapping: experiments on a robotic hand,” in 5th IEEE-RAS Int. Conf. on Humanoid Robots, 2005., 2005, pp. 189–195.
  • [13] M. Ito and J. Tani, “On-line imitative interaction with a humanoid robot using a dynamic neural network model of a mirror system,” Adaptive Behavior, vol. 12, no. 2, pp. 93–115, 2004.
  • [14] J. Hwang, J. Kim, A. Ahmadi, M. Choi, and J. Tani, “Dealing with large-scale spatio-temporal patterns in imitative interaction between a robot and a human by using the predictive coding framework,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 50, no. 5, pp. 1918–1931, 2020.
  • [15] A. Clark, Surfing uncertainty: Prediction, action, and the embodied mind.   Oxford University Press, 2015.
  • [16] A. Ahmadi and J. Tani, “A novel predictive-coding-inspired variational rnn model for online prediction and recognition,” Neural computation, vol. 31, no. 11, pp. 2025–2074, 2019.
  • [17] S. Murata, Y. Yamashita, H. Arie, T. Ogata, S. Sugano, and J. Tani, “Learning to perceive the world as probabilistic or deterministic via interaction with others: A neuro-robotics experiment,” IEEE Transactions on Neural Networks and Learning Systems, vol. 28, no. 4, pp. 830–848, 2017.
  • [18] P. Lanillos and G. Cheng, “Adaptive robot body learning and estimation through predictive coding,” CoRR, vol. abs/1805.03104, 2018.
  • [19] C. Lang, G. Schillaci, and V. Hafner, “A deep convolutional neural network model for sense of agency and object permanence in robots,” 2018 Joint IEEE 8th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob), pp. 257–262, 2018.
  • [20] A. Philippsen and Y. Nagai, “Deficits in prediction ability trigger asymmetries in behavior and internal representation,” Frontiers in Psychiatry, vol. 11, p. 1253, 2020.
  • [21] Y. Yamashita and J. Tani, “Emergence of functional hierarchy in a multiple timescale neural network model: a humanoid robot experiment,” PLoS computational biology, vol. 4, no. 11, 2008.
  • [22] H. F. Chame and J. Tani, “Cognitive and motor compliance in intentional human-robot interaction,” arXiv preprint arXiv:1911.01753, 2019, accepted for publication in IEEE ICRA2020.
  • [23] W. Ohata and J. Tani, “Investigation of the sense of agency in social cognition, based on frameworks of predictive coding and active inference: A simulation study on multimodal imitative interaction,” Frontiers in Neurorobotics, vol. 14, p. 61, 2020.
  • [24] L. Pio-Lopez, A. Nizard, K. Friston, and G. Pezzulo, “Active inference and robot control: a case study,” Journal of the Royal Society Interface, vol. 16, 2016.
  • [25] G. Schillaci, A. Ciria, and B. Lara, “Tracking emotions: Intrinsic motivation grounded on multi-level prediction error dynamics,” 10th Joint IEEE ICDL-EPIROB, pp. 1–8, 2020.
  • [26] J. Chung, K. Kastner, L. Dinh, K. Goel, A. C. Courville, and Y. Bengio, “A recurrent latent variable model for sequential data,” in Advances in neural information processing systems, 2015, pp. 2980–2988.
  • [27] D. P. Kingma and M. Welling, “Auto-encoding variational bayes,” 2014.
  • [28] D. E. Rumelhart, G. E. Hinton, and R. J. Williams, “Learning internal representations by error propagation,” California Univ San Diego La Jolla Inst for Cognitive Science, Tech. Rep., 1985.
  • [29] F. M. Bianchi, S. Scardapane, S. Løkse, and R. Jenssen, “Reservoir computing approaches for representation and classification of multivariate time series,” IEEE transactions on neural networks and learning systems, 2020.