This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Persistent-Transient Duality in Human Behavior Modeling

Hung Tran, Vuong Le, Svetha Venkatesh, Truyen Tran
Applied AI Institute, Deakin University, Geelong, Australia
{tduy,vuong.le,svetha.venkatesh,truyen.tran}@deakin.edu.au
Abstract

We propose to model the persistent-transient duality in human behavior using a parent-child multi-channel neural network, which features a parent persistent channel that manages the global dynamics and children transient channels that are initiated and terminated on-demand to handle detailed interactive actions. The short-lived transient sessions are managed by a proposed Transient Switch. The neural framework is trained to discover the structure of the duality automatically. Our model shows superior performances in human-object interaction motion prediction.

1 Introduction

Figure 1: We model the Persistent-transient duality of human behaviors. In the default Persistent process (lower row), the human subjects consider the whole scene from the global view with dense relational connections (purple links). For an interactive action, they switch on a Transient process (upper row) that zooms into a small set of local objects that are directly interacted. When done with such task, they switch back to the default Persistent mode.

Human behavior is highly contextualized, reacting to the rapid changes in the situation. This fast-changing nature requires a model to adapt quickly in structure, representation, and inference mechanism to follow the true patterns of the behavior. Such requirement is critical in the problem of human-object interaction (HOI) motion prediction, where the subjects follow an overall plan but occasionally deviate from it to solve emerging tasks[1].

Motion models on this task reflect the changing situations by gradually adapting their relational structure. However, with a fixed inference mechanism, they cannot account for the discrete switching between distinctive mechanisms and fail to keep up with the movement patterns. For example, when the human subject deviates from the overall path to interact with an object, these models will continue to consider the interacted object as an equal member of the scene and miss its importance as the action’s direct recipient.

We address this limitation by factorizing the human behaviors into two processes: a slow-changing persistent process, which maintains a continuous default dynamic, and a fast-changing transient process, which has an adaptive life-cycle and a personalized structure that reflects the human’s perspective in emerging events (See Fig. 1). The two processes are modeled into a parent-child multi-channel neural network called Persistent-Transient Duality. The Persistent channel is a recurrent relational network operating on the global scene spatially and throughout the session temporally. The Transient channels instead have a contextualized graphical structure constructed on the spot whenever the human subjects shift the priority toward interacting with other entities. The life cycles of these channels are managed by a neural Transient Switch, which can learn to anticipate when a Transient channel will be needed and trigger it in time.

Our model establishes the SOTA performance on the HOI subset of the KIT Whole-Body Human Motion Database.

2 Method

Figure 2: The architecture of Persistent-transient Duality Networks (PTD). The persistent channel contains fully-connected recurrent graph networks of all entities (squares) similar to current SoTA works. We introduce the new Transient channel to zoom into the local context of each human (circles) when they interact with surrounding entities (triangles). The transient channels are initialized and terminated on-demand, controlled by the neural Transient Switches (diamonds).

2.1 Preliminaries

We consider the problem of modeling the sequential behaviors of NN entities in a dynamic system, where each entity ii is represented by a class label cic_{i} and sequential features Xi={xit}t=1TX_{i}=\left\{x_{i}^{t}\right\}_{t=1}^{T}. After observing TT steps, we predict the features in the next LL steps, Yi={yit}t=T+1T+L+1Y_{i}=\left\{y_{i}^{t}\right\}_{t=T+1}^{T+L+1}. The entity classes include the human and object (cic_{i}=”human”/”object”), which decide the entity’s feature spaces and behaviors.

In this paper, we use a customized attention function defined over the query qRdq\in R^{d} and the identical key/value pairs V={vj}j=1NRd×NV=\{v_{j}\}_{j=1}^{N}\in R^{d\times N}:

Attn(q,V)=σ(j=1Nsoftmaxj(𝑾αT[𝑾qq;𝑾vvj])𝑾vvj)\displaystyle\text{Attn}\left(q,V\right)=\sigma\left(\sum_{j=1}^{N}\text{softmax}_{j}\left(\boldsymbol{W}_{\alpha}^{T}\left[\boldsymbol{W}_{q}q;\boldsymbol{W}_{v}v_{j}\right]\right)\boldsymbol{W}_{v}v_{j}\right)

where [;]\left[\cdot;\cdot\right] is the concatenation, σ\sigma is a non-linear activation function, 𝑾v\boldsymbol{W}_{v}, 𝑾q\boldsymbol{W}_{q}, and 𝑾α\boldsymbol{W}_{\alpha} are learnable weights.

2.2 The Persistent-Transient Duality

We model the persistent-transient duality of human behavior by a hierarchical neural network called Persistent-Transient Duality Networks (PTD) (See Fig. 2). The network has three main components: The Persistent Channel, the Transient Channels, and the Transient Switch.

2.3 Persistent Channel

The Persistent channel oversees the global view of the scene, including the humans and other entities. It has the form of a recurrent relational network where entities interact in the spatio-temporal space spanned by the video. The temporal evolution of each entity is modeled as:

hi𝒫,t=RNNci([zi𝒫,t,mi,𝒯𝒫t],hi𝒫,t1),h_{i}^{\mathcal{P},t}=\text{RNN}_{c_{i}}\left(\left[z_{i}^{\mathcal{P},t},m_{i,\mathcal{T}\rightarrow\mathcal{P}}^{t}\right],h_{i}^{\mathcal{P},t-1}\right), (1)

where RNNci\text{RNN}_{c_{i}} is a recurrent unit that corresponds to the class of the ithi^{\text{th}} entity, maintaining hidden states hi𝒫,t1h_{i}^{\mathcal{P},t-1} and consuming input vector zi𝒫,tz_{i}^{\mathcal{P},t}. This input is formed as zi𝒫,t=[xit;mit]z_{i}^{\mathcal{P},t}=\left[x_{i}^{t};m_{i}^{t}\right], where xitx_{i}^{t} is the entity’s feature and mitm_{i}^{t} is the message aggregated from the entity’s neighbor through attention, mit=Attn(uit,{ujt}ji)m_{i}^{t}=\text{Attn}\left(u_{i}^{t},\left\{u_{j}^{t}\right\}_{j\neq i}\right) with u_t=[x_t;h_𝒫,t1]u\_^{t}=\left[x\_^{t};h\_^{\mathcal{P},t-1}\right].

The unit also uses an optional transient-persistent message mi,𝒯𝒫tm_{i,\mathcal{T}\rightarrow\mathcal{P}}^{t} which is non-zero if a corresponding transient process (Sec. 2.4) is currently active.

The Persistent channel generates two outputs from its hidden state: the future prediction y^i𝒫,t\hat{y}_{i}^{\mathcal{P},t} and the message mi,𝒫𝒯tm_{i,\mathcal{P}\rightarrow\mathcal{T}}^{t} to the Transient channel:

y^i𝒫,t=MLP(hi𝒫,t),mi,𝒫𝒯t=MLP(hi𝒫,t).\hat{y}_{i}^{\mathcal{P},t}=\text{MLP}\left(h_{i}^{\mathcal{P},t}\right),\hskip 8.5359ptm_{i,\mathcal{P}\rightarrow\mathcal{T}}^{t}=\text{MLP}\left(h_{i}^{\mathcal{P},t}\right). (2)

The channel’s prediction output y^i𝒫,t\hat{y}_{i}^{\mathcal{P},t} are combined with those from the Transient channel as detailed in Sec. 2.6.

This persistent channel is equivalent in modeling with the major state-of-the-art HOI-M by using relational recurrent models [2]. Our key novelty is the consideration of the second side of the duality - the Transient process, presented in the next section.

2.4 Transient Channel

Within the persistent-transient duality, the Transient process allows the model to zoom in at relevant context and take the local viewpoint of the human entity when it starts to interact with objects. This human-specific process is implemented by a Transient channel. The egocentric property of this channel separates it from the global view of its parent persistent channel and reflects in three aspects of feature representation, computational structure, and inference logic.

The egocentric graph structure

reflects the relations between the active subject and the surrounding passive entities. We define the Transient graph for a human entity of index ii at time tt to be 𝒢it=(𝒱it,it)\mathcal{G}_{i}^{t}=\left(\mathcal{V}_{i}^{t},\mathcal{E}_{i}^{t}\right). For a particular human, subscript ii will be omitted for conciseness.

The egocentric characteristic of 𝒢t\mathcal{G}^{t} reflects in its star-like structure: the nodes 𝒱t\mathcal{V}^{t} includes a single center node rr for the considering human, and leaf nodes of indices {l}lr\left\{l\right\}_{l\neq r} for other entities. The dynamic edges t\mathcal{E}^{t} connect the center with the leaves in two directions: inward edges elrte_{l\rightarrow r}^{t} reflect which objects the human pay attention to, and the outward edges erlte_{r\rightarrow l}^{t} represents the objects are being manipulated by the human. These edges are determined at each time step by thresholding the center-leaf distances dlrtd_{lr}^{t}, making the graph’s topology evolve within one single Transient session.

(a) Transient Channel
(b) Transient Switch
Figure 3: The Transient Channel (a) and Transient Switch (b)

The Egocentric representation

of the entities are computed by transforming the geometrical features into the egocentric coordinate system corresponding to the viewpoint of the human center node rr, x¯_t=fego(x_t,xrt)=x_tcentroid(xrt)\bar{x}\_^{t}=f_{\text{ego}}\left(x\_^{t},x_{r}^{t}\right)=x\_^{t}-\text{centroid}(x_{r}^{t}). This change of system puts various patterns of the human’s motion into the same aligned space, filters out the irrelevant global information, and facilitates efficient inference of the egocentric model.

The Egocentric inference

is made by updating the RNN hidden state hth_{-}^{t} of each node in the transient graph structure (see Fig. 3a). In detail, for the center node, the inward messages from its leaves are aggregated into: mlrt=Attn([x¯rt;hrt1],{[x¯lt;hlt1]}elrtt)m_{l\rightarrow r}^{t}=\text{Attn}\left(\left[\bar{x}_{r}^{t};h_{r}^{t-1}\right],\left\{\left[\bar{x}_{l}^{t};h_{l}^{t-1}\right]\right\}_{e_{l\rightarrow r}^{t}\in\mathcal{E}^{t}}\right). It is then combined with the egocentric features x¯rt\bar{x}_{r}^{t} and the message from the persistent channel m𝒫𝒯tm_{\mathcal{P}\rightarrow\mathcal{T}}^{t} (Eq. 2) to update the RNN:

zrt=[x¯rt;mlrt;m𝒫𝒯t],hrt=RNNr(zrt,hrt1).z_{r}^{t}=\left[\bar{x}_{r}^{t};m_{l\rightarrow r}^{t};m_{\mathcal{P}\rightarrow\mathcal{T}}^{t}\right],\hskip 8.5359pth_{r}^{t}=\text{\text{RNN}}_{r}\left(z_{r}^{t},h_{r}^{t-1}\right). (3)

For leaf nodes, they only update their states if the center node interacts with them, indicated by the outward edge

zlt=[x¯lt;mrlt],hlt={RNNl(zlt,hlt1)if erltthlt1otherwise,z_{l}^{t}=\left[\bar{x}_{l}^{t};m_{r\rightarrow l}^{t}\right],h_{l}^{t}=\begin{cases}\text{\text{RNN}}_{l}\left(z_{l}^{t},h_{l}^{t-1}\right)&\textrm{if }e_{r\rightarrow l}^{t}\in\mathcal{E}^{t}\\ h_{l}^{t-1}&\text{otherwise}\end{cases}, (4)

where mrltm_{r\rightarrow l}^{t} is the outward message from the center to its leaves, calculated from its hidden state through an MLP.

The updated hidden states are used generate the transient predictions y^i𝒯,t\hat{y}_{i}^{\mathcal{T},t} and the messages sent to the persistent process m𝒯𝒫tm_{\mathcal{T}\rightarrow\mathcal{P}}^{t} (used in Eq. 1):

y^𝒯,t\displaystyle\hat{y}_{-}^{\mathcal{T},t} =fego1(MLP(ht)),m𝒯𝒫t=ϕ(MLP(hrt)),\displaystyle=f_{\text{ego}}^{-1}\left(\text{MLP}\left(h_{-}^{t}\right)\right),m_{\mathcal{T}\rightarrow\mathcal{P}}^{t}=\phi\left(\text{MLP}\left(h_{r}^{t}\right)\right), (5)

where fego1f_{\text{ego}}^{-1} is the function converting the egocentric back to global coordinates. The transient predictions y^𝒯,t\hat{y}_{-}^{\mathcal{T},t} are combined with those from the Persistent process in Sec. 2.6.

2.5 Switching Transient Processes

The life cycles of the Transient processes are managed based on the situation of human’s activity by a neural Transient Switch (See Fig. 3b).

The switch first considers the current persistent state and the surrounding environment of the center entity rr to update its switch RNN hrs,th_{r}^{s,t}:

hrs,t\displaystyle h_{r}^{s,t} =RNNs([hr𝒫,t,mrs,t],hrs,t1),\displaystyle=\text{RNN}_{s}\left(\left[h_{r}^{\mathcal{P},t},m_{r}^{s,t}\right],h_{r}^{s,t-1}\right), (6)

where mrs,t=Attn(hrs,t1,{x¯lt}lr)m_{r}^{s,t}=\text{Attn}\left(h_{r}^{s,t-1},\left\{\bar{x}_{l}^{t}\right\}_{l\neq r}\right), and x¯lt=fego(xlt,xrt)\bar{x}_{l}^{t}=f_{\text{ego}}\left(x_{l}^{t},x_{r}^{t}\right) defined in Sec. 2.4.

The RNN unit is important in maintaining the switch’s sequential properties, making it a state-full machine that can handle patterns of on and off switchings and avoid spurious decisions caused by noises. The switch-on probability prtp_{r}^{t} is:

p^rt=γrtsigmoid(Whrs,t),\hat{p}_{r}^{t}=\gamma_{r}^{t}\cdot\text{sigmoid}\left(Wh_{r}^{s,t}\right), (7)

where WW is learnable weights. The discount factor γrt[0,1]\gamma_{r}^{t}\in[0,1] responds to the distance from the subject to the nearest neighbor: γrt=exp(βmin{dlrt2}lr),\gamma_{r}^{t}=\exp\left(-\beta\cdot\text{min}\left\{\left\|d_{lr}^{t}\right\|_{2}\right\}_{l\neq r}\right), where dlrtd_{lr}^{t} are the center-leaf geometrical distances and β\beta is a learnable decay rate. This factor acts as a disruptive shortcut gate that modulates the switching decision based on the spatial evidence of the interaction.

Finally, the binary switch decision s^rt\hat{s}_{r}^{t} is decided by thresholding the switch score: the switch is on (s^rt=1)\hat{s}_{r}^{t}=1) when p^rt>=0.5\hat{p}_{r}^{t}>=0.5, and is off otherwise. When it changes from 0 to 11, a new Transient process is created at time tt for person rr. This transient process will keep running until the switch turns off, then the persistent process again becomes the single operator.

2.6 Future prediction

In PTD, future motions are predicted by unrolling the model into the future of LL time steps. At each future time step tt, the predictions from persistent channel Y^𝒫,t\hat{Y}^{\mathcal{P},t} (Eq. 2) and those from Transient channel Y^𝒯,t\hat{Y}^{\mathcal{T},t} (Eq. 5) are combined with the priority on the Transient predictions. For a human entity, if its Transient channel is activated, the Transient prediction will be chosen; otherwise, the Persistent prediction will be used. For an object entity, if it receives an active outward Transient edge, it will take that channel’s prediction. If it receives multiple outward edges, it uses the prediction from the channel with the highest transient score p^rt\hat{p}_{r}^{t}. Otherwise, it uses the persistent prediction by default.

2.7 Model Training

The model is trained end-to-end with two losses: prediction loss and switch loss, =pred+λswitch.\mathcal{L}=\mathcal{L}_{\text{pred}}+\lambda\mathcal{L}_{\text{switch}}.

The Prediction loss pred\mathcal{L}_{\text{pred}} measures the mismatch between predicted values Y^\hat{Y} and groundtruth YY, pred=Y^T:T+LYT:T+L22.\mathcal{L}_{\text{pred}}=||\hat{Y}^{T:T+L}-Y^{T:T+L}||_{2}^{2}.

The Switch loss switch\mathcal{L}_{\text{switch}} is used to supervise the Transient Switch and is implemented as: switch=BCE(P^1:T+L,P1:T+L)\mathcal{L}_{\text{switch}}=\text{BCE}\left(\hat{P}^{1:T+L},P^{1:T+L}\right), where P^t\hat{P}^{t} are the switch scores (Eq. 7) of all human entities, and PtP^{t}are binary ground-truth switch scores collected at time step tt.

3 Experiment

Figure 4: Quantitative performances and model size in HOI-M.

Experiment settings.

Following the standard protocol [2], we extract a subset of the Whole-Body Human Motion Database (WBHM) [4], which includes scenes that contain body poses of at least one human entity, multiple movable objects, and a stationary object such as “table”. This set includes 233 videos with 20 entity classes. The selected features include 3D skeleton poses of 18 joints for human entities and 3D bounding boxes for objects, sampled at 10Hz, consistent with the compared methods [2].

We compare PTD with CRNN[2], Structural RNN[3], and the simple GRU, whose implementations are redone for consistency. We follow the common settings of observing 10 time steps (1 second) and predicting future human and object motions for the next 20 time steps (2 seconds). The models are conventionally trained on 80% of the videos and evaluated on the remaining 20%.

Quantitative evaluation.

The performances are measured by prediction errors of joint positions in the Euclidean distance (in mm). The means and standard deviations from five independent runs are plotted in Fig. 4 and show that PTD consistently outperforms the state-of-the-art, especially in long-term prediction.

Figure 5: Persistent-transient duality in HOI-M. When the situation changes from interaction-free (a) to interaction-involved (b), PTD (Upper row) switches on its Transient channel with egocentric structures and handles the interaction accurately; In contrast, CRNN [2] (Lower row) uses a single mechanism, resulting in the sluggish adaptation of the attention map, leading to inaccurate predictions.

Visual analysis.

We further verify the benefit of the duality by visualizing the internal output predictions and graph structures of PTD compared to CRNN [2]. The upper row of Fig. 5 shows that PTD could learn to switch from the Persistent dense graph to the Transient egocentric graph when the situation changes from interaction-free (a) to interaction-involved (b). Particularly, in (b), the Transient graph reflects the interactions correctly thanks to it being trained only on targeted interaction samples free of noises.

In contrast, CRNN [2] (lower row) holds on to a single global mechanism and does not evolve adequately for the swift change in the true relational topology, resulting in inaccurate and unrealistic interactions.

4 Conclusion

In this work, we have introduced a new concept of the Persistent-Transient duality to model the interleaving of global dynamics and the short-lived interactions in human behavior. We model this conceptual duality into a parent-child multi-channel network that can switch between the two processes seamlessly. The superior performance of our model on the HOI subset of WBHM confirms the effectiveness of this duality in human behavior modeling.

References

  • [1] Adam P Baker, Matthew J Brookes, Iead A Rezek, Stephen M Smith, Timothy Behrens, Penny J Probert Smith, and Mark Woolrich. Fast transient networks in spontaneous human brain activity. Elife, 3:e01867, 2014.
  • [2] Enric Corona, Albert Pumarola, Guillem Alenya, and Francesc Moreno-Noguer. Context-aware human motion prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6992–7001, 2020.
  • [3] Ashesh Jain, Amir R Zamir, Silvio Savarese, and Ashutosh Saxena. Structural-rnn: Deep learning on spatio-temporal graphs. In Proceedings of the ieee conference on computer vision and pattern recognition, pages 5308–5317, 2016.
  • [4] Christian Mandery, Ömer Terlemez, Martin Do, Nikolaus Vahrenkamp, and Tamim Asfour. The kit whole-body human motion database. In International Conference on Advanced Robotics (ICAR), pages 329–336, 2015.