This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Integrated Decision Making and Trajectory Planning for Autonomous Driving Under Multimodal Uncertainties: A Bayesian Game Approach

Zhenmin Huang, Tong Li, Shaojie Shen, and Jun Ma Zhenmin Huang, Tong Li, Shaojie Shen, and Jun Ma are with the Department of Electronic and Computer Engineering, The Hong Kong University of Science and Technology, Hong Kong SAR, China (e-mail: [email protected]; [email protected]; [email protected]; [email protected]). This work has been submitted to the IEEE for possible publication. Copyright may be transferred without notice, after which this version may no longer be accessible.
Abstract

Modeling the interaction between traffic agents is a key issue in designing safe and non-conservative maneuvers in autonomous driving. This problem can be challenging when multi-modality and behavioral uncertainties are engaged. Existing methods either fail to plan interactively or consider unimodal behaviors that could lead to catastrophic results. In this paper, we introduce an integrated decision-making and trajectory planning framework based on Bayesian game (i.e., game of incomplete information). Human decisions inherently exhibit discrete characteristics and therefore are modeled as types of players in the game. A general solver based on no-regret learning is introduced to obtain a corresponding Bayesian Coarse Correlated Equilibrium, which captures the interaction between traffic agents in the multimodal context. With the attained equilibrium, decision-making and trajectory planning are performed simultaneously, and the resulting interactive strategy is shown to be optimal over the expectation of rivals’ driving intentions. Closed-loop simulations on different traffic scenarios are performed to illustrate the generalizability and the effectiveness of the proposed framework.

Index Terms:
Autonomous driving, game theory, multi-agent systems, decision-making, trajectory planning, game of incomplete information, Bayesian coarse correlated equilibrium, no-regret learning.

I Introduction

Recent years have witnessed great advancements in the technology of autonomous vehicles, which aims to provide a new transportation style that improves traffic efficiency and safety. On the other hand, many technical difficulties remain unsolved in decision-making and trajectory planning for autonomous vehicles across highly complex traffic scenarios. Within this topic, one of the key issues is to effectively model the interaction between autonomous vehicles and other traffic participants, such as human drivers. This problem is particularly challenging, as the rationality of humans is often limited and the preferences of human drivers vary, leading to large uncertainties in decisions and actions. Many of the approaches in the literature adopt an architecture such that prediction and trajectory planning are essentially decoupled. In particular, a prediction module is utilized to forecast the future trajectories of human-driving vehicles. Afterward, the human-driving vehicles are fed into the planning module as immutable objects, and the planning process is grounded on this immutable hypothesis. These approaches are limited as they largely fail to capture the interactive nature of driving behaviors by ignoring the influence of the planned trajectories on other traffic participants. Also, they can lead to the “frozen robot” problem [1] and overly conservative behaviors when all possible trajectories are potentially unsafe according to the predictions.

In light of this, a branch of methods based on game theory is receiving increasing attention owing to its strength in modeling the interactions between agents. The game theory models the self-interested nature of all participants that drives them to a particular equilibrium, which is optimal in the sense that none of the participants can gain more profits by unilaterally altering its own strategy. By calculating the equilibrium of the game, the prediction and the planning are essentially coupled, resulting in an optimal trajectory with interaction properly considered. Efforts have been paid in this direction to model various traffic scenarios such as intersections [2, 3], ramp-merging [4, 5], lane-changing [6, 7], and roundabouts [8]. Among game-based methods, various types of games are intensively studied such as potential games [9, 10], congestion game [11], auction game [12], and Stackelberg games [13, 14]. Nevertheless, the majority of game-based methods focus on achieving a single equilibrium and largely neglect the multimodal nature of the solution. In reality, the decision space of human drivers is often discrete, and multiple distinct decisions can be reasonable with respect to a specific traffic scenario, while a single equilibrium can only reveal one of them. Catastrophic results are prone to occur when there is a discrepancy between the mode of the calculated equilibrium and the actual decisions made by other traffic participants. A few game-based methods aim to solve this problem by considering the multimodality in behaviors of the rival traffic participants [15, 16, 17], but the game settings are simple and scenario-specific, which prevent these methods from generalizing to complex game settings and a variety of traffic scenarios. Also, in these methods, the decision-making and the planning are separated, and therefore the results are prone to be suboptimal.

To address the aforementioned challenges, we present an innovative integrated decision-making and trajectory planning framework for autonomous vehicles, which properly considers the multimodal nature of driving behaviors utilizing game theory. In particular, interactions of traffic participants are modeled as a general Bayesian game (i.e., game of incomplete information) that can potentially be applied to various traffic scenarios. By computing the corresponding Bayesian Coarse Correlated Equilibrium (Bayes-CCE), the decision and the corresponding trajectory are obtained simultaneously, which are shown to be optimal over the expectations of the underlying intentions of other participants. An updating strategy based on Bayesian filtering is also introduced to update the estimation of the driving intentions. The main contribution of this paper is listed as follows:

  • \bullet

    The interactions between traffic participants are effectively modeled from the perspective of a Bayesian game, such that the multimodalities in driving behaviors are properly addressed. Unlike most of the existing methods, the game formulation is generic with multiple game stages and the information structures are potentially adaptive. Therefore, it exhibits the generalizability to various traffic scenarios.

  • \bullet

    A parallelizable Monte Carlo search algorithm is proposed to solve the introduced game with arbitrarily complex information structures. Meanwhile, rigorous proof is established to demonstrate the almost-sure convergence of the introduced algorithm to a Bayes-CCE.

  • \bullet

    With the attained Bayes-CCE, an integrated decision-making and trajectory planning strategy is presented to obtain the decision and the corresponding trajectory simultaneously, which bridges the gap between decision-making and trajectory planning to avoid suboptimality. Moreover, theoretic analysis is performed to demonstrate the optimality of the resulting strategy in terms of the expectation, given the up-to-date beliefs over the underlying driving intentions of other vehicles.

  • \bullet

    With a belief updating strategy based on Bayesian filtering to maintain the estimation of the participants’ driving intentions, a closed-loop autonomous driving framework is given. A series of simulations are performed to illustrate the superiority of the proposed method over traditional game methods.

II Related Works

II-A Non-Game-Theoretic Decision-Making and Trajectory Planning

In the past years, efforts have been made to tackle the decision-making and trajectory-planning problems for autonomous driving. State-machine-based methods [18, 19, 20], which utilize hand-crafted rules to generate decisions and trajectories based on the current state of the autonomous vehicle, are popular at the early stage due to their simplicity and interpretability. However, the innumerable nature of traffic scenarios prevents the wide application of these methods, as the difficulty in maintaining such a state machine increases dramatically with increasing scenario complexities. The research focus soon shifts to a more generic formulation of the problem. In light of this, various methods are proposed by actively employing the concept of searching [21, 22, 23], optimization [24, 25, 26], sampling [27, 28], and POMDP [29], etc. Recently, a branch of multi-policy planning methods, including MPDM [30], EUDM [31], EPSILON [32], and MARC [33], are gaining attention due to their ability in efficient behavior planning. These methods rely on dedicated forward simulators to evaluate the performance of each policy and perform the best one selectively. Nonetheless, the aforementioned approaches generally suffer from difficulties in modeling the interactions, particularly in the effect of the trajectory of the autonomous vehicle on the behaviors of the other vehicles. Meanwhile, a series of recent works address the decision-making and trajectory planning problem using a unified neural network trained through end-to-end imitation learning  [34, 35, 36]. However, the lack of interpretability brings safety concerns, which prevents the wide application of these methods in real-world situations.

II-B Game-Theoretic Decision-Making and Trajectory Planning

To tackle the problem mentioned in the previous subsection, game-based methods are extensively researched. In this area, many of the existing methods calculate a single equilibrium and assume that all participants will follow the computed equilibrium [37, 4, 13, 38, 9]. These methods generally assume no uncertainty in agents’ behaviors. In [37, 4], efficient game-solving strategies are proposed to resolve differential games and reach typical Nash equilibriums. In [13], the interaction between traffic agents is formulated as a Stackelberg game, and a solving scheme based on dynamic programming is proposed to solve the game in real time. In [9], a general decision-making framework is introduced, which shows that typical traffic scenarios can be modeled as a potential game. In potential games, a pure-strategy Nash equilibrium is proven to exist and reachable, and therefore effective decision-making strategies can be derived. In [38], an efficient game-theoretic trajectory planning algorithm is introduced based on Monte Carlo Tree Search, together with an online estimation algorithm to identify the driving preferences of road participants. Although different types of games and various solving schemes are intensively studied in these researches, the uncertainties in human driving behaviors are not effectively revealed in the attained equilibria. Meanwhile, a few studies try to actively model different kinds of uncertainties involved in a game process. In [39], uncertainties in the observation model are actively addressed by formulating the interaction between agents as a stochastic dynamic game in belief space. However, the proposed method only considers observation uncertainties with Gaussian noise. In [40], the maximum entropy model is utilized to model the uncertainties in the equilibrium strategies for all agents. Nevertheless, the objectives and intentions of all agents are assumed to be known and fixed, which is unrealistic in complicated urban traffic scenarios.

To address the uncertainties residing in objectives and obtain the optimal strategy against the resulting multi-modal behaviors of rival agents, some of the researches study the game of incomplete information and the corresponding Bayesian equilibrium, where the equilibrium strategy is optimal over the distribution of rivals’ possible objectives and their equilibrium strategies. Representative works include [15, 17, 41, 42, 16, 43]. Nonetheless, these methods adopt overly simplified game settings of tiny scale and use hand-crafted solvers for obtaining the corresponding Bayesian equilibrium, which confines the application of these methods to simple scenarios like lane-switching. Besides, the decision-making and trajectory-planning processes are usually separated, and this discrepancy can lead to suboptimal decisions. A recent research [44] takes a game-theoretic perspective on contingency planning, and the resulting contingency game enables interaction with other agents by generating strategic motion plans conditioned on multiple possible intents for other actors. However, it is a motion planning framework lacking in the capacity to perform high-level decision-making. Also, the dedicated and unsymmetrical game settings make it hard to generalize to different types of games with multiple players, multiple policies, and multiple stages.

II-C No-Regret Learning

Another branch of methods closely related to this research is no-regret learning, which aims at providing general solutions to various extensive-form games (games with multiple stages). Despite their effectiveness in modeling sequential decision-making processes, solving for large extensive games has been a long-standing challenge. In [45], a counter-factual regret minimization (CFR) algorithm is introduced to iteratively solve this form of games, which is extended by MCCFR [46] with a Monte Carlo sampling process to avoid the difficulties traversing the entire game tree in each iteration. A follow-up work is proposed in [47], which further extends the MCCFR method to enable dynamic construction of the game tree. The convergence of these methods to a corresponding Nash equilibrium is only guaranteed on two-player zero-sum games. To extend these methods to multi-player general-sum games, CFR-S is introduced in [48], which performs sampling at each iteration and illustrates that the frequency distribution of the sampled strategies converges to a Coarse Correlated Equilibrium (CCE). For games of incomplete information with Bayes settings, the condition of convergence to a Bayes-CCE is provided in [49] without detailed solution schemes.

III Problem Statement

Consider a set of NN vehicles 𝒩={1,2,,N}\mathcal{N}=\{1,2,...,N\} in a traffic scenario. For each vehicle i𝒩i\in\mathcal{N}, we denote the set of its potential intentions as TiT_{i}. For each driving intention tiTit_{i}\in T_{i}, the associated action space and utility function are denoted as Ai(ti)A_{i}(t_{i}) and ui(ti)u_{i}(t_{i}), respectively. Further, we use T=×i𝒩TiT=\times_{i\in\mathcal{N}}T_{i} to denote the product space of {Ti}i𝒩\{T_{i}\}_{i\in\mathcal{N}}, such that each tTt\in T is a vector of intentions of all vehicles, t=[ti]i𝒩t=[t_{i}]_{i\in\mathcal{N}}. It should be noted that the intention of vehicle ii, tit_{i}, is private and unknown to all other vehicles and vice versa. Instead, each vehicle is trying to maintain a probability distribution over the intentions of all other vehicles. We further assume that the historical trajectories of all vehicles, together with the surroundings, are fully observable by each vehicle. Since the predictions on driving intentions are determined by historical trajectories and surroundings, we can reasonably assume that all vehicles can arrive at the same belief on the intentions. Formally, the following assumption is introduced.

Assumption 1.

A common probability distribution pp over TT can be established and is known to all vehicles, such that p=[pi]i𝒩p=[p^{i}]_{i\in\mathcal{N}}, where pip_{i} is the common belief of all vehicles j𝒩,jij\in\mathcal{N},j\neq i over the intention of vehicle ii.

Note that although vehicle ii has complete knowledge of its own intention, it also recognizes the fact that its intention is not known to others and that the belief of other vehicles over its own intention is pip_{i}.

With the aforementioned notations and assumptions, we formally introduce the following definition of games with incomplete information.

Definition 1.

(Harsanyi game of incomplete Information [50, Chapter 9.4]) A Harsanyi game of incomplete information is a vector (𝒩,(Ti)i𝒩,p,(gt)tT)(\mathcal{N},(T_{i})_{i\in\mathcal{N}},p,(g_{t})_{t\in T}), where gt=((Ai(ti))i𝒩,(ui(ti))i𝒩)g_{t}=((A_{i}(t_{i}))_{i\in\mathcal{N}},(u_{i}(t_{i}))_{i\in\mathcal{N}}) is the state game for the type vector tt.

Note that the driving intentions of vehicles are modeled as types of players in one-to-one correspondence. Through playing a Harsanyi game of incomplete information, a desirable result is the corresponding Bayesian Nash equilibrium strategy. In particular, a behavior strategy σi\sigma_{i} of vehicle ii, is a map from each type tiTit_{i}\in T_{i} to a probability distribution over the available actions: σi(ti)Δ(Ai(ti))\sigma_{i}(t_{i})\in\Delta(A_{i}(t_{i})). Further, given the strategy vector of all vehicles as σ=[σi]i𝒩\sigma=[\sigma_{i}]_{i\in\mathcal{N}}, we denote the expected payoff of vehicle ii with type tit_{i} as

Ui(σ|ti)=𝔼tipi𝔼aσ(t)ui(ti;a).U_{i}(\sigma|t_{i})=\mathbb{E}_{t_{-i}\sim p_{-i}}\mathbb{E}_{a\sim\sigma(t)}u_{i}(t_{i};a). (1)

The corresponding Bayesian Nash equilibrium is defined as follows.

Definition 2.

(Bayesian Nash Equilibrium [50, Chapter 9.4]) a strategy vector σ=[σi]i𝒩\sigma^{*}=[\sigma_{i}^{*}]_{i\in\mathcal{N}} is a (mixed-strategy) Bayesian Nash Equilibrium if for all i𝒩i\in\mathcal{N}, for all tiTit_{i}\in T_{i}, and for all aiAi(ti)a_{i}\in A_{i}(t_{i}), the following inequality holds:

Ui(σ|ti)Ui((ai,σi)|ti).U_{i}(\sigma|t_{i})\geq U_{i}((a_{i},\sigma^{*}_{i})|t_{i}). (2)

When a Bayesian Nash equilibrium is reached, no player of any type can obtain a higher expected payoff by unilaterally modifying its own behavioral strategy. In the context of autonomous driving, it means that for an arbitrary vehicle ii, once its intention is determined, its strategy is optimal over the expectation of the intentions of all other vehicles, given that their strategies are fixed. Since this optimality condition holds for all vehicles, the bilateral signaling effects are modeled, such that the influence of own actions on the belief of other vehicles about the intention of ego vehicle, and further, their preferences on actions, can be properly considered. Thus, an interaction model that handles multimodality on both sides is established.

However, solving for such a mixed-strategy Bayesian Nash Equilibrium is not trivial, especially for a multi-player general-sum game, which is typically the case in common urban traffic scenarios. To leverage the merit of game of incomplete information, we consider instead the Bayes-CCE, which is a generalization of the Bayesian Nash equilibrium, and propose a general solver that is proven to converge to a Bayes-CCE with minimum assumptions required. To introduce the concept of Bayes-CCE, we first define the plan sis_{i} of player ii as a direct map from type to a specific action: si(ti)Ai(ti)s_{i}(t_{i})\in A_{i}(t_{i}), and further, s=[si]i𝒩s=[s_{i}]_{i\in\mathcal{N}} is the concatenated vector of plans of all players. Σ\Sigma is the set of ss such that sΣs\in\Sigma. The formal definition of a Bayes-CCE is as follows.

Refer to caption
Figure 1: Settings of the game of incomplete information for traffic interaction. The game is an extensive-form game containing multiple stages. In the beginning, chance players select types for all players in a chance stage (corresponding to the chance actions in the figure), where the selections are subject to distribution pp. Then in each of the stages, all players perform a single action in turn. Eventually, the actions of players form a path leading from the root node to one of the leaf nodes, where the payoffs of all players are determined. The notation Ii,knI^{n}_{i,k} represents the kthk^{th} information set of player ii at stage nn, and aI,ma_{I,m} denotes the mthm^{th} action under information set II. Each of the information sets contains one or more states, such that the agent to play, under that information set, cannot distinguish between the game states contained in the information set. Therefore it will take the same action (or the same strategic profile of actions) for all the states contained in the information sets. Further, we assume that at the end of each stage except for the chance stage, all the states are distinguishable, and therefore each game state at the end of a stage constitutes an information set at the beginning of the next stage.
Definition 3.

(Bayes-CCE [49]) A strategy profile ψΔ(Σ)\psi\in\Delta(\Sigma) is a Bayes-CCE if for every player i𝒩i\in\mathcal{N}, for every type tiTit_{i}\in T_{i}, and for every possible plan sis^{\prime}_{i}, the following inequality holds:

𝔼sψ𝔼tipiui(ti;s(ti,ti))\displaystyle\mathbb{E}_{s\sim\psi}\mathbb{E}_{t_{-i}\sim p_{-i}}u_{i}(t_{i};s(t_{i},t_{-i})) \displaystyle\geq (3)
𝔼sψ𝔼tipi\displaystyle\mathbb{E}_{s\sim\psi}\mathbb{E}_{t_{-i}\sim p_{-i}} ui(ti;si(ti),si(ti)).\displaystyle u_{i}(t_{i};s^{\prime}_{i}(t_{i}),s_{-i}(t_{-i})).

In this paper, we formulate the driving game in urban traffic scenarios as a Harsanyi game of incomplete information. Specifically, the action spaces are represented as sets of sampled candidate trajectories, each associated with a particular driving intention (accelerating, lane switching, etc.) The utility functions include pertinent driving performance indices like smoothness and safety. Solving for the Bayes-CCE yields an open-loop equilibrium strategy, which is a joint probability distribution over candidate trajectories. We assume that all vehicles perform this equilibrium strategy. As time evolves and new information is gained, the prediction of vehicle intentions can be updated through Bayesian filtering, and the corresponding driving game can be resolved repeatedly, resulting in a closed-loop driving scheme.

IV Proposed Approach

IV-A Parallel Solver for Bayes-CCE

In this section, a parallel solver utilizing classic Monte Carlo counter-factual regret minimization (MCCFR) is introduced to resolve the game and enable a real-time solution for the open-loop Bayes-CCE in general traffic settings. In particular, the structure of a typical extensive-form game of incomplete information in traffic settings is shown as the game tree in Fig. 1. Each of the nodes on the game tree corresponds to a particular game state, and these game states are partitioned into information sets, such that players cannot distinguish between two states in the same information set. Therefore, the behavior strategy of each player is made with respect to information sets, σi(ti)=[σi(ti,I)]Ii\sigma_{i}(t_{i})=[\sigma_{i}(t_{i},I)]_{I\in\mathcal{I}_{i}}, where i\mathcal{I}_{i} is the information partition of player ii, or the set of information sets under which it is the turn of player ii to move. σi(ti,I)\sigma_{i}(t_{i},I) is a distribution over possible actions Ai(ti,I)A_{i}(t_{i},I) for player ii of type tit_{i} under Information set II. At the beginning stage of the game, an extra chance player (also known as the nature) will take chance moves that will randomly select a type for each player subject to the common probability distribution pp. After the types of all players are determined, players will take moves in turns to proceed through the game tree until a leaf node is reached and the utilities of all players are thus determined.

Remark 1.

We assume that at each stage, the moves of all players are taken simultaneously, which means that each player cannot distinguish between the moves taken by others before it takes its own move. This setting is clearly revealed by the structure of the information sets in the game tree.

Remark 2.

Fig. 1 is an illustrative figure corresponding to a game with 2 players. A game with 3 or more players can be defined similarly. Meanwhile, the game tree description of a game is general enough such that other forms of games, such as a Stackelberg game, can also be well-represented by simply changing the configurations of the information sets. It will be shown that the proposed solving scheme does not rely on the particular structure of the game tree and therefore is also general.

Before we formally propose the solving scheme, we give a brief introduction to two relevant algorithms as preliminaries.

CFR [45] and MCCFR [46]: The vanilla CFR only considers a two-player zero-sum game without types of players. Therefore, σi(ti,I)\sigma_{i}(t_{i},I), ui(ti;a)u_{i}(t_{i};a), and Ai(ti,I)A_{i}(t_{i},I) can be rewritten as σi(I)\sigma_{i}(I), ui(a)u_{i}(a), and Ai(I)A_{i}(I), respectively. Given the overall strategy as σ\sigma, the counterfactual value vi(σ,I)v_{i}(\sigma,I) for information set II is defined as

vi(σ,I)=zZIπiσ(z[I])πσ(z[I],z)ui(z)v_{i}(\sigma,I)=\sum_{z\in Z_{I}}\pi^{\sigma}_{-i}(z[I])\pi^{\sigma}(z[I],z)u_{i}(z) (4)

where ZIZ_{I} is the set of all the paths on the game tree from root to leaf that pass through I, and zz is one of the paths in ZIZ_{I} containing all historical actions of all players. z[I]z[I] is the prefix of zz from root to II. πiσ(z[I])\pi^{\sigma}_{-i}(z[I]) is the cumulative product of the probabilities of moves taken by other agents along z[I]z[I], and πσ(z[I],z)\pi^{\sigma}(z[I],z) is the cumulative product of the probabilities of moves taken by all agents from z[I]z[I] to zz. In general, the counterfactual value vi(σ,I)v_{i}(\sigma,I) describes the expected utility of Information set II given strategy profile σ\sigma with the extra assumption that agent ii plays to reach II.

The immediate counterfactual regrets of each action aAi(I)a\in A_{i}(I) is r(σ,I,a)=vi(σ(Ia),I)vi(σ,I)r(\sigma,I,a)=v_{i}(\sigma_{(I\rightarrow a)},I)-v_{i}(\sigma,I), with σ(Ia)\sigma_{(I\rightarrow a)} define as the strategy profile identical to σ\sigma except that σ(aI)(a|I)=1\sigma_{(a\rightarrow I)}(a|I)=1, namely that action aa is selected with probability 11. In the entire no-regret learning process, rounds of playing are conducted. In each round, the regret-matching strategy is conducted to improve the strategy. Particularly in round k+1k+1, the strategy σk+1\sigma^{k+1} is determined by σik+1(a|I)=Rik,+(I,a)/aAi(I)Rik,+(I,a)\sigma^{k+1}_{i}(a|I)=R^{k,+}_{i}(I,a)/\sum_{a^{\prime}\in A_{i}(I)}R^{k,+}_{i}(I,a^{\prime}), where Rik,+(I,a)=max{τ=1kr(στ,I,a),0}R^{k,+}_{i}(I,a)=\max\{\sum_{\tau=1}^{k}r(\sigma^{\tau},I,a),0\}. It is proven that when kk\rightarrow\infty, the average strategy σ¯k\overline{\sigma}^{k} defined as σ¯ik(a|I)=τ=1kπiστ(I)στ(a|I)/τ=1kπiστ(I)\overline{\sigma}^{k}_{i}(a|I)=\sum_{\tau=1}^{k}\pi^{\sigma^{\tau}}_{i}(I)\sigma^{\tau}(a|I)/\sum_{\tau=1}^{k}\pi^{\sigma^{\tau}}_{i}(I) converges to a Nash equilibrium for a two-player zero-sum game.

A drawback of the vanilla CFR method is that in each round of playing, traversing the entire game tree is required, which is time-consuming and even intractable when the size of the game is large. To cope with this problem, the method MCCFR is proposed, such that in each round of playing, only part of the game tree is explored. Particularly in outcome-sampling MCCFR, only one leaf node of the game tree is sampled according to some pre-defined distribution, zq(z)z\sim q(z), and all other leaves are dropped (equivalently, their utilities are zeroed out). The following sampled counterfactual value is introduced:

v~i(σ,I;z)={1q(z)πiσ(z[I])πσ(z[I],z)ui(z)ifzZI0else.\tilde{v}_{i}(\sigma,I;z)=\left\{\begin{array}[]{lll}\frac{1}{q(z)}\pi^{\sigma}_{-i}(z[I])\pi^{\sigma}(z[I],z)u_{i}(z)&&\textup{if}\ z\in Z_{I}\\ 0&&\textup{else.}\end{array}\right. (5)

It is shown that 𝔼zp(z)v~i(σ,I;z)=vi(σ,I)\mathbb{E}_{z\sim p(z)}\tilde{v}_{i}(\sigma,I;z)=v_{i}(\sigma,I). Replacing vi(σ,I)v_{i}(\sigma,I) by v~i(σ,I;z)\tilde{v}_{i}(\sigma,I;z) yields the outcome-sampling MCCFR. It is obvious that in each round of playing, the only information sets that require updating are those passed through by the sampled path zz.

The previous restatement of CFR and MCCFR describes a solution to obtain the (mixed-strategy) Nash equilibrium in a two-player zero-sum extensive-form game. We integrate MCCFR, the no-regret learning method in game of incomplete information [49], and CFR-S [48] to introduce MCCFR-S for game of incomplete information, a parallelizable solving scheme that is proven to converge to a Bayes-CCE for a multi-player general-sum game of incomplete information settings. Details are shown in Algorithm 1. We give a brief explanation of each procedure used in this algorithm.

Bayes-CCE(G,M,p)(G,M,p): this is the main procedure of Algorithm 1. GG is the game described in Fig.1. MM is the number of iterations and pp is the common distribution over intentions of vehicles. rere, σ\sigma, and ϕ\phi are the array of cumulative counterfactual regrets, the array of behavioral strategies, and the frequency of plan ss, respectively. vava is the array of the cumulative sampled counterfactual values. All those arrays are zero-initialized. In each iteraction, a type vector tt representing vehicles’ intentions are sampled from distribution pp, and then the outcome-sampling MCCFR is performed once to update the regret array rere and the strategy array σ\sigma. A sampling process is then performed based on updated σ\sigma strategy such that a plan ss is sampled from σ\sigma and the frequency ϕ\phi is updated.

MCCFR(G,re,σ,t,va,z,pr,l)(G,re,\sigma,t,va,z,pr,l): this is a recursive implementation of MCCFR [46], where G,re,σ,tG,re,\sigma,t are defined as above. zz is the history of play or the past actions of all players concatenated together. For a particular zz, prpr is the vector containing probabilities of all players play to reach zz according to current strategy σ\sigma, namely {πiσ(z)}i𝒩\{\pi^{\sigma}_{i}(z)\}_{i\in\mathcal{N}}. ll is the probability of sampling zz, namely q(z)q(z). The algorithm contains a forward pass and a backward pass. In the forward pass we randomly proceed through the game tree from the root to one of the leaf by sampling action from the ϵ\epsilon-modulated strategy at each node (line 17 and 18) and calculate the prpr and ll for current zz (line 19 and 20). Once a leaf is reached, we obtain the utility for all players (line 11 to 14). In the backward pass, we revisit all the information sets we visited in the forward pass in reverse order. For an information set II, we obtain the probability playing from II to the sampled leaf according to current strategy (line 23), and updates the cumulative counterfactual regret for II (line 24 to 31). A standard regret matching is performed to update the behavioral strategy associated with II (line 32).

RM(re,i,ti,I)(re,i,t_{i},I): this is the standard regret matching algorithm for updating behavioral strategy associated with information set II called by the MCCFR procedure.

Sample(σ,ϕ)(\sigma,\phi): this is the sampling procedure. given updated strategy σ\sigma, a plan sis_{i} is sampled for each i𝒩i\in\mathcal{N}. The concatenated plan sΣs\in\Sigma is then obtained and record. ϕ\phi measures the time when ss is sampled, such that ϕ/M\phi/M is an empirical distribution over the space of plan Σ\Sigma.

Algorithm 1 MCCFR-S for Game of Incomplete Information
1:procedure Bayes-CCE(G,M,pG,M,p)
2:     Init re,σ,ϕ,vare,\sigma,\phi,va
3:     for 11 to MM do
4:         tpt\sim p
5:         MCCFR(re,σ,t,va,{},[1]i𝒩,1)\textup{MCCFR}(re,\sigma,t,va,\{\},[1]_{i\in\mathcal{N}},1)
6:         Sample(σ,ϕ)\textup{Sample}(\sigma,\phi)
7:     end for
8:     return ϕ,va\phi,va
9:end procedure
10:
11:procedure MCCFR(G,re,σ,t,va,z,pr,lG,re,\sigma,t,va,z,pr,l)
12:     if z𝒵z\in\mathcal{Z} then
13:         {ui}i𝒩getUtility(G,z)\{u_{i}\}_{i\in\mathcal{N}}\leftarrow\textup{getUtility}(G,z)
14:         return (1,l,{ui}i𝒩)(1,l,\{u_{i}\}_{i\in\mathcal{N}})
15:     end if
16:     igetPlayer(G,z)i\leftarrow\textup{getPlayer}(G,z)
17:     IgetInfoset(G,z,i)I\leftarrow\textup{getInfoset}(G,z,i)
18:     a(1ϵ)σi(ti,I)+ϵ/|Ai(ti,I)|a\sim(1-\epsilon)\sigma_{i}(t_{i},I)+\epsilon/|A_{i}(t_{i},I)|
19:     zz+az\leftarrow z+a
20:     ll((1ϵ)σi(a|ti,I)+ϵ/|Ai(ti,I)|)l\leftarrow l*((1-\epsilon)\sigma_{i}(a|t_{i},I)+\epsilon/|A_{i}(t_{i},I)|)
21:     pr[i]pr[i]σi(a|ti,I)pr[i]\leftarrow pr[i]*\sigma_{i}(a|t_{i},I)
22:     x,l,{ui}i𝒩MCCFR(G,re,σ,t,va,z,pr,l)x,l,\{u_{i}\}_{i\in\mathcal{N}}\leftarrow\textup{MCCFR}(G,re,\sigma,t,va,z,pr,l)
23:     cxc\leftarrow x
24:     xxσi(a|ti,I)x\leftarrow x*\sigma_{i}(a|t_{i},I)
25:     wui/ljipr[j]w\leftarrow u_{i}/l*\prod_{j\neq i}pr[j]
26:     for aAi(ti,I)a^{\prime}\in A_{i}(t_{i},I) do
27:         if a=aa^{\prime}=a then
28:              re[i,ti,I,a]re[i,ti,I,a]+(cx)wre[i,t_{i},I,a^{\prime}]\leftarrow re[i,t_{i},I,a^{\prime}]+(c-x)*w
29:         else
30:              re[i,ti,I,a]re[i,ti,I,a]+(x)wre[i,t_{i},I,a^{\prime}]\leftarrow re[i,t_{i},I,a^{\prime}]+(-x)*w
31:         end if
32:     end for
33:     σ[i,ti,I]RM(re,i,ti,I)\sigma[i,t_{i},I]\leftarrow\textup{RM}(re,i,t_{i},I)
34:     va[i,ti,I]va[i,ti,I]+xwva[i,t_{i},I]\leftarrow va[i,t_{i},I]+x*w
35:     return (x,l,u)(x,l,u)
36:end procedure
37:
38:procedure RM(re,i,ti,Ire,i,t_{i},I)
39:     total0total\leftarrow 0
40:     for aAi(ti,I)a\in A_{i}(t_{i},I) do
41:         totaltotal+max(re[i,ti,I,a],0)total\leftarrow total+\max(re[i,t_{i},I,a],0)
42:     end for
43:     for aAi(ti,I)a\in A_{i}(t_{i},I) do
44:         σi,ti,I[a]max(re[i,ti,I,a],0)/total\sigma_{i,t_{i},I}[a]\leftarrow\max(re[i,t_{i},I,a],0)/total
45:     end for
46:     return σi,ti,I\sigma_{i,t_{i},I}
47:end procedure
48:
49:procedure Sample(σ,ϕ\sigma,\phi)
50:     Init {si}i𝒩\{s_{i}\}_{i\in\mathcal{N}}
51:     for i𝒩,tiTi,Ii(ti)i\in\mathcal{N},t_{i}\in T_{i},I\in\mathcal{I}_{i}(t_{i}) do
52:         aσ[i,ti,I]a\sim\sigma[i,t_{i},I]
53:         si(ti,I)as_{i}(t_{i},I)\leftarrow a
54:     end for
55:     s{si}i𝒩s\leftarrow\{s_{i}\}_{i\in\mathcal{N}}
56:     ϕ[s]ϕ[s]+1\phi[s]\leftarrow\phi[s]+1
57:end procedure

Algorithm 1 is easily parallelizable and implemented with multiple processes, as each process samples the type vector tt, proceeds along the game tree, and samples the overall plan ss, independently. The only coupled part is the line 32, as each process needs to gather the rere computed by all other processes to perform the regret matching and update the strategy. For Algorithm 1, we have the following theorem.

Theorem 1.

The empirical distribution ϕ/M\phi/M obtained with Algorithm 1 converges almost surely to a Bayes-CCE with common prior pp when MM\rightarrow\infty.

Proof.

For a particular type of vehicle ii, tiTit_{i}\in T_{i}, let ti\mathcal{M}_{t_{i}} denote the set of iteration number such that tiτ=tit^{\tau}_{i}=t_{i} when τti\tau\in\mathcal{M}_{t_{i}}. Define riτ(σi,στ)=ui(tiτ;σi(tiτ),σiτ(tiτ))ui(tiτ;στ(tτ))r^{\tau}_{i}(\sigma_{i},\sigma^{\tau})=u_{i}(t_{i}^{\tau};\sigma_{i}(t_{i}^{\tau}),\sigma^{\tau}_{-i}(t^{\tau}_{-i}))-u_{i}(t_{i}^{\tau};\sigma^{\tau}(t^{\tau})), which is the regret value that the player ii experience at iteration τ\tau when its strategy is replaced by σi\sigma_{i} from the overall strategy στ\sigma^{\tau}. Furthermore, the cumulative regret value for player ii with type tit_{i}, supposed that player ii adopts strategy σi\sigma_{i} all the time, is defined as

Rti(σi)=1|ti|τtiriτ(σi,στ)).R_{t_{i}}(\sigma_{i})=\frac{1}{|\mathcal{M}_{t_{i}}|}\sum_{\tau\in\mathcal{M}_{t_{i}}}r^{\tau}_{i}(\sigma_{i},\sigma^{\tau})). (6)

Since each player of each type drawn from line 4 of Algorithm 1 maintains a separate MCCFR regret minimizer, it follows directly from [46, Theorem 5] that for all possible σi\sigma_{i}, we have Rti(σi)(1+2p)(1δ)Gti/|ti|R_{t_{i}}(\sigma_{i})\leq(1+\frac{\sqrt{2}}{\sqrt{p}})(\frac{1}{\delta})G_{t_{i}}/\sqrt{|\mathcal{M}_{t_{i}}|} with probability 1p1-p for any p(0,1]p\in(0,1], where GtiG_{t_{i}} is a game constant and δ\delta is the smallest probability of a leaf being sampled in the forward pass in the MCCFR process. It is obvious that when MM\rightarrow\infty, it is also surely that Rti(σi)0R_{t_{i}}(\sigma_{i})\leq 0 if δ>0\delta>0. In particular, a plan sis_{i} is also a strategy, and therefore this conclusion naturally holds for all Rti(si)R_{t_{i}}(s_{i}), namely

Rti(si)(1+2p)(1δ)Gti/|ti|R_{t_{i}}(s_{i})\leq\left(1+\frac{\sqrt{2}}{\sqrt{p}}\right)\left(\frac{1}{\delta}\right)G_{t_{i}}/\sqrt{|\mathcal{M}_{t_{i}}|} (7)

holds for all sis_{i} with probability 1p1-p.

Next, at each iteration τ\tau, we sample a plan sτs^{\tau} from the strategy στ\sigma^{\tau}, and we define the sampled cumulative regret as

R^ti(si)=1|ti|τtiriτ(si,sτ).\hat{R}_{t_{i}}(s_{i})=\frac{1}{|\mathcal{M}_{t_{i}}|}\sum_{\tau\in\mathcal{M}_{t_{i}}}r^{\tau}_{i}(s_{i},s^{\tau}). (8)

Since sτστs^{\tau}\sim\sigma^{\tau}, the following equations hold naturally:

𝔼sτστui(tiτ;sτ(tτ))=ui(tiτ;στ(tτ)),\displaystyle\mathbb{E}_{s^{\tau}\sim\sigma^{\tau}}u_{i}(t_{i}^{\tau};s^{\tau}(t^{\tau}))=u_{i}(t^{\tau}_{i};\sigma^{\tau}(t^{\tau})), (9)
𝔼sτστui(tiτ;si(tiτ),si(tiτ))=ui(tiτ;si(tiτ),σiτ(tiτ)).\displaystyle\mathbb{E}_{s^{\tau}\sim\sigma^{\tau}}u_{i}(t_{i}^{\tau};s_{i}(t^{\tau}_{i}),s_{-i}(t^{\tau}_{-i}))=u_{i}(t^{\tau}_{i};s_{i}(t^{\tau}_{i}),\sigma_{-i}^{\tau}(t^{\tau}_{-i})).

Therefore, the following conclusion is directly obtained:

𝔼{sτστ}τti[R^ti(si)Rti(si)]=0.\mathbb{E}_{\{s^{\tau}\sim\sigma^{\tau}\}_{\tau\in\mathcal{M}_{t_{i}}}}[\hat{R}_{t_{i}}(s_{i})-R_{t_{i}}(s_{i})]=0. (10)

For simplicity we rewrite 𝔼{sτστ}τti\mathbb{E}_{\{s^{\tau}\sim\sigma^{\tau}\}_{\tau\in\mathcal{M}_{t_{i}}}} as 𝔼ti\mathbb{E}_{\mathcal{M}_{t_{i}}}, 𝔼sτστ\mathbb{E}_{s^{\tau}\sim\sigma^{\tau}} as 𝔼τ\mathbb{E}_{\tau}, riτ(si,sτ)riτ(si,στ))r^{\tau}_{i}(s_{i},s^{\tau})-r^{\tau}_{i}(s_{i},\sigma^{\tau})) as ζτ\zeta_{\tau}, and we inspect the following expectation:

𝔼ti[(R^ti(si)Rti(si))2]\displaystyle\mathbb{E}_{\mathcal{M}_{t_{i}}}\left[(\hat{R}_{t_{i}}(s_{i})-R_{t_{i}}(s_{i}))^{2}\right] (11)
=1|ti|2𝔼ti[(τtiζτ)2]\displaystyle=\frac{1}{|\mathcal{M}_{t_{i}}|^{2}}\mathbb{E}_{\mathcal{M}_{t_{i}}}\left[\left(\sum_{\tau\in\mathcal{M}_{t_{i}}}\zeta_{\tau}\right)^{2}\right]
=1|ti|2(τti𝔼τ[ζτ2]+2τtiτti,ττ𝔼τ[ζτζτ])\displaystyle=\frac{1}{|\mathcal{M}_{t_{i}}|^{2}}\left(\sum_{\tau\in\mathcal{M}_{t_{i}}}\mathbb{E}_{\tau}[\zeta_{\tau}^{2}]+2\sum_{\tau\in\mathcal{M}_{t_{i}}}\sum_{\tau^{\prime}\in\mathcal{M}_{t_{i}},\tau^{\prime}\neq\tau}\mathbb{E}_{\tau}[\zeta_{\tau}\zeta_{\tau^{\prime}}]\right)
=1|ti|2(τti𝔼τ[ζτ2])Δuti2|ti|.\displaystyle=\frac{1}{|\mathcal{M}_{t_{i}}|^{2}}\left(\sum_{\tau\in\mathcal{M}_{t_{i}}}\mathbb{E}_{\tau}[\zeta_{\tau}^{2}]\right)\leq\frac{\Delta u_{t_{i}}^{2}}{|\mathcal{M}_{t_{i}}|}.

The last two steps follow from the facts that ζτ\zeta_{\tau} and ζτ\zeta_{\tau^{\prime}} are independent with zero expectations, and that 𝔼τ[ζτ2]Δuti2\mathbb{E}_{\tau}[\zeta_{\tau}^{2}]\leq\Delta u_{t_{i}}^{2} where Δuti2\Delta u_{t_{i}}^{2} is the range of utility of player ii with type tit_{i}.

Combining (10) and (LABEL:Var), and following [46, Lemma 1], we obtain that

R^ti(si)Rti(si)+Δutip|ti|\hat{R}_{t_{i}}(s_{i})\leq R_{t_{i}}(s_{i})+\frac{\Delta u_{t_{i}}}{\sqrt{p^{\prime}}\sqrt{|\mathcal{M}_{t_{i}}|}} (12)

with probability 1p1-p^{\prime}. Combining (7) and (12), we have

R^ti(si)1|ti|((1+2p)(Gtiδ)+Δutip)\hat{R}_{t_{i}}(s_{i})\leq\frac{1}{\sqrt{|\mathcal{M}_{t_{i}}|}}\left(\left(1+\frac{\sqrt{2}}{\sqrt{p}}\right)\left(\frac{G_{t_{i}}}{\delta}\right)+\frac{\Delta u_{t_{i}}}{\sqrt{p^{\prime}}}\right) (13)

with probability (1p)(1p)(1-p)(1-p^{\prime}). As a result, it is almost sure that for any sis_{i}, we have τtiriτ(si,sτ)o(|ti|)\sum_{\tau\in\mathcal{M}_{t_{i}}}r^{\tau}_{i}(s_{i},s^{\tau})\leq o(|\mathcal{M}_{t_{i}}|) for all ii, all tit_{i}, and all sis_{i}. Following [49, Lemma 10], we conclude that the empirical distribution ϕ/M\phi/M converges almost surely to a Bayes-CCE. ∎

Sampling an entire plan sis_{i} requires traversing the entire game tree and sampling action for each information set IiI\in\mathcal{I}_{i}, which is computationally heavy for extensive-form games with multiple stages, as the number of information sets grows exponentially with respect to the number of stages. In the proposed framework, the receding horizon control is adopted, as each agent only performs the action computed in the first stage (details will be discussed in Section IV-D). Therefore, we only need to record the actions taken by players at the first stage, resulting in a partial plan. The resulting distribution can be viewed as a marginal distribution of the origin Bayes-CCE.

IV-B Integrated Decision Selection and Trajectory Planning

In a typical game of incomplete information, the type of each player is usually drawn by nature, a chance player, from the prior distribution. In the settings of autonomous driving, however, the types are self-determined by players, and therefore we can select the best type to perform. In the proposed framework, we view the selection of the best type tit^{*}_{i} from the type list TiT_{i} as equivalent to the decision-making process. Formally given the computed Bayes-CCE ψ\psi, we denote the expected utility of a type tit^{\prime}_{i} as

V(ti)=𝔼sψ𝔼tipiui(ti;s(ti,ti)).V(t^{\prime}_{i})=\mathbb{E}_{s\sim\psi}\mathbb{E}_{t_{-i}\sim p_{-i}}u_{i}(t^{\prime}_{i};s(t^{\prime}_{i},t_{-i})). (14)

Then the best type tit^{*}_{i} and equivalently the best decision, is determined by ti=argmaxtiTiV(ti).t^{*}_{i}=\operatorname*{argmax}_{t^{\prime}_{i}\in T_{i}}V(t^{\prime}_{i}). However, as we mentioned in Section IV-A, we only keep a partial record of the Bayes-CCE ψ\psi for computation efficiency, and therefore it is not feasible to directly calculate V(ti)V(t^{\prime}_{i}). Instead, we introduce an estimator of this expectation and use this estimator as a criterion of decision-making. We denote the root information set corresponding to tit_{i} as I0,tiI_{0,t_{i}}, and the estimator is given as V^(ti)=va[i,ti,I0,ti]/|ti|\hat{V}(t^{\prime}_{i})=va[i,t^{\prime}_{i},I_{0,t^{\prime}_{i}}]/|\mathcal{M}_{t^{\prime}_{i}}|. Immediately, we have the following theorem.

Theorem 2.

Suppose that the empirical distribution ϕ/M\phi/M converges to a Bayes-CCE ψ\psi when MM\rightarrow\infty, then V^(ti)V(ti)\hat{V}(t^{\prime}_{i})\rightarrow V(t^{\prime}_{i}) when MM\rightarrow\infty.

Proof.

For V(ti)V(t^{\prime}_{i}) we have

V(ti)\displaystyle V(t^{\prime}_{i}) =𝔼sψ𝔼tipiui(ti;s(ti,ti))\displaystyle=\mathbb{E}_{s\sim\psi}\mathbb{E}_{t_{-i}\sim p_{-i}}u_{i}(t^{\prime}_{i};s(t^{\prime}_{i},t_{-i})) (15)
=1pi(ti)𝔼sψpi(ti)𝔼tipiui(ti;s(ti,ti))\displaystyle=\frac{1}{p_{i}(t^{\prime}_{i})}\mathbb{E}_{s\sim\psi}p_{i}(t^{\prime}_{i})\mathbb{E}_{t_{-i}\sim p_{-i}}u_{i}(t^{\prime}_{i};s(t^{\prime}_{i},t_{-i}))
=1pi(ti)𝔼sψ𝔼tipi1{ti=ti}𝔼tipiui(ti;s(ti,ti))\displaystyle=\frac{1}{p_{i}(t^{\prime}_{i})}\mathbb{E}_{s\sim\psi}\mathbb{E}_{t_{i}\sim p_{i}}\textbf{1}\{t_{i}=t^{\prime}_{i}\}\mathbb{E}_{t_{-i}\sim p_{-i}}u_{i}(t_{i};s(t_{i},t_{-i}))
=1pi(ti)𝔼sψ𝔼tp1{ti=ti}ui(ti;s(t))\displaystyle=\frac{1}{p_{i}(t^{\prime}_{i})}\mathbb{E}_{s\sim\psi}\mathbb{E}_{t\sim p}\textbf{1}\{t_{i}=t^{\prime}_{i}\}u_{i}(t_{i};s(t))

where 1{}\textbf{1}\{\cdot\} is a binary function such that it equals to 11 if the condition holds and 0 otherwise. pi(ti)p_{i}(t^{\prime}_{i}) is the probability of tit^{\prime}_{i}. For V^(ti)\hat{V}(t^{\prime}_{i}), we examine its expectation

𝔼[V^(ti)]\displaystyle\mathbb{E}[\hat{V}(t^{\prime}_{i})] =1|ti|τti𝔼zq(z)1q(z)πστ(tτ)(z)ui(ti;z)\displaystyle=\frac{1}{|\mathcal{M}_{t^{\prime}_{i}}|}\sum_{\tau\in\mathcal{M}_{t^{\prime}_{i}}}\mathbb{E}_{z\sim q(z)}\frac{1}{q(z)}\pi^{\sigma^{\tau}(t^{\tau})}(z)u_{i}(t^{\prime}_{i};z) (16)
=1|ti|τtiui(ti;στ(tτ))\displaystyle=\frac{1}{|\mathcal{M}_{t^{\prime}_{i}}|}\sum_{\tau\in\mathcal{M}_{t^{\prime}_{i}}}u_{i}(t^{\prime}_{i};\sigma^{\tau}(t^{\tau}))

which follows from  [46, Lemma 1]. Plug in (9) and we have

𝔼[V^(ti)]=M|ti|1MτM𝔼sτστ1{tiτ=ti}ui(ti;sτ(tτ))\displaystyle\mathbb{E}[\hat{V}(t^{\prime}_{i})]=\frac{M}{|\mathcal{M}_{t^{\prime}_{i}}|}\frac{1}{M}\sum_{\tau\in M}\mathbb{E}_{s^{\tau}\sim\sigma^{\tau}}\textbf{1}\{t^{\tau}_{i}=t^{\prime}_{i}\}u_{i}(t_{i};s^{\tau}(t^{\tau})) (17)
=M|ti|1MτMlimQ1QkQ1{tiτ=ti}ui(ti;sτ,k(tτ))\displaystyle=\frac{M}{|\mathcal{M}_{t^{\prime}_{i}}|}\frac{1}{M}\sum_{\tau\in M}\lim_{Q\rightarrow\infty}\frac{1}{Q}\sum_{k\in Q}\textbf{1}\{t^{\tau}_{i}=t^{\prime}_{i}\}u_{i}(t_{i};s^{\tau,k}(t^{\tau}))
=M|ti|limQ1QkQ1MτM1{tiτ=ti}ui(ti;sτ,k(tτ))\displaystyle=\frac{M}{|\mathcal{M}_{t^{\prime}_{i}}|}\lim_{Q\rightarrow\infty}\frac{1}{Q}\sum_{k\in Q}\frac{1}{M}\sum_{\tau\in M}\textbf{1}\{t^{\tau}_{i}=t^{\prime}_{i}\}u_{i}(t_{i};s^{\tau,k}(t^{\tau}))

where sτ,ks^{\tau,k} is sampled from στ\sigma^{\tau}. The above equation follows from the law of large numbers. Obviously, when MM\rightarrow\infty, we have

1MτM1{tiτ=ti}ui(ti;sτ,k(tτ))\displaystyle\frac{1}{M}\sum_{\tau\in M}\textbf{1}\{t^{\tau}_{i}=t^{\prime}_{i}\}u_{i}(t_{i};s^{\tau,k}(t^{\tau}))\rightarrow (18)
𝔼sψ𝔼tp1{ti=ti}ui(ti;s(t))\displaystyle\mathbb{E}_{s\sim\psi}\mathbb{E}_{t\sim p}\textbf{1}\{t_{i}=t^{\prime}_{i}\}u_{i}(t_{i};s(t))

as {sτ,k}τM\{s^{\tau,k}\}_{\tau\in M} can be viewed as samples from the Bayes-CCE ψ\psi, and {tτ}τM\{t^{\tau}\}_{\tau\in M} are samples from pp. Again, the law of large numbers is applied. Therefore,

limM𝔼[V^(ti)]\displaystyle\lim_{M\rightarrow\infty}\mathbb{E}[\hat{V}(t^{\prime}_{i})] (19)
=limMM|ti|limQ1QkQ𝔼sψ𝔼tp1{ti=ti}ui(ti;s(t))\displaystyle=\lim_{M\rightarrow\infty}\frac{M}{|\mathcal{M}_{t^{\prime}_{i}}|}\lim_{Q\rightarrow\infty}\frac{1}{Q}\sum_{k\in Q}\mathbb{E}_{s\sim\psi}\mathbb{E}_{t\sim p}\textbf{1}\{t_{i}=t^{\prime}_{i}\}u_{i}(t_{i};s(t))
=limMM|ti|𝔼sψ𝔼tp1{ti=ti}ui(ti;s(t))\displaystyle=\lim_{M\rightarrow\infty}\frac{M}{|\mathcal{M}_{t^{\prime}_{i}}|}\mathbb{E}_{s\sim\psi}\mathbb{E}_{t\sim p}\textbf{1}\{t_{i}=t^{\prime}_{i}\}u_{i}(t_{i};s(t))
=1pi(ti)𝔼sψ𝔼tp1{ti=ti}ui(ti;s(t))=V(ti).\displaystyle=\frac{1}{p_{i}(t^{\prime}_{i})}\mathbb{E}_{s\sim\psi}\mathbb{E}_{t\sim p}\textbf{1}\{t_{i}=t^{\prime}_{i}\}u_{i}(t_{i};s(t))=V(t^{\prime}_{i}).

Noted that each term V^τ(ti)=πστ(tτ)/q(z)ui(ti;z)\hat{V}^{\tau}(t^{\prime}_{i})=\pi^{\sigma^{\tau}(t^{\tau})}/q(z)u_{i}(t^{\prime}_{i};z), is independent and with bounded covariance, following the Kolmogorov’s strong law of large numbers  [51, Theorem 2.3.10], we conclude that V^(ti)V(ti)\hat{V}(t^{\prime}_{i})\rightarrow V(t^{\prime}_{i}) when MM\rightarrow\infty. ∎

Theorem 2 indicates that V^(ti)\hat{V}(t^{\prime}_{i}) is an unbiased and consistent estimator of V(ti)V(t^{\prime}_{i}). Moreover, it can be seen from Algorithm 1 that calculating V^(ti)\hat{V}(t^{\prime}_{i}), which corresponds to line 33 of the algorithm, adds only minimal computation load to the algorithm as no extra sampling is required. Therefore, we propose the decision-making strategy as

ti=argmaxtiTiV^(ti)t^{*}_{i}=\operatorname*{argmax}_{t^{\prime}_{i}\in T_{i}}\hat{V}(t^{\prime}_{i}) (20)

such that by selecting its type as tit^{*}_{i}, the player ii makes the decision that maximizes its overall expected utility corresponding to the given Bayes-CCE ψ\psi.

Remark 3.

The decision-making strategy presented does not distinguish between ego vehicle and surrounding vehicles. However, we should not assume that surrounding vehicles will adopt the same strategy since their decisions are subject to many factors (navigation, driving preference, bounded rationality, etc.) and are usually sub-optimal. Potential danger may occur if we assume that surrounding vehicles act optimally. Therefore, (20) is only used to make decisions for the ego vehicle but not to predict the driving purposes of surrounding vehicles.

Once the type tit^{*}_{i} is determined, we can select the optimal plan sis^{*}_{i} from the Bayes-CCE ψ\psi, such that the trajectory is given by si(ti)s^{*}_{i}(t^{*}_{i}). We introduce two different schemes to perform the Bayes-CCE ψ\psi and select the optimal plan.

Accurate implementation of Bayes-CCE ψ\psi: since a Bayes-CCE is correlated, the accurate performance of a Bayes-CCE will require either negotiation beforehand (which is unlikely in urban traffic scenarios) or a leader-follower assumption which is often adopted in the Stackelberg game formulation of traffic interactions  [8, 13]. Without loss of generality, we assume that player 1 is the leader, and player ii is a direct follower of player i1i-1 for all i𝒩i\in\mathcal{N} and i1i\neq 1. The joint probability distribution ψ\psi can be decomposed into a series of marginal distributions and conditional distributions, such that the probability of a joint plan s={si}i𝒩s=\{s_{i}\}_{i\in\mathcal{N}} is given by

ψ(s)=ψ1(s1)i𝒩,i1ψi(si|{sj}j<i)\psi(s)=\psi_{1}(s_{1})\prod_{i\in\mathcal{N},i\neq 1}\psi^{\prime}_{i}(s_{i}|\{s_{j}\}_{j<i}) (21)

such that ψ1\psi_{1} is the marginal distribution of the plan corresponding to player 11, and ψi\psi^{\prime}_{i} is the conditional distribution of the plans of player ii conditioned on the plan of all the players before ii. If plans are selected by maximum likelihood, the plans selected are

s1=argmaxs1ψ1(s1),si=argmaxsiψi(si|{sj}ji).s^{*}_{1}=\operatorname*{argmax}_{s_{1}}\psi_{1}(s_{1}),s^{*}_{i}=\operatorname*{argmax}_{s_{i}}\psi^{\prime}_{i}(s_{i}|\{s^{*}_{j}\}_{j\leq i}). (22)

We consider this scheme to be overly idealistic due to the following reasons:

  • Although being widely adopted, the leader-follower assumption is not always valid, especially when more than two vehicles are involved.

  • Due to partial observation, vehicle ii cannot know for sure the plan sjs_{j} adopted by vehicle jj. At best, it can only maintain a belief over sjs_{j}.

Due to these reasons, we introduce the following marginal implementation of the Bayes-CCE.

Marginal implementation of Bayes-CCE ψ\psi: to avoid the problem mentioned before, we propose an approximation performance scheme of the Bayes-CCE ψ\psi. We define ψi\psi_{i} as the marginal distribution of the plan of player ii, and then the plan is determined as

si=argmaxsiψi(si).s^{*}_{i}=\operatorname*{argmax}_{s_{i}}\psi_{i}(s_{i}). (23)

The marginal implementation of Bayes-CCE is used in the simulation. Once the optimal plan sis^{*}_{i} is determined by one of the schemes introduced before, the planning process is conducted by taking si(ti)s^{*}_{i}(t^{*}_{i}) with the optimal decision tit^{*}_{i}. As a result, an integrated decision-making and trajectory planning scheme is introduced based on the Bayes-CCE.

IV-C Bayesian Intention Update

Following a Bayes-CCE, the probabilities of actions taken by vehicles depend on their underlying driving intentions. Therefore, the actual actions performed by vehicles provide extra information gains over the belief of driving intentions. In this section, we introduce strategies to update the common belief over the driving intentions of all vehicles based on the observed actions. Corresponding to different implementations of the Bayes-CCE, different updating strategies are introduced.

Update strategy for the accurate implementation: suppose that at time stamp λ\lambda the common prior on intentions of vehicles is given as pλp^{\lambda}, the Bayes-CCE is given as ψλ\psi^{\lambda}, and the action performed by all vehicles is aλ={aiλ}i𝒩a^{\lambda}=\{a_{i}^{\lambda}\}_{i\in\mathcal{N}}. The following equation gives the updated common prior pλ+1p^{\lambda+1}

pλ+1(t)=f(aλ,t;ψλ)pλ(t)tTf(aλ,t;ψλ)pλ(t)p^{\lambda+1}(t)=\frac{f(a^{\lambda},t;\psi^{\lambda})p^{\lambda}(t)}{\sum_{t^{\prime}\in T}f(a^{\lambda},t^{\prime};\psi^{\lambda})p^{\lambda}(t^{\prime})} (24)

which follows the standard Bayes’ rule of posterior probability. ff is the observation model such that f(aλ,t;ψλ)f(a^{\lambda},t;\psi^{\lambda}) gives the probability that action aλa^{\lambda} is taken if all players follow the Bayes-CCE ψλ\psi^{\lambda} and their underlying types are defined by tt. To effectively model ff, we adopt the Mixture-of-Gaussians distributions, and ff is defined as

f(aλ,t;ψλ)=sη(μ(aλ);s(t))ψλ(s(t))f(a^{\lambda},t;\psi^{\lambda})=\sum_{s}\eta(\mu(a^{\lambda});s(t))\psi^{\lambda}(s(t)) (25)

where η(;s(t))\eta(\cdot;s(t)) is the probability density function of a Gaussian distribution N(μ(s(t)),Σ)N(\mu(s(t)),\Sigma). μ(aλ)\mu(a^{\lambda}) and μ(s(t))\mu(s(t)) are the end states of vehicles when action aλa^{\lambda} and s(t)s(t) are taken, respectively. Σ\Sigma is a predefined covariance matrix.

Update strategy for the marginal implementation: under this implementation, the update rule is similarly given as

piλ+1(ti)=fi(aiλ,ti;ψiλ)piλ(ti)tiTifi(aiλ,ti;ψiλ)piλ(ti)p_{i}^{\lambda+1}(t_{i})=\frac{f_{i}(a^{\lambda}_{i},t_{i};\psi^{\lambda}_{i})p_{i}^{\lambda}(t_{i})}{\sum_{t^{\prime}_{i}\in T_{i}}f_{i}(a^{\lambda}_{i},t^{\prime}_{i};\psi^{\lambda}_{i})p_{i}^{\lambda}(t^{\prime}_{i})} (26)

where

fi(aiλ,ti;ψiλ)=siη(μ(aiλ);si(ti))ψiλ(si(ti)).f_{i}(a^{\lambda}_{i},t_{i};\psi^{\lambda}_{i})=\sum_{s_{i}}\eta(\mu(a^{\lambda}_{i});s_{i}(t_{i}))\psi^{\lambda}_{i}(s_{i}(t_{i})). (27)

This update strategy is performed individually for all i𝒩i\in\mathcal{N}.

Remark 4.

In the updating process, the belief of intentions is updated for both the surrounding vehicles and also the ego vehicle. Following Assumption 1, the updated belief is still common among all the vehicles involved in the game.

IV-D System Overview and Details of Implementation

Combining all aforementioned modules, we introduce Algorithm 2, which is an integrated decision-making and trajectory planning framework for autonomous driving. Detailed explanations are given as follow:

Algorithm 2 Integrated Decision Selection and Trajectory Planning for Autonomous Vehicle
1:Init p,Mp,M
2:jgetOwnID()j\leftarrow\textup{getOwnID}()
3:while True do
4:     for iNi\in N do
5:         xigetPosition(i)x_{i}\leftarrow\textup{getPosition}(i)
6:         for tiTit_{i}\in T_{i} do
7:              Tree(i,ti)getTrajTree(xi,i,ti)Tree(i,t_{i})\leftarrow\textup{getTrajTree}(x_{i},i,t_{i})
8:         end for
9:     end for
10:     GgetGame(Tree,{xi}iN)G\leftarrow\textup{getGame}(Tree,\{x_{i}\}_{i\in N})
11:     ϕ,vaBayes-CCE(G,M,p)\phi,va\leftarrow\textup{Bayes-CCE}(G,M,p)
12:     ψϕ/M\psi\leftarrow\phi/M
13:     for tjTjt_{j}\in T_{j} do
14:         V^(tj)va[j,tj,I0,tj]/pj(tj)\hat{V}(t_{j})\leftarrow va[j,t_{j},I_{0,t_{j}}]/p_{j}(t_{j})
15:     end for
16:     tjargmaxtjTjV^(tj)t^{*}_{j}\leftarrow\operatorname*{argmax}_{t_{j}\in T_{j}}\hat{V}(t_{j})
17:     sjargmaxsjψj(sj)s^{*}_{j}\leftarrow\operatorname*{argmax}_{s_{j}}\psi_{j}(s_{j})
18:     ajsj(tj)a_{j}\leftarrow s^{*}_{j}(t^{*}_{j})
19:     for iNi\in N do
20:         aigetObservation(i)a_{i}\leftarrow\textup{getObservation(i)}
21:     end for
22:     pgetUpdate(p,ψ,{ai}iN)p\leftarrow\textup{getUpdate}(p,\psi,\{a_{i}\}_{i\in N})
23:end while
  1. 1.

    A two-stage trajectory tree is constructed for each driving intention of each vehicle, which constitutes the action space (line 4 to 9). The methodology used to construct those trajectory trees is the same as the one used in [27].

  2. 2.

    Given the sets of trajectory trees and the prior distribution, a game of incomplete information is established, and the solver introduced in Section IV-A is invoked to solve for the Bayes-CCE (line 10 to 12).

  3. 3.

    Given the Bayes-CCE, the optimal decision is selected following the strategy described in equation (20) (line 13 to 16).

  4. 4.

    One of the two implementation strategies of Bayes-CCE described in Section IV-B is applied to select the optimal plan. Together with the optimal decision selected in the previous step, the trajectory is determined and executed (line 17 to 18).

  5. 5.

    Given the Bayes-CCE and the new trajectories, the corresponding update strategy is invoked to update the driving intentions of all vehicles. Thus, a new common belief over the driving intentions is established, and the algorithm goes back to Step 1)1) for a new round of playing (line 19 to 22).

The prior belief over the driving intentions is initialized with uniform distribution. Note that data-driven prediction methods can also be applied to perform the initialization and obtain a better estimation of the driving intentions based on historical trajectories.

Considering important performance indices such as passenger comfort and safety, the utility is defined as

ui(ti;z)=Jci(z)Jsi(z)Jpi(z)Jrefi(z).u_{i}(t_{i};z)=-J^{i}_{c}(z)-J^{i}_{s}(z)-J^{i}_{p}(z)-J^{i}_{\textup{ref}}(z). (28)

The trajectories of all vehicles are determined by the action history zz. To compute for the above performance indices, we sample along the trajectories of all vehicles with equal time spacing h[0,1,,H]h\in[0,1,...,H]. The comfort indice JciJ^{i}_{c} is given by

Jci(z)=\displaystyle J^{i}_{c}(z)= h=0Hωa,lat|ai,lath(z)|2+ωj,lat|ji,lath(z)|2\displaystyle\sum_{h=0}^{H}\omega_{a,lat}|a^{h}_{i,lat}(z)|^{2}+\omega_{j,lat}|j^{h}_{i,lat}(z)|^{2} (29)
+ωa,long|ai,longh(z)|2+ωj,long|ji,longh(z)|2\displaystyle+\omega_{a,long}|a^{h}_{i,long}(z)|^{2}+\omega_{j,long}|j^{h}_{i,long}(z)|^{2}

where ai,latha^{h}_{i,lat}, ai,longha^{h}_{i,long}, ji,lathj^{h}_{i,lat}, and ji,longhj^{h}_{i,long} are the lateral acceleration, the longitudinal acceleration, the lateral jerk, and the longitudinal jerk of vehicle ii at hh, respectively. ωa,lat\omega_{a,lat}, ωa,long\omega_{a,long}, ωj,lat\omega_{j,lat}, and ωj,long\omega_{j,long} are the corresponding weighting coefficients. The progress indice JpiJ^{i}_{p} is given by

Jpi(z)=ωph=0Hmin{vi,longh(z)vslow,0}2J^{i}_{p}(z)=\omega_{p}\sum_{h=0}^{H}\min\{v^{h}_{i,long}(z)-v_{slow},0\}^{2} (30)

such that longitudinal velocity vi,longhv^{h}_{i,long} smaller than vslowv_{slow} will be penalized. ωp\omega_{p} is the weighting coefficient. Indice JrefiJ^{i}_{\textup{ref}} is given by

Jrefi(z)=ωrefh=0Hxi,lath(z)2J^{i}_{\textup{ref}}(z)=\omega_{\textup{ref}}\sum_{h=0}^{H}x^{h}_{i,lat}(z)^{2} (31)

which penalizes the deviations from the reference line by penalizing the lateral displacement xi,lathx^{h}_{i,lat}. ωref\omega_{\textup{ref}} is the corresponding weighting coefficient. For collision avoidance, we represent each vehicle with two identical circles aligned along the longitudinal axis of the vehicle. For vehicle ii, the positions of the centers of circles are denoted as xi,fx_{i,f} and xi,rx_{i,r}, respectively. The safety indice JsiJ^{i}_{s} is defined as

Jsi(z)=ωsh=0Hj,β,γmin{D(xi,βh(z)xj,γh(z))d,0}2J^{i}_{s}(z)=\omega_{s}\sum_{h=0}^{H}\sum_{j,\beta,\gamma}\min\{D(x^{h}_{i,\beta}(z)-x^{h}_{j,\gamma}(z))-d,0\}^{2} (32)

where D(,)D(\cdot,\cdot) defines the Euclidean distance, j𝒩{i}j\in\mathcal{N}-\{i\}, and β,γ{f,r}\beta,\gamma\in\{f,r\}. ωs\omega_{s} is the weighting coefficient. As a result, a negative utility (namely a penalty) is added towards the overall utility when vehicle ii fails to maintain a safe distance dd with respect to other vehicles. Values of parameters are shown in Table I.

TABLE I: Parameter settings used in simulations
Param. Value Param. Value Param. Value Param. Value
ωa,lat\omega_{a,lat} 0.5 ωa,long\omega_{a,long} 1.0 ωj,lat\omega_{j,lat} 0.5 ωj,long\omega_{j,long} 1.0
ωp\omega_{p} 20.0 ωref\omega_{\textup{ref}} 10.0 ωs\omega_{s} 2000.0 dd 4.0 m

V Simulation Results

To illustrate the effectiveness of the introduced method in handling multi-modal driving behaviors and to verify the generalizability, two case studies of different traffic scenarios and a quantitative analysis compared with a baseline are included in this section. In both case studies, each agent maintains its game solver and solves the game separately. The types of Human-driving vehicles are given and fixed while the autonomous vehicle determines its own type adopting the strategy described in (20). All algorithms are implemented in Python 3.8 running on a server with Intel(R) Xeon(R) Platinum 8358P CPU @ 2.60GHz.

V-A Case I: Ramp Merging

Refer to caption
Refer to caption
Refer to caption
Refer to caption
(a) Simulation results of Scenario A at 0.8 s, 1.6 s, 2.4 s, and 3.2 s, where HV1 yields and AV merges into the gap between the two HVs.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
(b) Simulation results of Scenario B at 0.8 s, 1.6 s, 2.4 s, and 3.2 s, where HV1 does not yield and AV merges behind both HVs.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
(c) Simulation results of Scenario C at 0.8 s, 1.6 s, 2.4 s, and 3.2 s. where HV2 yields and AV merges into the gap between the two HVs.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
(d) Simulation results of Scenario D at 0.8 s, 1.6 s, 2.4 s, and 3.2 s, where HV2 does not yield and AV merges behind both HVs.
Figure 2: Simulation results for Case I. In all scenarios, AV manages to merge into the main road without collision. Depending on the differentiated behaviors of HVs, AV either merges into the gap between two HVs or merges behind both HVs.

In this case, the goal of the autonomous driving vehicle (AV) is to merge into a two-lane road from the ramp. One human-driving vehicle (HV1) is currently driving in the target lane of AV, while another human-driving vehicle (HV2) is running in another lane of the main road but also tries to merge into the target lane of AV. The initial velocity of all three vehicles is 7m/s7\,\textup{m}/\textup{s}. For the type settings, we consider each vehicle to be either a conservative-type vehicle or an aggressive-type vehicle. For each type, we sample the terminal longitudinal velocities that constitute the action space. In particular, for the aggressive type, the action space is determined to be [7,8,10,12]m/s[7,8,10,12]\,\textup{m}/\textup{s}, while for the conservative type, the action space is set to be [6,4,2,0]m/s[6,4,2,0]\,\textup{m}/\textup{s}. Furthermore, in this case, four different scenarios are considered. In Scenario A, the longitudinal initial positions of AV, HV1, and HV2 are 10m10\,\textup{m}, 8m8\,\textup{m}, and 12m12\,\textup{m}, respectively. HV1 and HV2 are conservative and aggressive, respectively. In Scenario B, the other settings are the same as in Scenario A but HV1 is also aggressive. In Scenario C, the longitudinal initial positions of AV, HV1, and HV2 are 10m10\,\textup{m}, 12m12\,\textup{m}, and 8m8\,\textup{m}, respectively. HV1 and HV2 are aggressive and conservative, respectively. In Scenario D, the other settings are the same as in Scenario C but the HV2 is also aggressive. The time span for both stages in the trajectory tree is 1s1\,\textup{s}.

The simulation results are shown in Fig. 2. It can be seen from the results that when the rear HV decides to yield, the AV manages to identify the type of the HV and successfully merge into the gap between the two HVs, such that the driving efficiency is enhanced. On the contrary, when the rear HV decides not to yield, the AV also identifies and performs the merging after both HVs to avoid collisions. Fig. 3 shows the longitudinal velocities of AV for each scenario. It can be seen that for all 4 scenarios, the AV decelerates at the beginning because at the moment it is uncertain about the driving intentions of the HVs. With incoming observations, the AV keeps updating the belief on HVs’ intentions. When it is certain that the rear HV is going to yield (in Scenarios A and C), it accelerates to merge into the gap between the two HVs. On the contrary, when AV is certain that the rear HV is not going to yield (in Scenarios B and D), it takes further braking to avoid collisions and merges behind both HVs.

Refer to caption
Figure 3: Longitudinal velocities of AV in different scenarios in Case I.

V-B Case II: Unprotected Left-Turn

Refer to caption
Refer to caption
Refer to caption
(a) Simulation results of Scenario A at 1.5 s, 2.5 s, and 3.5 s.
Refer to caption
Refer to caption
Refer to caption
(b) Simulation results of Scenario B at 1.5 s, 2.5 s, and 4.0 s.
Refer to caption
Refer to caption
Refer to caption
(c) Simulation results of Scenario C at 1.5 s, 2.5 s, and 3.5 s.
Refer to caption
Refer to caption
Refer to caption
(d) Simulation results of Scenario D at 1.5 s, 2.5 s, and 4.0 s.
Refer to caption
Refer to caption
Refer to caption
(e) Simulation results of Scenario E at 1.5 s, 2.5 s, and 3.5 s.
Refer to caption
Refer to caption
Refer to caption
(f) Simulation results of Scenario F at 1.5 s, 2.5 s, and 3.5 s.
Refer to caption
Refer to caption
Refer to caption
(g) Simulation results of Scenario G at 1.5 s, 2.5 s, and 3.5 s.
Refer to caption
Refer to caption
Refer to caption
(h) Simulation results of Scenario H at 2.0 s, 4.0 s, and 5.0 s.
Figure 4: Simulation results for Case II. Successful unprotected left-turning is performed in all scenarios without collisions. Adaptive strategies of AV are demonstrated in these figures, which react responsively to different driving modes of HVs to ensure security and successful passing.

In this case, the goal of AV is to perform an unprotected left-turning in an unsignalized intersection. Two human-driving vehicles are considered in this scenario, including HV1 traveling from the left-hand side of AV and HV2 traveling in the opposite lane. The initial positions of AV, HV1, and HV2 are (15,5)m(15,-5)\,\textup{m}, (5,10)m(-5,10)\,\textup{m}, and (10,35)m(10,35)\,\textup{m}, respectively, and the initial velocity is 7m/s7\,\textup{m}/\textup{s} for all three vehicles. For each vehicle, four different types are considered. Specifically, those types are combinations of navigation types and longitudinal types. For AV, the types concerning navigation include straight-traveling and left-turning and the types concerning longitudinal strategies include aggressive and conservative, such that AV can be an aggressive straight-traveling agent or a conservative left-turning agent, etc. Although the driving intention of the AV is to perform left-turning, it is clearly not known by HVs, and therefore straight-traveling types of AV also need to be modeled (but not selected). The same types are considered for HV1. For HV2, the navigation types include straight-going and right-turning, while longitudinal types are the same. Since vehicles can swerve away from the reference line to avoid each other in an intersection, the action space should include lateral actions and longitudinal actions. The longitudinal actions are the same as in Case I, while the set of lateral action is the set of eligible longitudinal displacement with respect to the corresponding reference lines [1,0.5,0,0.5,1]m[-1,-0.5,0,0.5,1]\,\textup{m}. In this case, 8 different scenarios are considered with respect to different types of HVs, and those scenarios are listed in Table II. The time span for the first stage in the trajectory tree is 1s1\,\textup{s} and that for the second stage is 2s2\,\textup{s}.

Refer to caption
Figure 5: Longitudinal velocities of AV in different scenarios in Case II.
Refer to caption
Figure 6: Lateral displacements of AV in different scenarios in Case II.
TABLE II: Scenario settings of Case I
Scenario HV1 HV2
A Straight-going, Aggressive Straight-going, Aggressive
B Straight-going, Aggressive Straight-going, Conservative
C Straight-going, Conservative Straight-going, Aggressive
D Straight-going, Conservative Straight-going, Conservative
E Left-turning, Aggressive Straight-going, Aggressive
F Left-turning, Aggressive Straight-going, Conservative
G Left-turning, Conservative Straight-going, Aggressive
H Left-turning, Conservative Straight-going, Conservative

The qualitative simulation results are shown in Fig. 4. In general, AV successfully identifies the intentions and types of other vehicles and manages to pass the scenario in a smooth and secure manner. Diversified driving behaviors are exhibited with respect to different driving intentions of HVs. Specifically, in Scenario A where both HVs are aggressive and straight-going, AV slows down and passes through the intersection after both HVs. In Scenario B, AV slows down to yield to the HV1, which is straight-going and aggressive. Then it accelerates to pass in front of HV2, which is considered conservative. In Scenario C where HV1 is straight-going and conservative and HV2 is straight-going and aggressive, the AV swerve to the right to pass through the intersection from behind HV2. In Scenario D where both HVs are straight-going and conservative, AV accelerates and cuts across in front of both HVs. The behavior of AV in Scenario E, F, and G are similar to that in Scenario A, B, and C, respectively. In Scenario H, although both HVs are conservative, HV1 is performing a left-turning and therefore there is no room for AV to cut across. As a result, AV slows down to pass the intersection behind both HVs. In addition to the behaviors of AV, the proposed method also manages to simulate the interactive behaviors of the HVs. For example, in Scenario A, both HVs swerve to the left to keep safety distances from AV, and in Scenario B, HV2 steers to the right to pass through the intersection behind AV, etc. Fig. 3 shows the longitudinal velocities of AV in each scenario, and Fig. 6 shows the lateral displacements. In general, AV decelerates at the beginning to avoid collision when the intentions of HVs remain unclear and then takes responsive actions to different driving behaviors of HVs.

V-C Comparison

Refer to caption
(a) t=1.0st=1.0\,\textup{s}
Refer to caption
(b) t=1.5st=1.5\,\textup{s}
Refer to caption
(c) t=2.0st=2.0\,\textup{s}
Refer to caption
(d) t=2.5st=2.5\,\textup{s}
Figure 7: A failure mode of the baseline method corresponding to Scenario B, Case I. Due to ignorance of the driving modes of HVs, the AV fails to yield to the aggressive rear HV and therefore collision occurs.

To further illustrate the advantage of the proposed method, we compared it with a planning scheme provided by a standard game of complete information. Specifically, the AV is trying to play a game of complete information, in which the action space of each agent is exactly the union of the action spaces of that agent under different types, and other settings are the same as in the game of incomplete information. Meanwhile, all other HVs are still playing the game of incomplete information with determined type. Simple qualitative results are shown in Fig. 7. In Scenario B of the ramp-merging case, since the longitudinal position of HV1 is 2m2\,\textup{m} behind AV and the initial velocities are the same for both vehicles, a typical Nash equilibrium is that AV will merge into the main road in front of the HV1. However, this equilibrium contradicts the actual driving intention of HV1, which is an aggressive agent and intends to accelerate. Due to this discrepancy, a catastrophic result occurs such that AV collides with HV1. This example illustrates the limitations of a traditional game of complete information and demonstrates the necessity of identifying the intentions of other vehicles.

We also present a quantitative analysis to compare the performance of the proposed method and the baseline of the traditional game. Since randomness exists in Monte-Carlo sampling, we perform repeated experiments for both cases with scenario settings described in Section V-A and Section V-B. For passenger comfort, we compare the following statistics of AV: 1) the root mean square of longitudinal acceleration, 2) the average maximum longitudinal acceleration, 3) the root mean square of lateral acceleration, and 4) the average maximum lateral acceleration. For driving safety, we compare 1) the average minimum distances between AV and All of the HVs and 2) the collision rates between AV and HV. Results are shown in Table III. It can be seen that in Case I, the collision rate of the baseline is unacceptable. Although the performance in terms of passenger comfort seems to be better, this is mainly because the baseline method chooses the theoretically optimal strategy disregarding the actual driving intentions of the HVs. In Case II, the proposed method is clearly better than the baseline method in terms of both passenger comfort and driving safety. Meanwhile, we also compare the computation time of both methods. It can be seen from Table IV that the computation efficiency of the proposed method is higher than the baseline, mainly due to the smaller action space.

TABLE III: Comparison of performance between the proposed method and the baseline
Case Method Avg. Max. Long. Acc. RMS. Long. Acc. Avg. Max. Lat. Acc. RMS. Lat. Acc. Min. Dis. Collision Rate
I Proposed 5.204m/s25.204\,\textup{m}/\textup{s}^{2} 2.172m/s22.172\,\textup{m}/\textup{s}^{2} / / 0.716m 0/8\textbf{0}/8
Baseline 3.235m/s2\textbf{3.235}\,\textup{m}/\textup{s}^{2} 1.430m/s2\textbf{1.430}\,\textup{m}/\textup{s}^{2} / / 0.333m0.333\,\textup{m} 4/84/8
II Proposed 3.382m/s2\textbf{3.382}\,\textup{m}/\textup{s}^{2} 1.509m/s2\textbf{1.509}\,\textup{m}/\textup{s}^{2} 2.238m/s2\textbf{2.238}\,\textup{m}/\textup{s}^{2} 0.756m/s2\textbf{0.756}\,\textup{m}/\textup{s}^{2} 1.069m 0/40\textbf{0}/40
Baseline 4.953m/s24.953\,\textup{m}/\textup{s}^{2} 1.915m/s21.915\,\textup{m}/\textup{s}^{2} 2.887m/s22.887\,\textup{m}/\textup{s}^{2} 0.914m/s20.914\,\textup{m}/\textup{s}^{2} 0.547m0.547\,\textup{m} 15/4015/40
TABLE IV: Comparison of computation efficiency between the proposed method and the baseline
Case Method 10000 Iter. 20000 Iter. 50000 Iter.
I Proposed 0.09s 0.14s 0.28s
Baseline 0.10s0.10\,\textup{s} 0.15s0.15\,\textup{s} 0.31s0.31\,\textup{s}
II Proposed 0.09s 0.18s 0.47s
Baseline 0.20s0.20\,\textup{s} 0.35s0.35\,\textup{s} 0.85s0.85\,\textup{s}

VI Conclusion

In this paper, we present an integrated decision-making and trajectory planning framework for autonomous vehicles. Based on the game of incomplete information, multimodal behaviors of human drivers are properly handled, the optimal decision is reached, and the planning is performed in a fully interactive manner. Simulation demonstrates the effectiveness of the proposed approach across multiple traffic scenarios and the improvements in terms of passengers’ comfort and security over traditional game method with smaller accelerations, larger safety distances, and lower collision rates. Possible future works include modeling the correlation between the driving intentions of vehicles and integrating data-driven methods to learn the cost functions online for accurate imitation of human driving behaviors.

References

  • [1] P. Trautman and A. Krause, “Unfreezing the robot: Navigation in dense, interacting crowds,” in 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.   IEEE, 2010, pp. 797–803.
  • [2] S. Jia, Y. Zhang, X. Li, X. Na, Y. Wang, B. Gao, B. Zhu, and R. Yu, “Interactive decision-making with switchable game modes for automated vehicles at intersections,” IEEE Transactions on Intelligent Transportation Systems, vol. 24, no. 11, pp. 11 785–11 799, 2023.
  • [3] P. Hang, C. Huang, Z. Hu, and C. Lv, “Decision making for connected automated vehicles at urban intersections considering social and individual benefits,” IEEE Transactions on Intelligent Transportation Systems, vol. 23, no. 11, pp. 22 549–22 562, 2022.
  • [4] S. L. Cleac’h, M. Schwager, and Z. Manchester, “ALGAMES: A fast solver for constrained dynamic games,” arXiv preprint arXiv:1910.09713, 2019.
  • [5] P. Hang, C. Lv, C. Huang, Y. Xing, and Z. Hu, “Cooperative decision making of connected automated vehicles at multi-lane merging zone: A coalitional game approach,” IEEE Transactions on Intelligent Transportation Systems, vol. 23, no. 4, pp. 3829–3841, 2021.
  • [6] F. Fabiani and S. Grammatico, “Multi-vehicle automated driving as a generalized mixed-integer potential game,” IEEE Transactions on Intelligent Transportation Systems, vol. 21, no. 3, pp. 1064–1073, 2019.
  • [7] V. G. Lopez, F. L. Lewis, M. Liu, Y. Wan, S. Nageshrao, and D. Filev, “Game-theoretic lane-changing decision making and payoff learning for autonomous vehicles,” IEEE Transactions on Vehicular Technology, vol. 71, no. 4, pp. 3609–3620, 2022.
  • [8] P. Hang, C. Huang, Z. Hu, Y. Xing, and C. Lv, “Decision making of connected automated vehicles at an unsignalized roundabout considering personalized driving behaviours,” IEEE Transactions on Vehicular Technology, vol. 70, no. 5, pp. 4051–4064, 2021.
  • [9] M. Liu, I. Kolmanovsky, H. E. Tseng, S. Huang, D. Filev, and A. Girard, “Potential game-based decision-making for autonomous driving,” IEEE Transactions on Intelligent Transportation Systems, vol. 24, no. 8, pp. 8014–8027, 2023.
  • [10] T. Kavuncu, A. Yaraneri, and N. Mehr, “Potential iLQR: A potential-minimizing controller for planning multi-agent interactive trajectories,” arXiv preprint arXiv:2107.04926, 2021.
  • [11] A. Zanardi, P. G. Sessa, N. Käslin, S. Bolognani, A. Censi, and E. Frazzoli, “How bad is selfish driving? bounding the inefficiency of equilibria in urban driving games,” IEEE Robotics and Automation Letters, vol. 8, no. 4, pp. 2293–2300, 2023.
  • [12] R. Chandra and D. Manocha, “GamePlan: Game-theoretic multi-agent planning with human drivers at intersections, roundabouts, and merging,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 2676–2683, 2022.
  • [13] J. F. Fisac, E. Bronstein, E. Stefansson, D. Sadigh, S. S. Sastry, and A. D. Dragan, “Hierarchical game-theoretic planning for autonomous vehicles,” in 2019 International Conference on Robotics and Automation (ICRA).   IEEE, 2019, pp. 9590–9596.
  • [14] H. Yu, H. E. Tseng, and R. Langari, “A human-like game theory-based controller for automatic lane changing,” Transportation Research Part C: Emerging Technologies, vol. 88, pp. 140–158, 2018.
  • [15] Z. Deng, W. Hu, Y. Yang, K. Cao, D. Cao, and A. Khajepour, “Lane change decision-making with active interactions in dense highway traffic: A bayesian game approach,” in 2022 IEEE 25th International Conference on Intelligent Transportation Systems (ITSC).   IEEE, 2022, pp. 3290–3297.
  • [16] T. Zhao, W. ShangGuan, L. Chai, and Y. Cao, “A non-cooperative dynamic game-based approach for lane changing decision-making in mixed traffic diversion scenarios,” in 2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC).   IEEE, 2023, pp. 4776–4781.
  • [17] Y. Zhang, P. Hang, C. Huang, and C. Lv, “Human-like interactive behavior generation for autonomous vehicles: a bayesian game-theoretic approach with turing test,” Advanced Intelligent Systems, vol. 4, no. 5, p. 2100211, 2022.
  • [18] J. Ziegler, P. Bender, M. Schreiber, H. Lategahn, T. Strauss, C. Stiller, T. Dang, U. Franke, N. Appenrodt, C. G. Keller et al., “Making bertha drive—an autonomous journey on a historic route,” IEEE Intelligent Transportation Systems Magazine, vol. 6, no. 2, pp. 8–20, 2014.
  • [19] C. Urmson, J. Anhalt, D. Bagnell, C. Baker, R. Bittner, M. Clark, J. Dolan, D. Duggins, T. Galatali, C. Geyer et al., “Autonomous driving in urban environments: Boss and the urban challenge,” Journal of Field Robotics, vol. 25, no. 8, pp. 425–466, 2008.
  • [20] M. Montemerlo, J. Becker, S. Bhat, H. Dahlkamp, D. Dolgov, S. Ettinger, D. Haehnel, T. Hilden, G. Hoffmann, B. Huhnke et al., “Junior: The stanford entry in the urban challenge,” Journal of Field Robotics, vol. 25, no. 9, pp. 569–597, 2008.
  • [21] J. Wei, J. M. Snider, T. Gu, J. M. Dolan, and B. Litkouhi, “A behavioral planning framework for autonomous driving,” in 2014 IEEE Intelligent Vehicles Symposium Proceedings.   IEEE, 2014, pp. 458–464.
  • [22] W. Zhan, J. Chen, C.-Y. Chan, C. Liu, and M. Tomizuka, “Spatially-partitioned environmental representation and planning architecture for on-road autonomous driving,” in 2017 IEEE Intelligent Vehicles Symposium (IV).   IEEE, 2017, pp. 632–639.
  • [23] Z. Ajanovic, B. Lacevic, B. Shyrokau, M. Stolz, and M. Horn, “Search-based optimal motion planning for automated driving. in 2018 IEEE,” in RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4523–4530.
  • [24] T. Kessler and A. Knoll, “Cooperative multi-vehicle behavior coordination for autonomous driving,” in 2019 IEEE Intelligent Vehicles Symposium (IV).   IEEE, 2019, pp. 1953–1960.
  • [25] J. Ma, Z. Cheng, X. Zhang, M. Tomizuka, and T. H. Lee, “Alternating direction method of multipliers for constrained iterative LQR in autonomous driving,” IEEE Transactions on Intelligent Transportation Systems, vol. 23, no. 12, pp. 23 031–23 042, 2022.
  • [26] Z. Huang, S. Shen, and J. Ma, “Decentralized iLQR for cooperative trajectory planning of connected autonomous vehicles via dual consensus ADMM,” IEEE Transactions on Intelligent Transportation Systems, vol. 24, no. 11, pp. 12 754–12 766, 2023.
  • [27] K. A. Mustafa, D. J. Ornia, J. Kober, and J. Alonso-Mora, “RACP: Risk-aware contingency planning with multi-modal predictions,” IEEE Transactions on Intelligent Vehicles, 2024.
  • [28] M. Werling, S. Kammel, J. Ziegler, and L. Gröll, “Optimal trajectories for time-critical street scenarios using discretized terminal manifolds,” The International Journal of Robotics Research, vol. 31, no. 3, pp. 346–359, 2012.
  • [29] W. Liu, S.-W. Kim, S. Pendleton, and M. H. Ang, “Situation-aware decision making for autonomous driving on urban road using online pomdp,” in 2015 IEEE Intelligent Vehicles Symposium (IV).   IEEE, 2015, pp. 1126–1133.
  • [30] A. G. Cunningham, E. Galceran, R. M. Eustice, and E. Olson, “MPDM: Multipolicy decision-making in dynamic, uncertain environments for autonomous driving,” in 2015 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2015, pp. 1670–1677.
  • [31] L. Zhang, W. Ding, J. Chen, and S. Shen, “Efficient uncertainty-aware decision-making for automated driving using guided branching,” in 2020 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2020, pp. 3291–3297.
  • [32] W. Ding, L. Zhang, J. Chen, and S. Shen, “EPSILON: An efficient planning system for automated vehicles in highly interactive environments,” IEEE Transactions on Robotics, vol. 38, no. 2, pp. 1118–1138, 2022.
  • [33] T. Li, L. Zhang, S. Liu, and S. Shen, “MARC: Multipolicy and risk-aware contingency planning for autonomous driving,” IEEE Robotics and Automation Letters, 2023.
  • [34] Y. Hu, J. Yang, L. Chen, K. Li, C. Sima, X. Zhu, S. Chai, S. Du, T. Lin, W. Wang et al., “Planning-oriented autonomous driving,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 17 853–17 862.
  • [35] Z. Huang, H. Liu, J. Wu, and C. Lv, “Differentiable integrated motion prediction and planning with learnable cost function for autonomous driving,” IEEE transactions on neural networks and learning systems, 2023.
  • [36] Z. Huang, H. Liu, and C. Lv, “Gameformer: Game-theoretic modeling and learning of transformer-based interactive prediction and planning for autonomous driving,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 3903–3913.
  • [37] D. Fridovich-Keil, E. Ratner, L. Peters, A. D. Dragan, and C. J. Tomlin, “Efficient iterative linear-quadratic approximations for nonlinear multi-player general-sum differential games,” in 2020 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2020, pp. 1475–1481.
  • [38] C. Li, T. Trinh, L. Wang, C. Liu, M. Tomizuka, and W. Zhan, “Efficient game-theoretic planning with prediction heuristic for socially-compliant autonomous driving,” IEEE Robotics and Automation Letters, vol. 7, no. 4, pp. 10 248–10 255, 2022.
  • [39] W. Schwarting, A. Pierson, S. Karaman, and D. Rus, “Stochastic dynamic games in belief space,” IEEE Transactions on Robotics, vol. 37, no. 6, pp. 2157–2172, 2021.
  • [40] N. Mehr, M. Wang, M. Bhatt, and M. Schwager, “Maximum-entropy multi-agent dynamic games: Forward and inverse solutions,” IEEE Transactions on Robotics, vol. 39, no. 3, pp. 1801–1815, 2023.
  • [41] H. Shao, M. Zhang, T. Feng, and Y. Dong, “A discretionary lane-changing decision-making mechanism incorporating drivers’ heterogeneity: A signalling game-based approach,” Journal of Advanced Transportation, vol. 2020, no. 1, p. 8892693, 2020.
  • [42] R. Yao and X. Du, “Modelling lane changing behaviors for bus exiting at bus bay stops considering driving styles: A game theoretical approach,” Travel Behaviour and Society, vol. 29, pp. 319–329, 2022.
  • [43] L. Li, W. Zhao, and C. Wang, “Cooperative merging strategy considering stochastic driving style at on-ramps: A bayesian game approach,” Automotive Innovation, vol. 7, no. 2, pp. 312–334, 2024.
  • [44] L. Peters, A. Bajcsy, C.-Y. Chiu, D. Fridovich-Keil, F. Laine, L. Ferranti, and J. Alonso-Mora, “Contingency games for multi-agent interaction,” IEEE Robotics and Automation Letters, 2024.
  • [45] M. Zinkevich, M. Johanson, M. Bowling, and C. Piccione, “Regret minimization in games with incomplete information,” Advances in Neural Information Processing Systems, vol. 20, 2007.
  • [46] M. Lanctot, K. Waugh, M. Zinkevich, and M. Bowling, “Monte carlo sampling for regret minimization in extensive games,” Advances in Neural Information Processing Systems, vol. 22, 2009.
  • [47] V. Lisỳ, M. Lanctot, and M. H. Bowling, “Online monte carlo counterfactual regret minimization for search in imperfect information games.” in AAMAS, 2015, pp. 27–36.
  • [48] A. Celli, A. Marchesi, T. Bianchi, and N. Gatti, “Learning to correlate in multi-player general-sum sequential games,” Advances in Neural Information Processing Systems, vol. 32, 2019.
  • [49] J. Hartline, V. Syrgkanis, and E. Tardos, “No-regret learning in bayesian games,” Advances in Neural Information Processing Systems, vol. 28, 2015.
  • [50] M. Maschler, S. Zamir, and E. Solan, Game Theory.   Cambridge, U.K.: Cambridge University Press, 2013.
  • [51] P. K. Sen and J. M. Singer, Large sample methods in statistics: An introduction with applications.   Boca Raton, FL: Chapman & Hall CRC, 2000.