This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Observation Adaptation via Annealed Importance Resampling for Partially Observable Markov Decision Processes

Yunuo Zhang 1, Baiting Luo 1, Ayan Mukhopadhyay 1, Abhishek Dubey 1
Abstract

Partially observable Markov decision processes (POMDPs) are a general mathematical model for sequential decision-making in stochastic environments under state uncertainty. POMDPs are often solved online, which enables the algorithm to adapt to new information in real time. Online solvers typically use bootstrap particle filters based on importance resampling for updating the belief distribution. Since directly sampling from the ideal state distribution given the latest observation and previous state is infeasible, particle filters approximate the posterior belief distribution by propagating states and adjusting weights through prediction and resampling steps. However, in practice, the importance resampling technique often leads to particle degeneracy and sample impoverishment when the state transition model poorly aligns with the posterior belief distribution, especially when the received observation is highly informative. We propose an approach that constructs a sequence of bridge distributions between the state-transition and optimal distributions through iterative Monte Carlo steps, better accommodating noisy observations in online POMDP solvers. Our algorithm demonstrates significantly superior performance compared to state-of-the-art methods when evaluated across multiple challenging POMDP domains.

Introduction

Partially observable Markov decision processes (POMDPs) provide a general mathematical framework for modeling decision-making problems under uncertainty, where the true state of the environment is not fully observable and actions have probabilistic outcomes (Kaelbling, Littman, and Cassandra 1998). These models have been successfully applied to various real-world scenarios, including time-critical UAV search and rescue operations where efficient path planning must balance computational constraints with effective decision-making (Zhang et al. 2024). However, POMDPs face challenges to solve exactly due to the “curse of dimensionality” and the “curse of history”, which make the computation of optimal policies non-scalable (Papadimitriou and Tsitsiklis 1987). To address these computational and scalability issues, online planning algorithms have emerged as a prominent approach. Instead of computing a full policy offline, online planning interleaves planning and execution by focusing computational resources on the current belief state and the immediate decisions to be made (Ross et al. 2008).

A critical component of online planning algorithms is the belief update process where belief denotes a distribution over the possible states. At each node in the search tree, the algorithm needs to update the belief state based on simulated actions and observations. However, performing exact belief updates is computationally infeasible in large state spaces due to the high dimensionality of the belief space (Rodriguez, Parr, and Koller 1999). To manage this, online planning algorithms approximate the belief state using sampling-based methods. They represent the belief state as a collection of sampled states, or particles, rather than as explicit probability distributions (Silver and Veness 2010). During the planning process, belief updates are performed by propagating these particles through the state-transition and observation models, using methods such as direct sampling and sequential importance sampling (Doucet and Johansen 2009).

However, these approximation methods face a significant issue: the variance of their estimations increases exponentially with the search depth. This variance escalation is primarily due to the accumulation of uncertainty from observations, which provide imperfect information about the true current state. As the search depth increases, the divergence between the sampled belief and the target posterior can grow substantially, potentially limiting the effectiveness of these planning algorithms, especially in deeper searches where precise belief representation is crucial.

Recognizing this challenge, we propose an approach, AIROAS (Annealed Importance Resampling for Observation Adaptation Search), a novel approach for POMDP planning that combines tree search with particle-based belief representation and annealed importance resampling. It aims to reduce the increasing variance by gradually refining the sampled states. Instead of relying solely on samples drawn from the state-transition distribution, we implement an annealed importance sampling process that incrementally transforms these samples. The goal is to shift the distribution of these samples closer to the ideal state distribution, which would be conditioned on the latest observation and the previous state. While directly sampling from this ideal distribution is impossible, our approach approximates it by using Markov Chain Monte-Carlo approximation. By aligning the sampled states more closely with the most probable current states given the available information, we improve the quality of the belief representation. This enhancement leads to more accurate state estimations, particularly at greater search depths.

Background

POMDPs

Partially observable Markov decision processes (POMDPs) are a framework for modeling decision-making under uncertainty, defined by the tuple (S,A,T,R,Ω,O)(S,A,T,R,\Omega,O). Here, SS is the set of states, AA the set of actions, T(s|s,a)T(s^{\prime}|s,a) the transition function, R(s,a)R(s,a) the reward function, Ω\Omega the set of observations, and O(o|s,a)O(o|s^{\prime},a) the observation function. In POMDPs, at time t, the agent maintains a belief state btb_{t}, a probability distribution over states. This distribution is updated after action ata_{t} and observation ot+1o_{t+1} using the equation: bt+1(s)=ηO(ot+1|s,at)sST(s|s,at)bt(s)b_{t+1}(s^{\prime})=\eta O(o_{t+1}|s^{\prime},a_{t})\sum_{s\in S}T(s^{\prime}|s,a_{t})b_{t}(s), where η\eta is a normalizing factor. This belief update process enables the agent to estimate its current state despite partial observability, facilitating informed decision-making in uncertain environments.

Particle Filters

Particle filters  (Arulampalam et al. 2002), also known as sequential Monte Carlo (SMC) methods, are computational algorithms used to estimate the state of a dynamic system when it is observed through noisy measurements. In the context of Partially observable Markov decision processes (POMDPs), particle filters offer an efficient way to approximate the belief state by maintaining a set of weighted samples (particles) that represent the probability distribution over possible states. The widely-known bootstrap particle filter algorithm  (Gordon, Salmond, and Smith 1993) works by recursively propagating these particles through the state transition model and updating their weights based on new observations. In POMDPs, at each time step tt, the particle filter aims to approximate the optimal posterior distribution:

p(st|ot,st1,a)p(ot|st,a)p(st|st1,a)p(s_{t}|o_{t},s_{t-1},a)\propto p(o_{t}|s_{t},a)\cdot p(s_{t}|s_{t-1},a) (1)

where sts_{t} is the current state, oto_{t} is the current observation, st1s_{t-1} is the previous state, and aa is the action taken.

A crucial step in bootstrap particle filter is resampling (Kitagawa 1996), which addresses the problem of particle degeneracy. Over time, some particles may have negligible weights, leading to poor representation of the state space. Resampling involves randomly drawing particles with replacement from the current set, with probabilities proportional to their weights, and then resetting all weights to 1/N1/N, where NN is the number of particles. However overly frequent resampling can lead to sample impoverishment, where the particle set loses diversity and fails to adequately represent the full state space. To determine when to resample, the Effective Sample Size (ESS) (Forthofer, Lee, and Hernandez 2007) is often used:

ESS=1i=1N(win)2ESS=\frac{1}{\sum_{i=1}^{N}(w_{i}^{n})^{2}} (2)

where winw_{i}^{n} are the normalized weights. When the ESS falls below a predefined threshold (typically N/2N/2), resampling is triggered.

Although the bootstrap particle filter is relatively easy to implement, its accuracy can be significantly compromised when there is a substantial discrepancy between the optimal posterior distribution described in equation (1) and the state-transition distribution. This limitation becomes particularly pronounced in scenarios where the one-step received observation is highly informative about the true state, leading to a sharply peaked posterior distribution. In such cases, the bootstrap filter’s reliance on the state-transition model for proposal generation may lead to particles being proposed in regions where the posterior has significant mass but the proposal density is low, resulting in many particles having negligible weights. Consequently, this can exacerbate the problem of sample degeneracy, necessitating more frequent resampling and potentially leading to sample impoverishment.

Annealed Importance Sampling

Importance Sampling (IS) is a technique for estimating properties of a target distribution by sampling from a different, easier-to-sample proposal distribution and reweighting the samples (Kloek and van Dijk 1978). However, its major limitation lies in its inefficiency when the proposal distribution differs significantly from the target distribution — in such cases, only a few samples that happen to fall in the high-probability regions of the target distribution receive high importance weights, while most samples have negligible weights. This leads to high variance in the estimates and poor sample efficiency.

Annealed Importance Sampling (AIS) constructs a sequence of intermediate distributions that gradually bridge the gap between the proposal and target distributions (Neal 1998). Given a target density p(x)p(x) and a proposal density q(x)q(x), AIS defines a sequence of intermediate distributions:

πk(x)=p(x)βkq(x)(1βk)\pi_{k}(x)=p(x)^{\beta_{k}}q(x)^{(1-\beta_{k})} (3)

where 0=β0βK=10=\beta_{0}\leq\cdots\leq\beta_{K}=1 represents a sequence of inverse temperatures. The method proceeds by first drawing an initial sample x0π0(x)x_{0}\sim\pi_{0}(x), then evolving this sample through a series of transition kernels:

xkTk(xk|xk1) for k=1,,Kx_{k}\sim T_{k}(x_{k}|x_{k-1})\text{ for }k=1,\ldots,K (4)

where each TkT_{k} is constructed to leave πk\pi_{k} invariant. The final importance weight is computed as the product of ratios between successive distributions:

w=π1(x1)π0(x1)πK(xK)πK1(xK)w=\frac{\pi_{1}(x_{1})}{\pi_{0}(x_{1})}\cdots\frac{\pi_{K}(x_{K})}{\pi_{K-1}(x_{K})} (5)

This gradual transition through intermediate distributions helps AIS overcome the limitations of standard importance sampling by maintaining better overlap between successive distributions, resulting in more reliable estimates.

Related Work

Monte Carlo tree search (MCTS) has demonstrated success in solving large POMDPs online. POMCP (Silver and Veness 2010) pioneered this approach by combining a UCT-based tree search with particle filtering for belief updates, but faces challenges with continuous observation spaces. DESPOT and its variant AR-DESPOT (Ye et al. 2017) improve upon POMCP by focusing the search on a fixed set of scenarios and using dual bounds to guide exploration, making it more robust for large discrete problems. POMCPOW (Sunberg and Kochenderfer 2018) extends these ideas to continuous state-action-observation spaces by incorporating progressive widening and weighted particle filtering.

A common thread among these approaches is their reliance on particle filtering techniques, specifically bootstrap particle filtering or Sequential Importance Resampling (SIR), to update belief states. While these filtering methods are computationally efficient, they can sometimes lead to particle degeneracy issues, especially in continuous observation spaces or when observations are unlikely under the current belief.

Most recently, AdaOPS (Wu et al. 2021) attempts to address this challenge using KLD-sampling, which dynamically adapts the number of particles based on the Kullback-Leibler divergence between the true and approximated distributions. However, KLD-sampling exhibits two major limitations: First, it requires discretizing the state space into bins for divergence calculation, which becomes computationally prohibitive especially in high-dimensional spaces (Li, Sun, and Sattar 2013). Second, and more critically, while it adjusts particle quantities, it does not modify the particle values themselves, leading to potential underestimation of distribution variance and failure to capture multimodal aspects of the target distribution. In contrast, Annealed Importance Sampling (AIS) offers several compelling advantages: it constructs a sequence of intermediate distributions that gradually bridge the prior and posterior distributions, effectively maintaining particle diversity while preventing degeneracy. This annealing process enables particles to adaptively migrate toward regions of high posterior probability, making it particularly effective when dealing with concentrated observation likelihoods or significant disparities between prior and posterior distributions. The gradual transition through intermediate distributions allows AIS to better capture the full structure of multimodal distributions and provide more accurate representations of distribution tails compared to KLD-sampling approaches.

Previous work has explored adding tempering iterations to particle filters, where the optimal posterior distribution is constructed adaptively through a sequence of Monte Carlo steps. For example, M. Johansen (2015) developed a block-tempered particle filter that uses bridge distributions to gradually adapt particle values, while (Herbst and Schorfheide 2019) proposed a tempered particle filter that sequentially reduces inflated measurement error variance in Dynamic Stochastic General Equilibrium (DSGE) models. However, these tempering approaches have been primarily studied in the context of state estimation and system identification, but not yet explored within the POMDP planning literature. To our knowledge, we are the first to investigate using bridge distributions to connect target and proposal distributions in particle filtering specifically for POMDP planning, combining the strengths of both sequential Monte Carlo methods and POMDP planning algorithms.

Approach

In this section, we present AIROAS (Annealed Importance Resampling for Observation Adaptation Search), a novel approach for POMDP planning that combines tree search with particle-based belief representation and annealed importance resampling. The complete procedure is detailed in algorithm 1 and  fig. 1. AIROAS constructs a search tree that alternates between belief nodes and action nodes. Each belief node approximates the optimal posterior distribution using a set of weighted particles, as illustrated in fig. 1, and maintains both upper and lower bounds of the optimal value. Using a sequence of bridging distributions, the algorithm refines the value bounds and adjusts particle states and weights to better approximate the optimal posterior distribution described in equation (1). At each timestep, starting from the current belief b0b_{0} as the root node, AIROAS expands the belief tree by exploring various paths from the initial states (Line 3 in algorithm 1). A key characteristic of this structure is that sibling nodes share identical particle states but maintain distinct weight distributions. This configuration persists until annealed importance resampling is applied to mutate the particle values, as we will discuss below.

Selection

Refer to caption
Figure 1: Process of AIROAS Tree Search. AIROAS constructs a search tree that alternates between belief nodes and action nodes. AIR represents Annealed Importance Resampling.
Algorithm 1 AIROAS
1:initial belief b¯0\bar{b}_{0}, maximum depth MAX_DEPTH
2:while Time Allowed and l(b¯0)<u(b¯0)l(\bar{b}_{0})<u(\bar{b}_{0}) do
3:     b¯b¯0\bar{b}\leftarrow\bar{b}_{0}
4:     while depth(b¯)<MAX_DEPTH(\bar{b})<\text{MAX\_DEPTH} do
5:         if b¯\bar{b} is a leaf node then
6:              b¯Annealed Importance Resampling(b¯)\bar{b}\leftarrow\text{Annealed Importance Resampling}(\bar{b})
7:              ExpandAndBackup(b¯)(\bar{b})
8:         end if
9:         b¯\bar{b}\leftarrow Selection(b¯)(\bar{b})
10:     end while
11:     if depth(b¯)MAX_DEPTH(\bar{b})\geq\text{MAX\_DEPTH} then
12:         u(b¯)l(b¯)u(\bar{b})\leftarrow l(\bar{b})
13:     end if
14:end while
15:return argmaxa𝒜l(b¯0,a)\arg\max_{a\in\mathcal{A}}l(\bar{b}_{0},a)
Algorithm 2 Selection(b¯)(\bar{b})
1:a=argmaxaAu(b¯,a)a^{*}=\operatorname*{arg\,max}_{a\in A}u(\bar{b},a)
2:oargmaxoOb¯,ap^(o|b¯,a)EU(τ(b¯,a,o))o^{*}\leftarrow\arg\max_{o\in O_{\bar{b},a}}\hat{p}(o|\bar{b},a)\text{EU}(\tau(\bar{b},a^{*},o))
3:b¯τ(b¯,a,o)\bar{b}\leftarrow\tau(\bar{b},a^{*},o^{*})

Similar to (Wu et al. 2021; Ye et al. 2017; Smith and Simmons 2004), during online planning, AIROAS maintains both upper and lower bounds, u(b¯,a)u(\bar{b},a) and l(b¯,a)l(\bar{b},a), for each action node aa at a belief node b¯\bar{b}. These bounds estimate the optimal value that can be achieved by taking action aa at belief b¯\bar{b}. The action selection follows an optimistic strategy - at each belief node b¯\bar{b}, AIROAS chooses the action aa^{*} that maximizes the upper bound(Line 1 in algorithm 2):

a=argmaxaAu(b¯,a)a^{*}=\operatorname*{arg\,max}_{a\in A}u(\bar{b},a) (6)

After selecting an action, we need to choose which observation branch to explore next. This choice is guided by the probability-weighted excess uncertainty (EU) criterion (Ye et al. 2017; Smith and Simmons 2004) (Line 2 in algorithm 2). For each possible observation oo after taking action aa^{*}, we compute:

p^(o|b¯,a)EU(τ(b¯,a,o))\hat{p}(o|\bar{b},a^{*})\cdot EU(\tau(\bar{b},a^{*},o)) (7)

where p^(o|b¯,a)\hat{p}(o|\bar{b},a^{*}) is an estimation of the probability p(o|b¯,a)p(o|\bar{b},a^{*}) i.e estimates the probability of receiving observation oo after taking action aa^{*} at belief b¯\bar{b}, τ(b¯,a,o)\tau(\bar{b},a^{*},o) represents the updated belief after executing aa^{*} and receiving observation oo, and EU()EU(\cdot) measures the excess uncertainty at a belief node  (Wu et al. 2021). The excess uncertainty quantifies the degree of value uncertainty at each belief node:

EU(b¯)=[u(b¯)l(b¯)]ξ[u(b¯0)l(b¯0)]γd(b¯)EU(\bar{b})=[u(\bar{b})-l(\bar{b})]-\frac{\xi[u(\bar{b}_{0})-l(\bar{b}_{0})]}{\gamma^{d(\bar{b})}} (8)

where u(b¯)u(\bar{b}) and l(b¯)l(\bar{b}) are the upper and lower bounds at belief b¯\bar{b}, b¯0\bar{b}_{0} is the root belief, d(b¯)d(\bar{b}) denotes the depth of belief b¯\bar{b} in the tree, γ\gamma is the discount factor, and ξ(0,1)\xi\in(0,1) controls the target uncertainty reduction at the root. The observation oo^{*} that maximizes the weighted excess uncertainty is selected for exploration, effectively focusing the search on belief states with high uncertainty and high probability of being reached (Wu et al. 2021).

Expand And Backup

When encountering a leaf node, we first perform annealed importance resampling to adjust the particle states and weights (detailed in the next section). Then, AIROAS expands this leaf node by creating child nodes for each possible action aa. For each action node, the expansion process involves propagating the parent belief node’s particles forward using a simulator 𝒢\mathcal{G} that generates state transitions (Line 6 in algorithm 3):

(s,o,r)𝒢(s,a)(s^{\prime},o,r)\sim\mathcal{G}(s,a)
Algorithm 3 ExpandAndBackup(b¯)(\bar{b})
1:for aa \in 𝒜\mathcal{A} do
2:     Create Action Node (b¯,a\bar{b},a)
3:     bb^{\prime} \leftarrow Empty Particles Set
4:     𝒪\mathcal{O} \leftarrow Empty Observation Set
5:     for s \in b¯\bar{b} do
6:         (s,o,r)𝒢(s,a)(s\prime,o,r)\sim\mathcal{G}(s,a)
7:         𝒪\mathcal{O} \leftarrow 𝒪\mathcal{O} \cup o
8:         weight(ss\prime) \leftarrow 1
9:         bb^{\prime} \leftarrow bb^{\prime} \cup ss\prime
10:     end for
11:     for o \in 𝒪\mathcal{O} do
12:         b~\tilde{b} \leftarrow DeepCopy(bb^{\prime})
13:         for s \in b~\tilde{b} do
14:              weight(s) = p(o|s,a)p(o|s,a) \cdot weight(s)
15:         end for
16:         Create Belief Node b~\tilde{b}
17:     end for
18:end for
19:BackUp(b¯\bar{b})

The expansion continues by creating new belief nodes for each unique observation encountered during particle propagation. Specifically, after propagating each particle through 𝒢\mathcal{G}, we obtain a set of updated particle states, initially assigning each particle a weight of 1. For each observation oo generated during this process, we create a new belief node where the particle weights are updated according to the observation likelihood p(o|s,a)p(o|s,a) — the probability of receiving observation oo given the particle state ss and action aa. This results in a tree structure where each action node branches into multiple belief nodes based on possible observations, with each belief node containing weighted particles that represent the updated belief state.

After expanding belief nodes, we perform a backup operation to update the bounds of their ancestor nodes. Similar as (Ye et al. 2017) and (Wu et al. 2021), this backup process is implemented by recursively applying the Bellman equation, which calculates the optimal value of a belief state in two parts. First, we consider the expected immediate reward for taking an action in the current belief state, calculated by looking at each possible state, weighing its reward by how likely we think we are in that state according to our current belief. Second, we consider the long-term value by looking at all possible observations we might receive after taking an action. For each possible observation, we calculate how likely we are to see that observation, determine what our new belief state would be after seeing it, consider the value of that resulting belief state, and weight this value by the probability of getting that observation. The total value is the sum of these immediate and future components, and we choose the action that maximizes this total value. In practice, since we can’t compute exact values over continuous states and observations, we approximate this calculation using discrete sums over our particle-based belief representation.

Annealed Importance Resampling

As discussed above, the belief of each belief node is generated using a state-transition function from its parent belief node. Suppose the leaf node represents belief at timestep t and its parent belief node represents belief at timestep t-1, while our simulator 𝒢\mathcal{G} implements the state-transition function p(st|st1,a)p(s_{t}|s_{t-1},a) during expansion, our ultimate objective is to have the final particle distribution approximate the optimal posterior distribution p(st|ot,st1,a)p(s_{t}|o_{t},s_{t-1},a) referenced in eq. 1.

Algorithm 4 Annealed Importance Resampling(b¯\bar{b})
1:b¯\bar{b} from current belief node, sequence of tempering parameters 0=β0βK=10=\beta_{0}\leq\cdots\leq\beta_{K}=1, target inefficiency ratio rr^{*}, transition kernels T1(s,s),,TK(s,s)T_{1}(s,s^{\prime}),\ldots,T_{K}(s,s^{\prime})
2:for βk\beta_{k} in {β0,,βK}\{\beta_{0},\ldots,\beta_{K}\} do
3:     if k==0k==0 then
4:         Continue
5:     end if
6:     weights(b¯\bar{b}) \leftarrow Update_Weights(b¯\bar{b}, βk\beta_{k}, βk1\beta_{k-1})
7:     if In_eff(b¯\bar{b}) r\leq r^{*} then
8:         break
9:     else
10:         rr^{*} \leftarrow In_eff(b¯\bar{b})
11:     end if
12:     Resample b¯\bar{b}
13:     b¯\bar{b} \leftarrow Mutation(b¯\bar{b})
14:end for

The primary challenge in this context stems from the impossibility of directly sampling from the optimal posterior distribution. Previous work (Silver and Veness 2010; Ye et al. 2017; Sunberg and Kochenderfer 2018; Wu et al. 2021) approximates sampling from this optimal posterior distribution using Sequential Importance Resampling, treating the state-transition distribution as proposal distribution and the optimal posterior distribution as target distribution. However, this approach suffers from particle degeneracy issues, particularly when observations are highly informative about the true state, leading to a sharply peaked posterior distribution and a significant divergence between the optimal posterior and the proposal distribution derived from the current belief (Wu et al. 2021).

To address the issue, Annealed Importance Sampling (Neal 1998) introduced in Background offers an effective solution by constructing intermediate bridging distributions between the proposal distribution and target distribution, gradually guiding particles toward high-probability regions of the target distribution by constructing a sequence of intermediate distributions parameterized by tempering parameters 0=β0β1βK=10=\beta_{0}\leq\beta_{1}\leq\cdots\leq\beta_{K}=1.

Building upon these principles, we integrate Annealed Importance Resampling (AIR), which extends AIS with a resampling step, into our AIROAS algorithm to address the particle degeneracy issue inherent in POMDP planning. Our approach adapts this concept to the POMDP planning context by designing intermediate distributions that account for observations. By gradually bridging the proposal and target distributions, particles are adaptively reweighted and resampled, allowing them to migrate toward regions of high posterior probability. This gradual transition through intermediate distributions not only addresses degeneracy but also enables our approach to capture the multimodal characteristics of the optimal posterior distribution more effectively.

These tempering parameters are generated using a sigmoid function applied to a linearly spaced sequence between 10310^{-3} and 1. Specifically, for a sequence {x1,x2,xK}\{x_{1},x_{2},\cdots x_{K}\} linearly spaced between 10310^{-3} and 1, each tempering parameter βi\beta_{i} is computed as:

βi=11+e10(xi0.5)\beta_{i}=\frac{1}{1+e^{-10(x_{i}-0.5)}} (9)

This sigmoid transformation ensures a smooth progression of β\beta values from near zero to one, with denser sampling in the middle range.

The intermediate distributions are characterized by their probability densities:

πk=p(st|ot,st1,a)βkp(st|st1,a)(1βk)\pi_{k}=p(s_{t}|o_{t},s_{t-1},a)^{\beta_{k}}p(s_{t}|s_{t-1},a)^{(1-\beta_{k})} (10)

When k=Kk=K, this formulation exactly matches the optimal posterior distribution. By progressively transforming particles sampled from the prior distribution p(st|st1,a)p(s_{t}|s_{t-1},a) through the sequence π1,π2,,πK\pi_{1},\pi_{2},\ldots,\pi_{K}, we can effectively approximate particle states drawn from the optimal posterior distribution.

The overall procedure of annealed importance resampling is described in algorithm 4.

Weights

To update weights of the particles at iteration kk and timestep t, for every βk\beta_{k} in the sequence {β0,,βK}\{\beta_{0},\ldots,\beta_{K}\}, we transform the particle approximation from πk1\pi_{k-1} to πk\pi_{k} by recalculating the weights. Assume the particle set at iteration kk and timestep t is StkS_{tk}, for each particle stkjs_{tk}^{j} in the particle set StkS_{tk} at iteration kk, the weight is updated as:

wtkjp(otstkj,at)βkp(otstkj,at)βk1wt(k1)jw_{tk}^{j}\propto\frac{p(o_{t}\mid s_{tk}^{j},a_{t})^{\beta_{k}}}{p(o_{t}\mid s_{tk}^{j},a_{t})^{\beta_{k-1}}}\cdot w_{t(k-1)}^{j} (11)
Proof.

At iteration k-1 and timestep t, for particle st(k1)js_{t(k-1)}^{j}, we have computed its weight as:

wt(k1)j=π1(st1j)π0(st1j)πk1(st(k1)j)πk2(st(k1)j)w_{t(k-1)}^{j}=\frac{\pi_{1}(s_{t1}^{j})}{\pi_{0}(s_{t1}^{j})}\cdots\frac{\pi_{k-1}(s_{t(k-1)}^{j})}{\pi_{k-2}(s_{t(k-1)}^{j})}

Thus

wtkj=πk(stkj)πk1(stkj)wt(k1)jw_{tk}^{j}=\frac{\pi_{k}(s_{tk}^{j})}{\pi_{k-1}(s_{tk}^{j})}\cdot w_{t(k-1)}^{j}

From eq. 10 and eq. 1, we know:

πk(stkj)\displaystyle\pi_{k}(s_{tk}^{j}) =p(stkj|ot,st1j,at)βkp(stkj|st1j,at)(1βk)\displaystyle=p(s_{tk}^{j}|o_{t},s_{t-1}^{j},a_{t})^{\beta_{k}}p(s_{tk}^{j}|s_{t-1}^{j},a_{t})^{(1-\beta_{k})}
p(ot|stkj,at)βkp(stkj|st1j,at)βkp(stkj|st1j,at)(1βk)\displaystyle\propto p(o_{t}|s_{tk}^{j},a_{t})^{\beta_{k}}p(s_{tk}^{j}|s_{t-1}^{j},a_{t})^{\beta_{k}}p(s_{tk}^{j}|s_{t-1}^{j},a_{t})^{(1-\beta_{k})}
p(ot|stkj,at)βkp(stkj|st1j,at)\displaystyle\propto p(o_{t}|s_{tk}^{j},a_{t})^{\beta_{k}}p(s_{tk}^{j}|s_{t-1}^{j},a_{t})

Similarly, we can get:

πk1(stkj)p(ot|stkj,at)βk1p(stkj|st1j,at)\pi_{k-1}(s_{tk}^{j})\propto p(o_{t}|s_{tk}^{j},a_{t})^{\beta_{k-1}}p(s_{tk}^{j}|s_{t-1}^{j},a_{t})

where oto_{t} and aa is the observation and action leading to this belief node. Thus we have

wtkj\displaystyle w_{tk}^{j} =πk(stkj)πk1(stkj)wt(k1)j\displaystyle=\frac{\pi_{k}(s_{tk}^{j})}{\pi_{k-1}(s_{tk}^{j})}\cdot w_{t(k-1)}^{j}
p(ot|stkj,at)βkp(stkj|st1j,at)p(ot|stkj,at)βk1p(stkj|st1j,at)wt(k1)j\displaystyle\propto\frac{p(o_{t}|s_{tk}^{j},a_{t})^{\beta_{k}}p(s_{tk}^{j}|s_{t-1}^{j},a_{t})}{p(o_{t}|s_{tk}^{j},a_{t})^{\beta_{k-1}}p(s_{tk}^{j}|s_{t-1}^{j},a_{t})}\cdot w_{t(k-1)}^{j}
p(ot|stkj,at)βkp(ot|stkj,at)βk1wt(k1)j\displaystyle\propto\frac{p(o_{t}|s_{tk}^{j},a_{t})^{\beta_{k}}}{p(o_{t}|s_{tk}^{j},a_{t})^{\beta_{k-1}}}\cdot w_{t(k-1)}^{j}

This update reflects the incremental adjustment of weights based on the annealing parameter βk\beta_{k}. Specifically, at iteration kk, the importance weight of each particle stkjs_{tk}^{j} is updated by multiplying its previous weight wt(k1)jw_{t(k-1)}^{j} with an incremental weight. This incremental weight is computed as the ratio of two observation likelihood terms: the observation density function raised to the current tempering parameter βk\beta_{k} in the numerator, and the same function raised to the previous tempering parameter βk1\beta_{k-1} in the denominator. Specifically, the incremental weight is p(ot|stkj,at)βkp(ot|stkj,at)βk1\frac{p(o_{t}|s_{tk}^{j},a_{t})^{\beta_{k}}}{p(o_{t}|s_{tk}^{j},a_{t})^{\beta_{k-1}}}, where oto_{t} is the observation received at timestep tt and ata_{t} is the action taken.

After updating the particle weights, we calculate their inefficiency score in a similar way as (Herbst and Schorfheide 2019). For a particle set containing MM particles, the inefficiency score is computed as:

In_eff(b¯tk)=1Mj=1M(wtkj1Mj=1Mwtkj)2\text{In\_eff}(\bar{b}_{tk})=\frac{1}{M}\sum_{j=1}^{M}(\frac{w_{tk}^{j}}{\frac{1}{M}\sum_{j=1}^{M}w_{tk}^{j}})^{2} (12)

This score reflects the variance in particle weights. As in (Herbst and Schorfheide 2019) and specified in algorithm 4, we maintain a pre-defined target inefficiency ratio rr^{*}. If the inefficiency score exceeds rr^{*}, indicating that the variance of the resulting weights remains high, we continue adjusting the particle states and weights. Otherwise, we exit the current iteration.

Mutation

For each particle st(k1)js_{t(k-1)}^{j}, the algorithm employs a Markov transition kernel Tk(sk1j,skj)T_{k}(s_{k-1}^{j},s_{k}^{j}) to evolve the state from time k1k-1 to kk. This transition kernel is constructed to preserve πk\pi_{k} as its invariant distribution, as specified in eq. 4.

Algorithm 5 Mutation(b¯\bar{b})
1:for particle st(k1)js_{t(k-1)}^{j} in b¯\bar{b} do
2:     Tk(sk1j,)T_{k}(s_{k-1}^{j},\cdot) \leftarrow 𝒩(st(k1)j,σ2I)\mathcal{N}(s_{t(k-1)}^{j},\sigma^{2}\cdot I)
3:     stkjTk(sk1j,)s_{tk}^{j}\sim T_{k}(s_{k-1}^{j},\cdot)
4:     Compute the acceptance probability p
5:     stkjs_{tk}^{j} \leftarrow stkjs_{tk}^{j} for probability p, otherwise stkjs_{tk}^{j} \leftarrow st(k1)js_{t(k-1)}^{j}
6:end for

The transition mechanism implements a Metropolis-Hastings kernel (Hastings 1970) with a multivariate Gaussian proposal distribution. Given a state vector st(k1)js_{t(k-1)}^{j} that can be decomposed into components st(k1)j=<x,y,>s_{t(k-1)}^{j}=<x,y,\cdots>, the proposal distribution is centered at st(k1)js_{t(k-1)}^{j} with covariance structure. The forward proposal distribution i.e the transition kernel generates new states according to:

stkj𝒩(st(k1)j,σ2I)s_{tk}^{j}\sim\mathcal{N}(s_{t(k-1)}^{j},\sigma^{2}\cdot I)

where II denotes the identity matrix and σ2\sigma^{2} is a scaling parameter calibrated proportionally to the L1 distance between the current state st(k1)js_{t(k-1)}^{j} and the observation oto_{t} in the state space. Then we define its reverse transition kernel Tk~(skj)\tilde{T_{k}}(s_{k}^{j}) as:

𝒩(stkj,σ2I)\mathcal{N}(s_{tk}^{j},\sigma^{2}\cdot I)

The acceptance rate shown in algorithm 5 is computed as:

paccept=min{1,Tk(st(k1)j,stkj)p(ot|stkj,at)βkTk~(stkj,st(k1)j)p(ot|st(k1)j,at)βk)}p_{accept}=\min\left\{1,\frac{T_{k}(s_{t(k-1)}^{j},s_{tk}^{j})\cdot p(o_{t}|s_{tk}^{j},a_{t})^{\beta_{k}}}{\tilde{T_{k}}(s_{tk}^{j},s_{t(k-1)}^{j})\cdot p(o_{t}|s_{t(k-1)}^{j},a_{t})^{\beta_{k}})}\right\} (13)

Then we update the particle state:

stkj={stkjwith prob. pacceptst(k1)jwith prob. 1paccepts_{tk}^{j}=\begin{cases}s_{tk}^{j}\text{with prob. }p_{\text{accept}}\\ s_{t(k-1)}^{j}\text{with prob. }1-p_{\text{accept}}\end{cases} (14)

The iterative process continues until either the particle approximation converges to the optimal posterior distribution, or the variance of the particle weights becomes less than or equal to the threshold rr^{*}.

Experiments

Table 1: Performance Comparison
LD (α=0.5\alpha=0.5) LD (α=1.0\alpha=1.0) Tag Laser Tag RS(11, 11) RS(15,15)
|𝒮||\mathcal{S}| \infty \infty 870 4830 247,808 7,372,800
|𝒜||\mathcal{A}| 3 3 5 5 20 16
|𝒪||\mathcal{O}| \infty \infty 30 1.5×106\sim 1.5\times 10^{6} 3 3
POMCP 0.357±0.220.357\pm 0.22 0.691±0.410.691\pm 0.41 18.530±0.40-18.530\pm 0.40 18.112±0.18-18.112\pm 0.18 12.252±0.4412.252\pm 0.44 8.207±0.538.207\pm 0.53
ARDESPOT 1.24±0.25-1.24\pm 0.25 0.760±0.410.760\pm 0.41 13.704±0.71-13.704\pm 0.71 15.982±0.64-15.982\pm 0.64 18.113±0.6418.113\pm 0.64 16.111±0.6616.111\pm 0.66
POMCPOW 1.555±0.211.555\pm 0.21 1.468±0.491.468\pm 0.49 16.783±0.42-16.783\pm 0.42 15.263±0.40-15.263\pm 0.40 13.521±0.6713.521\pm 0.67 9.141±0.449.141\pm 0.44
AdaOPS 2.231±0.132.231\pm 0.13 2.762±0.272.762\pm 0.27 9.920±0.70-9.920\pm 0.70 13.427±0.67-13.427\pm 0.67 21.904±0.5921.904\pm 0.59 19.091±0.5719.091\pm 0.57
AIROAS 2.303±0.43\mathbf{2.303\pm 0.43} 3.102±0.36\mathbf{3.102\pm 0.36} 8.451±0.75\mathbf{-8.451\pm 0.75} 12.931±1.22\mathbf{-12.931\pm 1.22} 22.872±0.51\mathbf{22.872\pm 0.51} 20.246±0.50\mathbf{20.246\pm 0.50}
  • Note: LD(α\alpha) stands for Light Dark with step size α\alpha. RS(n,m) stands for the Rock Sample with n×nn\times n map and mm rocks. \infty means continuous state (or observation) space. The results represent the average discounted return and its standard error of mean (SEM) for each method and each domain. Higher discounter return indicates better performance.

In this section, we evaluate our method on several domains and conduct an ablation study to demonstrate the contribution of each component.

Baseline Approaches

We evaluate our approach against four state-of-the-art baselines: POMCP (Silver and Veness 2010), ARDESPOT (Ye et al. 2017), POMCPOW (Sunberg and Kochenderfer 2018), and AdaOPS (Wu et al. 2021) on four domains: Light Dark (LD), Tag, Laser Tag, and Rock Sample (RS) (Platt et al. 2010; Egorov et al. 2017; Smith and Simmons 2004). Detailed descriptions of these domains are provided in section A.1. For ARDESPOT and AdaOPS, we employ different initialization strategies: in Light Dark, Tag, and Laser Tag domains, we use domain-specific independent bounds based on maximum and minimum achievable discounted rewards, while for RockSample, bounds are initialized using heuristics. For POMCPOW, we optimize the maxUCB parameter—which governs action selection at each node—by testing values in {1.0, 10.0, 20.0} and selecting the best-performing configuration for each domain. The performance comparison across all domains is presented in table 1. Our experimental results for the baselines may differ from previously published results due to modifications in implementation code, differences in computing infrastructure, and alternative bound initialization methods. Specifically, while (Wu et al. 2021) uses heuristics for bound initialization in Light Dark, Tag, and Laser Tag, we initialize bounds based on maximum and minimum achievable discounted rewards. However, we maintain consistent initialization methods across all approaches within each domain to ensure fair comparison. Complete details of hyperparameter selection for each domain are provided in appendix B.

Experiment Settings

The main configurations for all baselines are detailed in Baseline Approaches. For fair comparison, we ensure AIROAS’s maximum allowed time per decision matches those of comparable baselines with similar parameters. While POMCP and POMCPOW use iterations rather than time limits, we maintain consistency by keeping the number of iterations identical during testing. All these parameters are documented in appendix B. AIROAS uses the same bound initialization methods as ARDESPOT and AdaOPS. For AIROAS-specific configurations, we generate 100 tempering parameters βk\beta_{k} ranging from 0 to 1 using the sigmoid transformation described in eq. 9. We evaluate AIROAS using different target inefficiency ratios r{2.0,3.0,5.0,10.0}r^{*}\in\{2.0,3.0,5.0,10.0\} and select the best-performing value for each domain. All experiments were conducted on a computer equipped with an Intel(R) Core(TM) i9-14900KS processor with 32 CPUs.

Results

The average discounted return and its standard error of mean (SEM) are presented in table 1. In all these domains, AIROAS outperforms the other solvers with suitable target inefficiency ratio and tempering parameters.

In the Light Dark Problem, we evaluate performance with step sizes α=0.5\alpha=0.5 and α=1.0\alpha=1.0. The state consists of position coordinates and termination status (terminated or not), while observations provide noisy measurements of the current position. For particle state mutation, we generate a Gaussian distribution over the position components, where the variance is proportional to the distance between the state’s position and the received observation. Through this mutation strategy in bridging distributions, combined with weight adjustments, we observe significant mitigation of the sample impoverishment problem, leading to improved performance.

In the Tag problem, the state consists of the robot’s position, target’s position, and tagging status. While the agent’s position is fully observable, the target’s position is only observed when both actors occupy the same cell. For particle state mutation, we generate a Gaussian distribution exclusively over the target’s position, as the agent’s position is known. Given this limited observability structure, we did not anticipate significant performance improvements from Annealed Importance Resampling. The observed improvements may be attributed to the compact state and observation spaces, where the mutation step effectively explores different states within this limited domain.

The Laser Tag problem represents a more challenging variant of the Tag problem. The state space comprises the agent’s position, target’s position, and terminal signal, with neither position directly observable. The observation consists of noisy range readings from laser sensors in eight different directions. For particle state mutation, we construct a multivariate Gaussian distribution over both agent and target positions. Given these high-dimensional observations, which induce multimodal distributions over the state space, we anticipate Annealed Importance Sampling to demonstrate significant advantages in this domain.

In the RockSample problem, the state space comprises the robot’s position and the quality (good or bad) of each rock. With scenarios containing 11 or 15 rocks, this results in an extremely high-dimensional state space with a large number of states. In contrast, the observation structure is simple, providing information about only one rock’s quality at a time (or no observation). For particle state mutation, we focus specifically on mutating the quality of the rock for which the observation provides information, rather than mutating the entire state vector.

Benefits of Annealed Importance Resampling

Refer to caption
Figure 2: Performance comparison between AIROAS variants with and without Annealed Importance Resampling (AIR) on the Light Dark domain (α=1.0\alpha=1.0). The performance gap widens in favor of AIR as the number of particles increases.

We conduct an ablation study to evaluate the impact of Annealed Importance Resampling (AIR) on our approach. Figure fig. 2 presents a comparative analysis between AIROAS with and without AIR on the Light Dark domain, using a step size of α=1.0\alpha=1.0. To ensure a comprehensive evaluation, we vary the number of particles from the set {100,200,500,1000,2000}\{100,200,500,1000,2000\}. The results demonstrate that the performance advantage of incorporating AIR becomes more pronounced as the number of particles increases. This suggests that AIR’s particle state and weight mutation strategy becomes increasingly effective with larger particle populations, while maintaining all other components of the algorithm unchanged.

Conclusion

We presented AIROAS, a novel online POMDP solver that leverages Annealed Importance Resampling to better handle observation uncertainty. By constructing a sequence of bridge distributions between the state-transition and optimal distributions, our approach effectively addresses the particle degeneracy problem common in traditional particle filtering methods. Empirical evaluations across diverse POMDP domains demonstrate AIROAS’s consistent performance advantages over state-of-the-art baselines, with ablation studies confirming that these benefits scale with the number of particles. These findings indicate that AIROAS represents a significant advancement in online POMDP solving, particularly for domains with complex observation spaces or highly informative observations that create sharply peaked posterior distributions.

While our current approach uses sigmoid transformation for generating tempering parameters, future work could explore two key improvements: adaptive scheduling of tempering parameters that dynamically adjusts based on belief uncertainty, and parallel implementation of the annealing process to reduce computational overhead. Additionally, incorporating temporal abstraction techniques would enable AIROAS to efficiently handle high-dimensional continuous action spaces through learned macro-actions, extending its applicability to more complex robotic control and autonomous driving domains (Luo et al. 2025). These enhancements could further improve AIROAS’s efficiency and effectiveness in solving complex POMDPs across both discrete and continuous action spaces.

Acknowledgements

This material is based upon work sponsored by the National Science Foundation (NSF) under Grant CNS-2238815 and by the Defense Advanced Research Projects Agency (DARPA) under the Assured Neuro Symbolic Learning and Reasoning program. Results presented in this paper were obtained using the Chameleon testbed supported by the National Science Foundation. Any opinions, findings, conclusions, or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the NSF or the DARPA.

References

  • Arulampalam et al. (2002) Arulampalam, M. S.; Maskell, S.; Gordon, N. J.; and Clapp, T. 2002. A tutorial on particle filters for online nonlinear/non-Gaussian Bayesian tracking. IEEE Trans. Signal Process., 50: 174–188.
  • Doucet and Johansen (2009) Doucet, A.; and Johansen, A. 2009. A Tutorial on Particle Filtering and Smoothing: Fifteen Years Later. Handbook of Nonlinear Filtering, 12.
  • Egorov et al. (2017) Egorov, M.; Sunberg, Z. N.; Balaban, E.; Wheeler, T. A.; Gupta, J. K.; and Kochenderfer, M. J. 2017. POMDPs.jl: A Framework for Sequential Decision Making under Uncertainty. Journal of Machine Learning Research, 18(26): 1–5.
  • Forthofer, Lee, and Hernandez (2007) Forthofer, R. N.; Lee, E. S.; and Hernandez, M. 2007. 15 - Analysis of Survey Data. In Forthofer, R. N.; Lee, E. S.; and Hernandez, M., eds., Biostatistics (Second Edition), 421–443. San Diego: Academic Press, second edition edition. ISBN 978-0-12-369492-8.
  • Gordon, Salmond, and Smith (1993) Gordon, N. J.; Salmond, D.; and Smith, A. F. M. 1993. Novel approach to nonlinear/non-Gaussian Bayesian state estimation. In IEE Proceedings F - Radar and Signal Processing.
  • Hastings (1970) Hastings, W. K. 1970. Monte Carlo Sampling Methods Using Markov Chains and Their Applications. Biometrika, 57(1): 97–109.
  • Herbst and Schorfheide (2019) Herbst, E.; and Schorfheide, F. 2019. Tempered particle filtering. Journal of Econometrics, 210(1): 26–44. Annals Issue in Honor of John Geweke “Complexity and Big Data in Economics and Finance: Recent Developments from a Bayesian Perspective”.
  • Kaelbling, Littman, and Cassandra (1998) Kaelbling, L. P.; Littman, M. L.; and Cassandra, A. R. 1998. Planning and acting in partially observable stochastic domains. Artificial Intelligence, 101(1): 99–134.
  • Kitagawa (1996) Kitagawa, G. 1996. Monte Carlo Filter and Smoother for Non-Gaussian Nonlinear State Space Models. Journal of Computational and Graphical Statistics, 5(1): 1–25.
  • Kloek and van Dijk (1978) Kloek, T.; and van Dijk, H. K. 1978. Bayesian Estimates of Equation System Parameters: An Application of Integration by Monte Carlo. Econometrica, 46(1): 1–19.
  • Li, Sun, and Sattar (2013) Li, T.; Sun, S.; and Sattar, T. 2013. Adapting sample size in particle filters through KLD‐resampling. Electronics Letters, 49(12): 740–742.
  • Luo et al. (2025) Luo, B.; Pettet, A.; Laszka, A.; Dubey, A.; and Mukhopadhyay, A. 2025. Scalable Decision-Making in Stochastic Environments through Learned Temporal Abstraction. In The Thirteenth International Conference on Learning Representations.
  • M. Johansen (2015) M. Johansen, A. 2015. On Blocks, Tempering and Particle MCMC for Systems Identification. IFAC-PapersOnLine, 48(28): 969–974. 17th IFAC Symposium on System Identification SYSID 2015.
  • Neal (1998) Neal, R. M. 1998. Annealed Importance Sampling. arXiv:physics/9803008.
  • Papadimitriou and Tsitsiklis (1987) Papadimitriou, C. H.; and Tsitsiklis, J. N. 1987. The Complexity of Markov Decision Processes. Mathematics of Operations Research, 12(3): 441–450.
  • Platt et al. (2010) Platt, R.; Tedrake, R.; Kaelbling, L. P.; and Lozano-Perez, T. 2010. Belief space planning assuming maximum likelihood observations. In Robotics: Science and Systems.
  • Rodriguez, Parr, and Koller (1999) Rodriguez, A.; Parr, R.; and Koller, D. 1999. Reinforcement Learning Using Approximate Belief States. In Solla, S.; Leen, T.; and Müller, K., eds., Advances in Neural Information Processing Systems, volume 12. MIT Press.
  • Ross et al. (2008) Ross, S.; Pineau, J.; Paquet, S.; and Chaib-draa, B. 2008. Online Planning Algorithms for POMDPs. Journal of Artificial Intelligence Research, 32: 663–704.
  • Silver and Veness (2010) Silver, D.; and Veness, J. 2010. Monte-Carlo Planning in Large POMDPs. In Lafferty, J.; Williams, C.; Shawe-Taylor, J.; Zemel, R.; and Culotta, A., eds., Advances in Neural Information Processing Systems, volume 23. Curran Associates, Inc.
  • Smith and Simmons (2004) Smith, T.; and Simmons, R. 2004. Heuristic search value iteration for POMDPs. In Proceedings of the 20th Conference on Uncertainty in Artificial Intelligence, UAI ’04, 520–527. Arlington, Virginia, USA: AUAI Press. ISBN 0974903906.
  • Sunberg and Kochenderfer (2018) Sunberg, Z.; and Kochenderfer, M. 2018. Online algorithms for POMDPs with continuous state, action, and observation spaces. arXiv:1709.06196.
  • Wu et al. (2021) Wu, C.; Yang, G.; Zhang, Z.; Yu, Y.; Li, D.; Liu, W.; and Hao, J. 2021. Adaptive Online Packing-guided Search for POMDPs. In Ranzato, M.; Beygelzimer, A.; Dauphin, Y.; Liang, P.; and Vaughan, J. W., eds., Advances in Neural Information Processing Systems, volume 34, 28419–28430. Curran Associates, Inc.
  • Ye et al. (2017) Ye, N.; Somani, A.; Hsu, D.; and Lee, W. S. 2017. DESPOT: Online POMDP Planning with Regularization. Journal of Artificial Intelligence Research, 58: 231–266.
  • Zhang et al. (2024) Zhang, Y.; Luo, B.; Mukhopadhyay, A.; Stojcsics, D.; Elenius, D.; Roy, A.; Jha, S.; Maroti, M.; Koutsoukos, X.; Karsai, G.; and Dubey, A. 2024. Shrinking POMCP: A Framework for Real-Time UAV Search and Rescue. In 2024 International Conference on Assured Autonomy (ICAA), 48–57.

Appendix

Appendix A Experiment

A.1 Domains

We use the following four environments to validate our approach:

  1. 1.

    Light Dark: The Light Dark  (Platt et al. 2010) is a one-dimensional continuous-state problem where an agent must navigate to a target area while dealing with state-dependent observation noise. The key challenge lies in the fact that observations, which are noisy measurements of the agent’s position, become more accurate as the agent moves toward a “light” region and less accurate in “darker” regions. Starting from an initial position on one side of the environment, the agent can move left, right, or declare goal completion. We increase the problem’s complexity by varying the step size α\alpha, where larger steps make precise positioning more challenging while smaller steps increase the number of actions required for information gathering.

  2. 2.

    Tag: The Tag problem  (Egorov et al. 2017) involves an agent that aims to tag an opponent by occupying the same grid cell and executing a tag action. The opponent moves stochastically according to a fixed policy, attempting to move away from the agent with probability pp while staying in the same cell with probability 1p1-p. The state space consists of both agent position and target position and a tag status indicator, while the action space allows the agent to move in four cardinal directions or perform a tag action. The agent can always observe its own position but can only observe the opponent’s position when both are in the same grid cell.

  3. 3.

    Laser Tag: The LaserTag problem  (Egorov et al. 2017), a more challenging variant of Tag, involves an agent attempting to tag an escaping target by occupying the same grid cell and executing a tag action. The state space consists of both agent position and target position and a tag status indicator, while the action space allows the agent to move in four cardinal directions or perform a tag action. Unlike Tag where the agent knows its own position, in LaserTag the agent initially has no knowledge of either its own or the target’s position, and must rely on sensor information from the environment to infer both positions.

  4. 4.

    RockSample: The RockSample problem  (Smith and Simmons 2004) models a robot’s exploration task where it must navigate through a grid environment to collect valuable rocks while avoiding bad ones before reaching an exit area. The key challenge lies in the uncertainty about rock qualities, which can only be determined through noisy sensor readings that decrease in accuracy exponentially with distance from the rock. The state space consists of the robot’s position and the status of rocks (good or bad), while the action space includes four movement actions (up, down, left, right), a sampling action to collect rocks, and KK sensing actions to check rock status. The robot receives perfect information about its position but noisy observations about rock qualities through a sensor with efficiency parameter α\alpha, with no observations during movement or sampling actions.

Appendix B Hyperparameters

Table 2: Hyperparameters Selected
LD(α=0.5\alpha=0.5) LD(α=1.0\alpha=1.0) Tag Laser Tag RS(11, 11) RS(15, 15)
POMCP Iterations 20000 20000 20000 20000 20000 20000
Max_Depth 100 100 100 100 100 100
ARDESPOT K 30 30 300 300 100 100
λ\lambda 0.1 0.1 0.01 0.01 0.0 0.0
T_max 5.0 5.0 5.0 5.0 5.0 5.0
bounds (-11.0, 11.0) (-11.0, 11.0) (-20.0, 0.0) (-20.0, 0.0) (FA, MDP) (FA, MDP)
POMCPOW c 10 10 10 10 10 10
α𝒪\alpha_{\mathcal{O}} 0.03 0.03 0.03 0.03 1.0 1.0
k𝒪k_{\mathcal{O}} 4.0 4.0 4.0 4.0 1.0 1.0
Iterations 20000 20000 20000 20000 20000 20000
Max_Depth 100 100 100 100 100 100
AdaOPS mminm_{min} 10 10 10 10 100 100
δ\delta 1.0 1.0 0.1 0.1 0.1 0.1
T_max 5.0 5.0 5.0 5.0 5.0 5.0
bounds (-11.0, 11.0) (-11.0, 11.0) (-20.0, 0.0) (-20.0, 0.0) (FA, MDP) (FA, MDP)
AIROAS rr^{*} 5.0 3.0 2.0 2.0 10.0 2.0
T_max 5.0 5.0 5.0 5.0 5.0 5.0
bounds (-11.0, 11.0) (-11.0, 11.0) (-20.0, 0.0) (-20.0, 0.0) (FA, MDP) (FA, MDP)
  • Note: LD(α\alpha) stands for Light Dark with step size α\alpha. RS(n,m) stands for the Rock Sample with n×nn\times n map and mm rocks.

  • Note: For bounds, (a,b) represents independent bound that lower value is set to a and upper value is set to b. For RockSample domain, we use heuristics to initialize the lower and upper bound values. FA means Fixed Action Policy and MDP means MDP approximation for upper bound.

The hyperparameters for AIROAS and baseline algorithms are presented in table 2. To ensure fair comparison, we maintain consistent initialization methods for the lower and upper bounds across ARDESPOT, AdaOPS, and AIROAS within each domain. Some of our results differ from previously published findings due to our use of different bound initialization methods in certain domains.

For the RockSample domain, we utilize the heuristics described in (Wu et al. 2021). Specifically, we initialize the lower bound using a Fixed Action Policy, which selects the same action regardless of observations received. The upper bound is initialized using MDP approximation. For the Light Dark, Tag, and Laser Tag domains, we employ domain-specific initialization values that differ from those reported in (Wu et al. 2021). While this leads to some discrepancies with previously reported results, our comparison remains fair as we consistently apply the same initialization method across all approaches within each domain.

For AIROAS evaluation, we tested different target inefficiency ratios (rr^{*}) for each domain, as discussed in the main paper. We explored rr^{*} values in the set {2.0, 5.0, 10.0, 20.0} and report the best-performing results. The optimal rr^{*} value for each domain is documented in table 2.