Learning Sampling Distributions Using Local 3d Workspace Decompositions for Motion Planning in High Dimensions
Abstract
Earlier work has shown that reusing experience from prior motion planning problems can improve the efficiency of similar, future motion planning queries. However, for robots with many degrees-of-freedom, these methods exhibit poor generalization across different environments and often require large datasets that are impractical to gather. We present spark and flame, two experience-based frameworks for sampling-based planning applicable to complex manipulators in 3d environments. Both combine samplers associated with features from a workspace decomposition into a global biased sampling distribution. spark decomposes the environment based on exact geometry while flame is more general, and uses an octree-based decomposition obtained from sensor data. We demonstrate the effectiveness of spark and flame on a real and simulated Fetch robot tasked with challenging pick-and-place manipulation problems. Our approaches can be trained incrementally and significantly improve performance with only a handful of examples, generalizing better over diverse tasks and environments as compared to prior approaches.
I Introduction
For many high-dimensional systems, sampling-based methods have proven very successful at efficient motion planning [1]. However, as noted by [2], they are sensitive to “challenging regions”, such as narrow passages in the search space, which have a low probability of being sampled. The existence of such “challenging regions” hinders the performance of sampling-based planners.
Many methods have improved the performance of sampling-based planners by using experience from prior motion planning problems, e.g., [3, 4]. In particular, experience can be used to learn how to explore the “challenging regions” of the search space. However, for high dimensional robots, learning-based methods typically face difficulties in generalizing even between similar environments [5, 6, 7].
This work presents two experience-based planning frameworks—one using exact geometry (spark) and one using sensor data (flame)—for high dof robots in 3d environments. Both frameworks build on the decomposition paradigm introduced by [5]. That work introduced the idea of local primitives, which capture important local workspace features. Each primitive is associated with a local sampler, which produces samples in “challenging regions” created by the primitive, and are stored in an experience database. The database is updated offline by adding experiences from past planning problems in an incremental manner. Given a planning query, similar local primitives to those in the current workspace are retrieved; their local samplers are combined and used by a sampling-based planner. In such a framework several questions arise on the definition and use of local primitives. [5] required a priori knowledge of local primitives and was designed for 2d environments with circular obstacles.
The contributions of this paper are the following. We present spark which scales to 3d environments when exact geometric information is available and flame which relies only on sensor information. In order to develop these frameworks, we revisit and enrich the framework of [5] by introducing novel local primitives, distance functions, and utilizing efficient retrieval structures. Furthermore, we propose a new criticality test that enables incremental training of spark and flame from novel scenes. Through the synergistic combination of the introduced components, we achieve real-time retrieval and training of both frameworks. Finally we validate our methods on a suite of simulated and real robot experiments with the Fetch robot, an 8 degree-of-freedom (dof) manipulator. Compared to prior work, our methods exhibit significantly better performance over diverse environments and tasks with relatively few examples. The datasets and the implementation of spark and flame are open-sourced 111https://github.com/KavrakiLab/pyre.

II Problem Statement
Consider a robot in a workspace . A robot configuration is a point in the configuration space (-space), . Obstacles in workspace induce -space obstacles where the robot is in collision, , which must be avoided. We are interested in finding a collision-free path, from start to goal as efficiently as possible.
Sampling-based planners are widely used for motion planning [1, 9]. These planners randomly sample to explore the -space. However, the performance of sampling-based planners is degraded when “challenging regions” exist in the -space [2]. These regions, introduced by , create difficult-to-explore areas in the -space due to their “low visibility” relative to the entire -space. It is understood [10, 11] that biased sampling that targets such regions can significantly improve performance. Thus, the problem we consider is how to learn to bias sampling towards “challenging regions” using workspace information and prior experiences.
III Related Work
One way to bias sampling is with rejection sampling, which draws uniform samples but only accepts them if a geometric test is passed, e.g., bridge sampling [2] or Gaussian sampling [12]. However, rejection sampling is inefficient, possibly drawing many samples before a valid one is found. Some approaches attempt to directly sample the “challenging regions,” such as medial axis sampling [13] or free-space dilation [14], but these distributions become difficult to compute as the state space increases in dimension.
Rather than using fixed distributions, many methods [15, 16] adapt sampling online based on collision information, such as utility-guided sampling [17], toggle-prm [18] or hybrid-prm [19]. However, these methods do not transfer knowledge between planning queries. Our methods bias based on prior motion plans, and thus is complementary to adaptive sampling methods.
Another category of techniques improves planning by learning problem invariants. Examples include using sparse roadmaps [3], informed lazy evaluations [20], path databases [21, 22, 23], or biased distributions [24]. However, these methods do not use workspace information. Thus, changes in the environment force these methods to repair now invalidated information, reducing performance.
On the other hand, many methods do use workspace features to guide sampling—typically, exact [25, 26] or inexact [27] workspace decompositions are used. For example, the authors of [28] use balls to decompose the workspace. However, to generate samples in “challenging regions,” these approaches require a mapping from workspace to -space, reliant on the robot’s inverse kinematics. Thus, these methods are typically limited to free-flying or low-dimensional robots, as manipulator kinematics can become complex. The proposed methods also use workspace decompositions (both an exact and inexact one), but instead learn the inverse mapping from workspace to -space.
Many recent techniques use deep learning to learn the mapping from workspace to -space, such as using neural networks to bias sampling [4, 29, 6, 30] or to guide planners [31, 32, 33]). However, due to the inherent discontinuity of motion planning [34, 35], this mapping is generally discontinuous, limiting the applicability of these methods [7, 5, 36]. That is, small changes in the workspace result in substantial (possibly discontinuous) changes in -space. Thus, these approaches usually apply only to manipulators with low workspace variance [4, 29, 31, 30].
On the other hand, a category of methods that addresses the discontinuity problem, are mixture model methods and similarity-based methods. Mixture model methods [7, 37] learn different functions for different discontinuous regions but are limited by the fixed number of mixtures. Similarity-based methods [38, 39, 5], which are most similar to ours, are usually non-parametric techniques that retrieve relevant information (e.g., “challenging regions” representations) based on workspace similarity. Retrieval is by design discontinuity-sensitive since multiple different “challenging regions” can be retrieved for similar workspaces. However, [38] is limited to free flying robots [39] addresses trajectory optimization-based problems rather than sampling and [5] is limited to 2d planar environments with circular objects. Our work shares the retrieval idea with these methods but applies to general 3d workspaces with realistic robots.
IV Methodology
We introduce two experience-based planning frameworks for arbitrary 3d environments and manipulators, spark and flame. spark requires an exact representation of the environment’s geometry, while flame only uses sensor information represented as an octomap [8].
Consider the Fetch robot reaching into a deep shelf, as shown in Fig. 1. The arrangement of the workspace obstacles constricts the manipulator’s movement, making the motion planning problem challenging. Both frameworks decompose the workspace into local primitives to capture local workspace features that create “challenging regions” in -space, e.g., the shelf’s sides that create a narrow region for reaching. Our frameworks learn by using a criticality test, which associates “critical” configurations from every successful motion plan to relevant local primitives in the current workspace. For example, the criticality test should capture that configurations of the Fetch’s arm deep in the shelf, should be associated with the sides of the shelf. These configurations are used to create a local sampler—a biased sampling distribution focused on “challenging regions” created by the associated local primitive. During learning, these primitive-sampler pairs are stored in a spatial database (Alg. 1). When inferring for a new problem, relevant local primitive-sampler pairs are retrieved from the database; local samplers are then combined into a global biased sampler (Alg. 2).
IV-A Overall Framework
Both spark and flame define their own decomposition of the workspace into local primitives, a criticality test for these primitives, and a metric to compare local primitives. The database that stores local primitives, their associated local samplers, and the synthesis of the global sampler is the same in spark and flame.
IV-A1 Local Primitives
Local primitives are workspace features that come from any valid decomposition of the workspace. Each local primitive attempts to capture the workspace features that create “challenging regions” in the -space. The definition of local primitives is a key component of the algorithm. The local primitives for spark and flame are given in their respective sections, and are shown in Fig. 2 and Fig. 3. We denote local primitives with . From a given workspace , local primitives are created from a workspace decomposition such that (Alg. 1 in Alg. 1, Alg. 2 in Alg. 2).
IV-A2 Local Samplers
Each local primitive is associated with a local sampler, which should produce samples in the “challenging regions” created by the associated local primitive. A local sampler is a generative distribution over a set of “critical” configurations gathered from past experience. Due to the potential complexity of this distribution and its natural multi-modality, we represent the local sampler as a Gaussian mixture model (gmm), similar to [24]. Contrary to [24], we do not use expectation maximization to calculate the parameters of the gmm. Instead, we place one mixture for each configuration in the local sampler with a fixed covariance (Alg. 1 in Alg. 1):
We choose uniforms weights—all mixtures are equiprobable.
IV-A3 Criticality Test
In 3d local primitives may or may not create “challenging regions,” and finding if a configuration belongs to such a region is not straightforward. The authors of [40] chose regions with maximum path density while [41] and [6] identified “challenging regions” using graph properties of a roadmap. However in our setting only finding these regions is not enough, since they must also correspond to a local primitive. To overcome this problem, and depart from earlier work, we introduce a criticality test (isCritical, Alg. 1 in Alg. 1), which associates key configurations from solution paths to relevant local primitives. The intuition is that a configuration in a critical region will be close to a -space obstacle, which often means that some part of the robot will be close to a workspace obstacle. Both spark and flame define their own criticality test, which enables learning incrementally without a priori knowledge.
IV-A4 Distance Function
Given a new motion planning problem, the workspace is decomposed into a set of local primitives. Based on a distance metric over local primitives , we search for similar local primitives in the existing database and retrieve their associated local samplers (Alg. 2 in Alg. 2). Both spark and flame define a distance metric over their local primitives.
Quick retrieval of relevant local primitives is essential to our method. We use gnat [42], a data structure with logarithmic scaling for nearest-neighbor retrieval in high-dimensional metric spaces. gnat also enables fast rejection of queries that do not exist within the database, which is useful as many local primitives are irrelevant e.g., outside the reachable workspace.
IV-A5 Global Sampling
Given a set of retrieved local samplers, they are combined in a composite gmm (Alg. 2 in Alg. 2):
To avoid redundancy as local samplers may be the same, a local sampler is used only if its unique. To retain probabilistic completeness, there is a chance to sample uniformly rather than from the gmm. If no local samplers are retrieved the biased sampler defaults to uniform sampling, i.e., .
IV-B spark

IV-B1 Local Primitives
In spark, local primitives are defined as pairs of box obstacles and , which have the following parameters:
where respectively denote the pose and size—a 3d vector for the scale of each dimension—of each box, for a total of 10 parameters per box. Non-box objects are approximated by bounding boxes. Boxes that are too far from each other (see Sec. IV-B3) are not considered valid local primitives. An example of pairs of local primitives can be seen in Fig. 2.
Intuitively, pairs of objects create “challenging regions” in the configuration space, as they constrict space in the workspace. Note that one or many (e.g., 3) objects per local primitive are also viable options. However, single objects are less discriminative than two and thus recall extraneous information. In a similar vein, considering many objects combinatorially increases the number of primitives in a scene. As we show in Sec. V, using pairs of objects provides good results in our tested scenarios.
IV-B2 Criticality Test
To check if a configuration is associated with a local primitive, we find the minimum distance between the robot at that configuration with the pair of obstacles. For every shape in the robot’s geometry, the distance to both objects in the local primitive is measured with the Gilbert-Johnson-Keerthi (GJK) algorithm [43]. If the minimum distance is smaller than a threshold , that configuration is related to that local primitive.
IV-B3 Distance
To measure the distance between poses in , we use [44]:
where respectively denote translation and orientation (a quaternion) of a pose , is a chosen weight, is the Euclidean norm and is the inner product. To measure distance between boxes, we use a sum of metrics:
where is a chosen weight. Local primitives for spark are all the pairs of boxes with distance less than . To compare local primitives, we use a metric that is the minimum distance between the boxes in each primitive:
IV-C flame

IV-C1 Local Primitives
In flame, local primitives are defined as “local” occupancy grids, or octoboxes, which are obtained from a global octree. An octree is a hierarchical volumetric tree that efficiently encodes obstacles in the workspace. At each level, the octree contains voxels—descending a level of the tree, these voxels are split in half along each of its axes, each giving 8 subvoxels. A voxel contains occupancy information at its resolution; a voxel is either occupied or free. Voxels at higher levels can compact information from lower levels—if all subvoxels are either occupied or free, then the tree does not need to descend further. flame is useful for application on real robots as octrees [8] are a common approach used to represent sensor data.
We define octoboxes as occupancy grids that correspond to intermediate nodes of an octree. Here, octoboxes contain the voxels of the two lowest levels of the octree, that is, 64 voxels as a 3d occupancy grid.
To decompose the workspace, an octree is built relative to the robot that captures the occupancy of the workspace. Non-empty octoboxes are quickly generated by traversing the octree at a specific depth. We also consider the pose of the center of the octobox in the global frame. An example is shown in Fig. 3. Local primitives in flame are , where is the pose of the octobox and is a -bit occupancy grid vector, where each bit corresponds to an occupied () or free () voxel.
IV-C2 Criticality Test
A configuration is associated with a local primitive if the robot’s geometry at configuration intersects the bounding box of the given octobox. Note that a configuration may be associated with many local primitives.
IV-C3 Distance
To compare two octoboxes, we simply check if the two octoboxes are the same. It may seem overly pessimistic to look for precise octobox matches, but our empirical results show that this simple metric yields good results. One possible explanation is, since the criticality test assigns a configuration to many different local primitives, only one primitive needs to be recovered to recall the relevant configuration. The distance metric is:
V Experiments
We evaluate the performance of spark and flame in the following varied and realistic scenarios using a Fetch (8-dof) robot manipulator. We believe both frameworks can also be used with mobile/flying robots, but focus on manipulators since we consider them more challenging for learning methods. The “Small Shelf” environment (Fig. 4a) with 3 cylindrical objects, the “Large Shelf” environment with 7-9 cylindrical objects (Fig. 6a), the “Real Shelf” (Fig. 1) with 3 box objects and the ”Box Table” with mutliple random objects (Fig. 5a). In each environment, we vary both the pose of “small” objects (cylinders, boxes) and more importantly the pose of “big” objects(shelves, box, table) relative to the robot. The ”big” objects variations of ”Small shelf” are shown in Fig. 4a) and described in Fig. 1a), Fig. 5a), and Fig. 6a) for the other scenarios. See the attached video for more details.
We consider these variations of environments highly challenging for learning-based motion planning frameworks for two reasons. First, the underlying motion planning problem is hard since it exhibits many “challenging regions.” Second, even small changes in the relative position of the objects can cause discontinuous changes in the configuration space obstacles, making it hard to learn functions that map from workspace to -space [35, 7]. For example, a small change in the relative angle of the shelf to the robot might require a solution that goes on the left side of the robot, versus the right side. Such diversity is realistic, as mobile manipulators can approach objects from many different angles and positions.
We test two tasks: a “pick” and a “place” task. In the “pick” task, the Fetch plans between its stow position and the farthest object on the shelf. In the “place” task in the “Large Shelf” environment, the Fetch starts rigidly grasping the object farthest back on the top or bottom shelf and places the object in the back of the middle shelf. In the “place” task in the “Box Table” environment, the Fetch starts grasping the cube inside the box and place randomly on the table.
We implemented our framework using the Open Motion Planning Library (ompl) [45] and Robowflex [46]. Note that spark and flame both generate sampling distributions and must be used by a sampling-based planner. For all experiments and frameworks, rrt-Connect [47] is used with default settings unless noted otherwise. All the experiments were run on an Intel® i7-6700 with four 4.00GHz cores and 32 gb of memory. The hyperparameters for spark (7) and flame (3) are given in Table I and were the same for all experiments. These are reasonable defaults for our methods, found heuristically. The methods performance was empirically observed to be robust over a wide range of values.
Parameter | Symbol | Value |
---|---|---|
Sampling Variance | 0.2 | |
Sampling Bias | 0.5 | |
Retrieval Radius | 0.4 | |
spark Clustering Distance | 0.15 | |
spark Pair Distance | 0.2 | |
spark Pose Weight | 0.75 | |
spark Size Weight | 0.5 |
V-A Generalization
Fig. 4b presents timing results for a “pick” task in “Small Shelf” over three types of variation, , , and as shown in Fig. 4a. We compare against rrt-Connect, an experience-based approach that uses a path database (thunder) [3], and a conditional variational autoencoder framework (cvae) [4]. thunder stores previous paths as a sparse roadmap—new problems are solved by searching the roadmap for a valid path; if no path is found (possibly due to changes in the environment), a planner is used to repair the roadmap or find a new path from scratch. The ompl version of thunder was used within Robowflex.
cvae learns a mapping from workspace to “interesting samples” in -space. To infer new samples the latent space of the cvae can be sampled and decoded into -space, providing promising samples to a sampling-based planner (here, rrt-Connect is used). We use the provided implementation of cvae [4]. We test two variants of cvae: “cvae w/o Task”, which has the same input as spark albeit with a fixed number of obstacles, and “cvae”, which includes the start and goal as recommended in [4]. The latent space dimension was increased as proposed by [6] for our high dof problems. The same training data (500 paths) were given to cvae as spark which is approximately 5000 training samples.
All methods show marked improvement when the variance is low. However, only spark, flame, and cvae retain performance as the axes of variation increases (Fig. 4b). As environment variance grows, thunder’s sparse roadmap becomes less informative and falls back to planning from scratch. Similarly, the performance of “cvae w/o Task” degrades as diversity increases (possibly due to discontinuities). cvae greatly benefits from the addition of the start and goal, indicating it is highly informative for this task. Increasing the size of training examples did not improve the performance of cvae Finally, note that spark and flame have lower variance in solution time, showing both become more effective at solving the “harder” problems in the test set.

V-B Convergence and Scalability

Fig. 4c, Fig. 5b and Fig. 6b present convergence results for spark and flame for the “pick” task in the “Small Shelf” and the place task for “Large Shelf,”“Box table” environments. spark and flame both achieve significant improvement in average performance given only a few examples and quickly converge ( for “Small Shelf,” “ Box Table” and for “Large Shelf”). spark has better asymptotic performance (especially in ”Box Table”) than flame. This is expected, as flame uses the less informative but more general decomposition. Fig. 5c shows the database retrieval time versus the size of the dataset using naïve linear search and gnat-based search. The number of local primitives for 1000 training examples is for spark and for flame. gnat scales much better than naive linear search. Although the retrieval overhead is negligible, we expect that it would affect performance if the database grows too large.

V-C Transfer
Fig. 6 presents transfer learning results for spark and flame on a “place” task in the “Large Shelf” environment. Here, we show how spark and flame can learn from examples in one task and apply their experience to a different, harder task. In Fig. 6, a baseline of spark and flame trained and tested both on “place” tasks (labeled as spark and flame) is compared to spark and flame trained on “pick” tasks and tested on the “place” task (labeled as spark (TX) and flame (TX)). Note that the “place” task changes the robot’s geometry by attaching an object in the end-effector’s grasp, making the planning problem harder. Even when using experience gathered on a different, easier motion planning problem both spark and flame demonstrate significant performance gains, successfully transferring their experience. We hypothesize that flame shows better performance in transfer learning as it uses an approximate decomposition, and thus can find more relevant experience than spark.
V-D Real Robot
Fig. 1 shows timing results for spark and flame applied to a real robot system. To increase the statistical significance of our results, we use 6-fold cross-validation. Additionally, we carefully tuned the range of rrt-Connect to obtain the best results possible for all methods (range was used). A Vicon camera system was used to gather object poses for spark. The Fetch’s depth camera was used to build an octomap [8] for flame. Note that flame can retrieve relevant experiences even from partial/noisy octomaps which is often the case for real systems. Both demonstrate significant improvement even with few given examples.
VI Conclusion
We have presented two experience-based planning frameworks for arbitrary manipulators in 3d environments, instantiated in spark and flame. spark uses an exact decomposition of the workspace into pairs of obstacles and achieves the best performance when this information is available. flame uses an approximate decomposition from a volumetric octree generated from sensor data and is comparable in performance with spark. We demonstrated that, with only a handful of examples, both approaches confer significant improvement on planning time on a Fetch robot (8 dof), on challenging manipulation tasks in environments that substantially vary, outperforming other learning-based approaches. One limitation of our approaches is that due to our conservative metrics between local primitives, it is hard to generalize to environments that are far outside of the training distribution. For example, the “Large Shelf” and “Real Shelf” share no local primitives (due to the height of the shelves) and thus do not transfer from one to the other. In the future, we will investigate more diverse environments and learning local primitive representations, metrics, and criticality tests.
References
- [1] H. M. Choset, S. Hutchinson, K. M. Lynch, G. Kantor, W. Burgard, L. E. Kavraki, and S. Thrun, Principles of Robot Motion: Theory, Algorithms, and Implementation. MIT Press, 2005.
- [2] D. Hsu, T. Jiang, J. Reif, and Z. Sun, “The bridge test for sampling narrow passages with probabilistic roadmap planners,” IEEE Int. Conf. Robot. Autom., vol. 3, pp. 4420–4426, 2003.
- [3] D. Coleman, I. A. Sucan, M. Moll, K. Okada, and N. Correll, “Experience-based planning with sparse roadmap spanners,” in IEEE Int. Conf. Robot. Autom., 2015, pp. 900–905.
- [4] B. Ichter, J. Harrison, and M. Pavone, “Learning sampling distributions for robot motion planning,” in IEEE Int. Conf. Robot. Autom., May 2018, pp. 7087–7094.
- [5] C. Chamzas, A. Shrivastava, and L. E. Kavraki, “Using local experiences for global motion planning,” in IEEE Int. Conf. Robot. Autom., May 2019, pp. 8606–8612.
- [6] R. Kumar, A. Mandalika, S. Choudhury, and S. S. Srinivasa, “LEGO: Leveraging experience in roadmap generation for sampling-based planning,” IEEE/RSJ Int. Conf. on Intell. Robots and Syst., 2019.
- [7] G. Tang and K. Hauser, “Discontinuity-sensitive optimal control learning by mixture of experts,” in IEEE Int. Conf. Robot. Autom., 2019, pp. 7892–7898.
- [8] A. Hornung, K. M. Wurm, M. Bennewitz, C. Stachniss, and W. Burgard, “OctoMap: An efficient probabilistic 3D mapping framework based on octrees,” Autonomous Robots, vol. 34, no. 3, pp. 189–206, 2013.
- [9] S. M. LaValle, Planning Algorithms. Cambridge Univ. Press, 2006.
- [10] D. Hsu, J.-C. Latombe, and R. Motwani, “Path planning in expansive configuration spaces,” Int. J. of Computational Geometry and Applications, vol. 9, no. 4n5, pp. 495–512, 1999.
- [11] D. Hsu, J.-C. Latombe, and H. Kurniawati, “On the probabilistic foundations of probabilistic roadmap planning,” Int. J. of Robotics Research, vol. 25, no. 7, pp. 627–643, 2006.
- [12] V. Boor, M. H. Overmars, and a. V. D. Stappen, “The gaussian sampling strategy for probabilistic roadmap planners,” in IEEE Int. Conf. Robot. Autom., vol. 2, May 1999, pp. 1018–1023.
- [13] J.-M. Lien, S. L. Thomas, and N. M. Amato, “A general framework for sampling on the medial axis of the free space,” in IEEE Int. Conf. Robot. Autom., vol. 3, 2003, pp. 4439–4444.
- [14] D. Hsu, L. E. Kavraki, J.-C. Latombe, R. Motwani, and S. Sorkin, “On finding narrow passages with probabilistic roadmap planners,” in Int. Wksp. on the Algorithmic Foundations of Robotics. Springer, 1998.
- [15] M. Morales, L. Tapia, R. Pearce, S. Rodriguez, and N. M. Amato, “A machine learning approach for feature-sensitive motion planning,” in Algorithmic Foundations of Robotics VI. Springer, pp. 361–376.
- [16] L. Tapia, S. Thomas, B. Boyd, and N. M. Amato, “An unsupervised adaptive strategy for constructing probabilistic roadmaps,” in IEEE Int. Conf. Robot. Autom. IEEE, 2009, pp. 4037–4044.
- [17] B. Burns and O. Brock, “Toward optimal configuration space sampling,” in Robotics: Science and Syst., 2005, pp. 105–112.
- [18] J. Denny and N. M. Amato, “Toggle PRM: A coordinated mapping of C-Free and C-Obstacle in arbitrary dimension,” in Springer Tracts in Advanced Robot., vol. 86, 2013, pp. 297–312.
- [19] D. Hsu, G. Sánchez-Ante, and Z. Sun, “Hybrid prm sampling with a cost-sensitive adaptive strategy,” in IEEE Int. Conf. Robot. Autom. IEEE, 2005, pp. 3874–3880.
- [20] M. Bhardwaj, S. Choudhury, B. Boots, and S. Srinivasa, “Leveraging experience in lazy search,” in Robotics: Science and Syst., 2019.
- [21] D. Berenson, P. Abbeel, and K. Goldberg, “A robot path planning framework that learns from experience,” in IEEE Int. Conf. Robot. Autom., 2012, pp. 3671–3678.
- [22] M. Phillips, B. J. Cohen, S. Chitta, and M. Likhachev, “E-Graphs: Bootstrapping planning with experience graphs,” in Robotics: Science and Syst., 2012.
- [23] T. S. Lembono, A. Paolillo, E. Pignat, and S. Calinon, “Memory of motion for warm-starting trajectory optimization,” vol. 5, no. 2, pp. 2594–2601, 2020.
- [24] P. Lehner and A. Albu-Schaeffer, “The repetition roadmap for repetitive constrained motion planning,” IEEE Robot. Autom. Letters, vol. 3, no. 3, pp. 3884–3891, 2018.
- [25] H. Kurniawati and D. Hsu, “Workspace importance sampling for probabilistic roadmap planning,” in IEEE/RSJ Int. Conf. on Intell. Robots and Syst., vol. 2, 2004, pp. 1618–1623.
- [26] ——, “Workspace-based connectivity oracle: An adaptive sampling strategy for prm planning,” in Int. Wksp. on the Algorithmic Foundations of Robotics. Springer, 2008, pp. 35–51.
- [27] M. Zucker, J. Kuffner, and J. A. Bagnell, “Adaptive workspace biasing for sampling-based planners,” IEEE Int. Conf. Robot. Autom., pp. 3757–3762, 2008.
- [28] O. Brock and L. E. Kavraki, “Decomposition-based motion planning: A framework for real-time motion planning in high-dimensional configuration spaces,” in IEEE Int. Conf. Robot. Autom., vol. 2, 2001, pp. 1469–1474.
- [29] C. Zhang, J. Huh, and D. D. Lee, “Learning implicit sampling distributions for motion planning,” in IEEE/RSJ Int. Conf. on Intell. Robots and Syst., 2018, pp. 3654–3661.
- [30] B. Chen, B. Dai, Q. Lin, G. Ye, H. Liu, and L. Song, “Learning to plan in high dimensions via neural exploration-exploitation trees,” in Int. Conf. on Learn. Repr., 2020.
- [31] A. H. Qureshi, A. Simeonov, M. J. Bency, and M. C. Yip, “Motion planning networks,” in IEEE Int. Conf. Robot. Autom., 2019, pp. 2118–2124.
- [32] T. Jurgenson and A. Tamar, “Harnessing reinforcement learning for neural motion planning,” in Robotics: Science and Syst., 2019.
- [33] R. Terasawa, Y. Ariki, T. Narihira, T. Tsuboi, and K. Nagasaka, “3d-cnn based heuristic guided task-space planner for faster motion planning,” in IEEE Int. Conf. Robot. Autom. IEEE, 2020, pp. 9548–9554.
- [34] K. Hauser and S. Emmons, “Global redundancy resolution via continuous pseudoinversion of the forward kinematic map,” IEEE Trans. on Autom. Science and Eng., vol. 15, no. 3, pp. 932–944, 2018.
- [35] M. Farber, “Topological complexity of motion planning,” Discrete and Computational Geometry, vol. 29, no. 2, pp. 211–221, 2003.
- [36] W. Merkt, V. Ivan, T. Dinev, I. Havoutis, and S. Vijayakumar, “Memory clustering using persistent homology for multimodality-and discontinuity-sensitive learning of optimal control warm-starts,” arXiv preprint arXiv:2010.01024, 2020.
- [37] S. Finney, L. P. Kaelbling, and T. Lozano-Pérez, “Predicting partial paths from planning problem parameters.” in Robotics: Science and Syst., 2007.
- [38] J.-M. Lien and Y. Lu, “Planning motion in environments with similar obstacles,” Robotics: Science and Syst., 2009.
- [39] N. Jetchev and M. Toussaint, “Fast motion planning from experience: trajectory prediction for speeding up movement generation,” IEEE J. Robot. Autom., vol. 34, no. 2, pp. 111–127, 2013.
- [40] D. Molina, K. Kumar, and S. Srivastava, “Learn and link: Learning critical regions for efficient planning,” in IEEE Int. Conf. Robot. Autom., 2020.
- [41] B. Ichter, E. Schmerling, T.-W. E. Lee, and A. Faust, “Learned critical probabilistic roadmaps for robotic motion planning,” in IEEE Int. Conf. Robot. Autom. IEEE, 2020, pp. 9535–9541.
- [42] S. Brin, “Near neighbor search in large metric spaces,” in Int. Conf. on Very Large Data Bases, 1995, pp. 574–584.
- [43] E. G. Gilbert, D. W. Johnson, and S. S. Keerthi, “A fast procedure for computing the distance between complex objects in three-dimensional space,” IEEE J. Robot. Autom., vol. 4, no. 2, pp. 193–203, 1988.
- [44] J. J. Kuffner, “Effective sampling and distance metrics for 3d rigid body path planning,” in IEEE Int. Conf. Robot. Autom., vol. 4, 2004, pp. 3993–3998.
- [45] I. A. Şucan, M. Moll, and L. E. Kavraki, “The Open Motion Planning Library,” IEEE Robot. Autom. Magazine, vol. 19, no. 4, pp. 72–82, 2012.
- [46] Z. Kingston and L. E. Kavraki, “Robowflex: Robot motion planning with MoveIt made easy,” IEEE Robotics and Automation Letters, 2021, under Review. [Online]. Available: https://arxiv.org/abs/2103.12826
- [47] J. J. Kuffner and S. M. LaValle, “RRT-Connect: An efficient approach to single-query path planning,” in IEEE Int. Conf. Robot. Autom., vol. 2, 2000, pp. 995–1001.