This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Learning Diverse and Physically Feasible
Dexterous Grasps with Generative Model and
Bilevel Optimization

Albert Wu
Computer Science Department
Stanford University, United States
[email protected]
&Michelle Guo
Computer Science Department
Stanford University, United States
[email protected]
&C. Karen Liu
Computer Science Department
Stanford University, United States
[email protected]
Abstract

To fully utilize the versatility of a multi-fingered dexterous robotic hand for executing diverse object grasps, one must consider the rich physical constraints introduced by hand-object interaction and object geometry. We propose an integrative approach of combining a generative model and a bilevel optimization (BO) to plan diverse grasp configurations on novel objects. First, a conditional variational autoencoder trained on merely six YCB objects predicts the finger placement directly from the object point cloud. The prediction is then used to seed a nonconvex BO that solves for a grasp configuration under collision, reachability, wrench closure, and friction constraints. Our method achieved an 86.7%86.7\% success over 120120 real world grasping trials on 2020 household objects, including unseen and challenging geometries. Through quantitative empirical evaluations, we confirm that grasp configurations produced by our pipeline are indeed guaranteed to satisfy kinematic and dynamic constraints. A video summary of our results is available at youtu.be/9DTrImbN99I.

Keywords: Dexterous grasping, Grasp planning, Bilevel optimization, Generative model

1 Introduction

Performing diverse grasps on a variety of objects is a fundamental skill in robotic manipulation. Diverse grasp configurations allow flexible interaction with the objects while satisfying requirements demanded by the higher level task of interest. Dexterous grasping, which refers to object grasping with a fully actuated, multi-finger dexterous robotic hand, has the potential to achieve diverse grasp configurations with applicability to a wide range of objects. This is in contrast with simple grasps achieved by parallel jaw grippers or underactuated multi-finger grippers, both of which have fingertip workspaces restricted to a subspace of the 3D task space.

We identify two major challenges in planning diverse dexterous grasps. Firstly, the solution space is multimodal due to the many finger placement possibilities and the lack of a metric defining an optimal grasp among valid answers. Secondly, each valid grasp is governed by nonconvex physical constraints such as collision, contact, and force closure.

The multimodality of dexterous grasp planning motivates the use of deep learning-based approaches. It is challenging to produce diverse and multimodal grasp plans with regression or naive supervised methods. Analytical approaches such as precomputing a grasp library or sampling-based grasp planning, while popular for planning simple grasps, are computationally intractable on high-dimensional dexterous grasps. Nevertheless, learning to plan grasps that strictly obey physical constraints is difficult. Model-based optimization has historically been applied to enforce exact constraints. However, in practice the dexterous grasp planning problem is arguably too nonconvex to solve directly. Consequently, previous works usually relax the constraints as scalar losses and solve the relaxed problem, sacrificing exact constraint satisfaction guarantees for practicality.

We propose an integrative approach that combines learning and optimization to produce diverse physically-feasible dexterous grasp configurations for unseen objects. Our method first predicts an initial set of contact points using a conditional variational autoencoder (CVAE). The contact points are then projected onto the manifold of kinematically and dynamically feasible grasps by solving a bilevel optimization (BO) problem. Our key contributions are summarized below:
\bullet Learning-based dexterous grasp planning pipeline that integrates CVAE and BO to produce diverse, fully-specified, and constraint-satisfying dexterous grasps from point clouds.
\bullet Bilevel grasp optimization formulation that takes a learning-based grasp prediction and outputs a dexterous grasp that satisfies reachability, collision, wrench closure, and friction cone constraints.
\bullet Successful and extensive hardware validation on 20 household objects. Our method achieved an overall success rate of 86.7%86.7\% over 120120 grasp trials.

Refer to caption
Figure 1: Our method can pick up different objects shapes with a diverse set of grasp configurations.

2 Related Work

We limit our discussion to literature on grasping in uncluttered scenes, with an emphasis on dexterous grasping. We exclude work that focuses on object segmentation in cluttered environments or other types of manipulation, such as in-hand manipulation.

2.1 Dexterous Grasping

Dexterous grasping is a long standing problem in the robotics community. In general, the literature can be split into learning-based and analytical methods. Early learning-based grasp planners seek to fit the space of possible grasps for rapid grasp generation (e.g., [1]). More recent papers have shifted to producing grasps directly with complex model architectures, such as generative models [2, 3, 4] and convolutional neural networks [5, 6]. However, most of these works do not account for physical laws and have restricted, or even missing, hardware evaluation. This issue is exacerbated by the difficulty of simulating contact-rich interactions [7, 8]. The quality of grasps produced by learning exclusively with simulation is questionable. Moreover, many papers do not emphasize the ability to learn diverse grasps, thus forgoing a key benefit of dexterous hands over simple grippers. Notably, some publications on learning dexterous grasping originate from computer vision and graphics communities (e.g., [9, 10, 11, 12, 13, 14]). While these papers cover diverse grasp generation on fully dexterous hands, their ultimate objective tends to be achieving photorealistic human grasps in simulation rather than satisfying strict real-world physical constraints.

On the other hand, analytical dexterous grasp planners are often inspired by physical constraints such as force closure, friction cone, robustness to external disturbance wrenches, and object contact (e.g., [15, 16, 17, 18]). Historically, analytical methods are standalone and includes both grasp modality exploration and physical constraint compliance. We point the readers to [19, 20] for a detailed review on these methods. Some limitations of these approaches include dependency on a high-fidelity object model and lack of diversity in the generated grasps [20].

In recent years, some learning-based approaches have adopted a “grasp refinement” step motivated by analytical metrics [4, 18, 21]. Nevertheless, these metrics tend to be formulated as “relaxed” optimizations to keep the problem tractable. Instead of enforcing the constraints directly, the violation of each constraint is cast as a scalar loss and summed together. This results in an unconstrained optimization which is significantly easier to solve. However, there is no guarantee that the optimization output will indeed satisfy the constraints that motivated the loss design.

2.2 Simple Grasping

Due to the geometry of parallel jaw grippers, simple grasp planning can be reduced to computing a 6-DoF gripper pose in space. As such, analytical simple grasp planners may directly reason about the object geometry (e.g., [22]) or rank grasps using grasp quality metrics [19]. Recently, learning-based approaches that predict wrist poses have gained significant traction (e.g., [23, 24, 25]). We refer the readers to [26, 19, 20, 27, 28] for a thorough review. These approaches seldom scale directly to dexterous grasp planning, which is a significantly more complex problem.

2.3 Bilevel Optimization (BO) for Planning

While BO theory is established in literature [29], application of BO on motion planning is relatively new. BO has been applied to continuous systems [30], robotic locomotion [31, 32], simple pushing and pivoting [33], collaborative object manipulation [34], and task and motion planning [35]. To our best knowledge, this work is the first application of BO on dexterous robotic manipulation.

3 Method

Our method consists of a learned model that predicts a plausible finger placement 𝒫3×3\mathcal{P}\in\mathbb{R}^{3\times 3}, and an analytical process that computes a physically feasible dexterous grasp 𝒒22\bm{q}^{*}\in\mathbb{R}^{22} (Definition 1) guided by 𝒫\mathcal{P}. At inference time, the pipeline takes in an object point cloud and outputs a fully specified grasp configuration 𝒒\bm{q}^{*}. Figure 2 gives an overview of our method. We assume the object is grasped with exactly 3 fingers. This section first formally specify the modeled physical constraints, then discuss each of the components in the pipeline. Our implementation is available at github.com/Stanford-TML/dex_grasp.

Refer to caption
Figure 2: Overview of our method. We train a CVAE that predicts finger placements 𝒫3×3\mathcal{P}\in\mathbb{R}^{3\times 3} given an object point cloud 𝑶\bm{O}. At inference time, we first obtain a finger placement prediction 𝒫\mathcal{P}, which is not guaranteed to be physically feasible. Next, we compute a grasp configuration initial guess 𝒒22\bm{q}^{\prime}\in\mathbb{R}^{22} from 𝒫\mathcal{P}. Finally, we apply BO to compute a physically feasible grasp 𝒒22\bm{q}^{*}\in\mathbb{R}^{22}.
Definition 1 (Physically Feasible Dexterous Grasp).

Given an object 𝑶\bm{O} in environment 𝑬\bm{E}, a grasp configuration 𝒒\bm{q}, and a desired finger placement 𝒫=(𝒑1,𝒑2,𝒑3)3×3\mathcal{P}=(\bm{p}_{1},\bm{p}_{2},\bm{p}_{3})\in\mathbb{R}^{3\times 3}, we consider (𝑶,𝑬,𝒒,𝒫)(\bm{O},\bm{E},\bm{q},\mathcal{P}) to be physically feasible if they satisfy following dynamic and kinematic constraints. In our setup, 𝒒22\bm{q}\in\mathbb{R}^{22} and 𝑬\bm{E} is the fixed tabletop (see Section 4.1).

3.1 Dynamic Constraints: Wrench Closure and Friction Cone

For a static dexterous grasp, the dynamic constraints include the wrench closure constraint and the friction cone constraint. The wrench closure constraint requires the sum of the external wrench from all contact points to be zero: i=13𝒇i=𝟎\sum_{i=1}^{3}\bm{f}_{i}=\bm{0} and i=13𝒑i×𝒇i=𝟎\sum_{i=1}^{3}\bm{p}_{i}\times\bm{f}_{i}=\bm{0}. 𝒇i3\bm{f}_{i}\in\mathbb{R}^{3} is the unknown contact force applied at position 𝒑i3\bm{p}_{i}\in\mathbb{R}^{3} from the dexterous hand to the object.

Given the static friction coefficient μ\mu and the outward-pointing surface normal 𝒏^i\hat{\bm{n}}_{i}, the friction cone constraint requires the normal force to be nonnegative and the contact force to be within the friction cone. Using a polyhedral cone approximation ([36]) with an orthogonal basis 𝒕^i,j𝒏^i,j{1,2}\hat{\bm{t}}_{i,j}\perp\hat{\bm{n}}_{i},\forall j\in\{1,2\}, we arrive at the following approximated friction cone constraints:

0𝒇i𝒏^iand|𝒇i𝒕^i,j|μ𝒇i𝒏^i,i{1,2,3},j{1,2}.0\leq-\bm{f}_{i}\cdot\hat{\bm{n}}_{i}\;\;\;\mathrm{and}\;\;\;|\bm{f}_{i}\cdot\hat{\bm{t}}_{i,j}|\leq-\mu\bm{f}_{i}\cdot\hat{\bm{n}}_{i},\forall i\in\{1,2,3\},\forall j\in\{1,2\}. (1)

A grasp is dynamically feasible if it satisfies both wrench closure and friction cone constraints. Leveraging the polyhedral approximation, dynamic feasibility can be cast as a quadratic program (QP):

min𝒇1,𝒇2,𝒇3i=13𝒇i22+i=13𝒑i×𝒇i22 subject to  0<fmin𝒇i𝒏^i and (1).\displaystyle\begin{split}\min_{\bm{f}_{1},\bm{f}_{2},\bm{f}_{3}\;\;\;}&\|\sum_{i=1}^{3}\bm{f}_{i}\|^{2}_{2}+\|\sum_{i=1}^{3}\bm{p}_{i}\times\bm{f}_{i}\|^{2}_{2}\;\;\;\textrm{ subject to}\;\;0<{f}_{min}\leq-\bm{f}_{i}\cdot\hat{\bm{n}}_{i}\textrm{ and~{}\eqref{eqn:friction_cone_approx}}.\end{split} (2)

The optimization in Equation (2) has objective value 0 if and only if the grasp is dynamically feasible. Note that we set an arbitrary lower bound fmin+{f}_{min}\in\mathbb{R}^{+} on the normal force to avoid the trivial solution of 𝒇i=𝟎\bm{f}_{i}=\bm{0}. The exact value of fmin{f}_{min} is irrelevant because changing fmin{f}_{min} represents scaling the optimal contact forces, which does not affect membership in the friction cone.

3.2 Kinematic Constraints: Reachability and Collision

The kinematic constraints include reachability constraints and collision constraints. Reachability constraints enforce the kinematic ability of the robot hand to reach the contact point 𝒑i\bm{p}_{i} on the object surface 𝑶\partial\bm{O}, i.e. 𝒒:Ki(𝒒)=𝒑i𝑶,i{1,2,3}\exists\bm{q}:K_{i}(\bm{q})=\bm{p}_{i}\in\partial\bm{O},\;\forall i\in\{1,2,3\}. Here Ki:223K_{i}:\mathbb{R}^{22}\mapsto\mathbb{R}^{3} is the forward kinematics function that maps the hand state 𝒒\bm{q} to the position of the ii-th fingertip. Assuming only fingertip contacts, collision constraints ensure that no robot link 𝑳(𝒒)\bm{L}(\bm{q}) is in collision with the object 𝑶\bm{O} except at the fingertip or with the environment 𝑬\bm{E}, i.e. 𝑳(𝒒){𝑶𝑬}=.\bm{L}(\bm{q})\cap\left\{\bm{O}\cup\bm{E}\right\}=\emptyset. Both reachability and collision constraints are nonconvex constraints.

3.3 Learning to Predict Finger Placement from a Physically Feasible Grasp Dataset

We leverage a generative model to sample potential finger placements 𝒫\mathcal{P} given an arbitrary object observed as a point cloud 𝑶^\hat{\bm{O}}. An immediate challenge faced by the learning approach is the lack of large-scale datasets for dexterous robotic hands. While datasets for real human hands do exist, retargeting human grasping configurations to a robot hand with different kinematic structures and joint limits presents numerous challenges. We opt to synthesize a large-scale dataset of physically feasible grasping configurations for the Allegro robot hand used in this paper.

3.3.1 Creating a Physically Feasible Dexterous Grasp Dataset

We generate physically feasible grasps of six YCB objects [37] in simulation. The chosen objects are “cracker box,” “sugar box,” “tomato soup can,” “mustard bottle,” “gelatin box,” and “potted meat can.” We consider a realistic scenario where the object is placed on a flat surface instead of free floating. As such, we need to consider different object rest poses of on the surface in addition to finger placements. In summary, we first generate random object rest poses and enumerate all possible finger placements 𝒫¯\bar{\mathcal{P}} using the corresponding simulated object point cloud 𝑶^\hat{\bm{O}}. If a kinematically and dynamically feasible grasp 𝒒\bm{q} can be found for 𝒫¯\bar{\mathcal{P}}, (𝑶^,𝒒,𝒫¯)(\hat{\bm{O}},\bm{q},\bar{\mathcal{P}}) is added to the dataset \mathbb{P}. The detailed procedure is described in the supplementary material.

3.3.2 Training a Conditional Variational Autoencoder (CVAE)

To compute diverse grasps for arbitrary objects directly from point cloud observations, we predict fingertip contact points on the object surface with a conditional variational autoencoder (CVAE) [38]. Our model is adapted from [24] and consists of an encoder EE and a decoder DD that are based on the PointNet++ architecture [39]. The model seeks to maximize the likelihood of producing a set of contact points 𝒫\mathcal{P} deemed feasible in Section 3.3.1 given the point cloud 𝑶^\hat{\bm{O}}.

The encoder E(𝒛𝒫¯,𝑶^)E(\bm{z}\mid\bar{\mathcal{P}},\hat{\bm{O}}) maps a grasp 𝒫¯\bar{\mathcal{P}} and a point cloud 𝑶^\hat{\bm{O}} to the latent space. We assume the latent variable has a normal distribution 𝒩(𝟎,)\mathcal{N}(\bm{0},\mathcal{I}). Meanwhile, given a latent sample 𝒛E\bm{z}\sim E, the decoder attempts to reconstruct the finger placement 𝒫\mathcal{P}. We seek to minimize the element-wise L1L^{1}-norm reconstruction loss Lrec(𝒫,𝒫¯)𝒫𝒫¯1L_{rec}(\mathcal{P},\bar{\mathcal{P}})\triangleq\|\mathcal{P}-\bar{\mathcal{P}}\|_{1} for a feasible grasp from the dataset 𝒫¯\bar{\mathcal{P}}\in\mathbb{P}. Additionally, a KL divergence loss 𝒟KL\mathcal{D}_{KL} is applied on the latent distribution E()E(\cdot\mid\cdot) to ensure a normally distributed latent variable. The complete loss function of the network is defined as

L𝒛E,𝒫¯Lrec(𝒫,𝒫¯)+α𝒟KL(E(𝒛𝒫¯,𝑶^)||𝒩(𝟎,)).L\triangleq\sum_{\bm{z}\sim E,\bar{\mathcal{P}}\in\mathbb{P}}L_{rec}(\mathcal{P},\bar{\mathcal{P}})+\alpha\mathcal{D}_{KL}\left(E(\bm{z}\mid\bar{\mathcal{P}},\hat{\bm{O}})\;\middle|\middle|\;\mathcal{N}(\bm{0},\mathcal{I})\right). (3)

At inference time, EE is removed and a latent sample 𝒛\bm{z} is drawn from 𝒩(𝟎,)\mathcal{N}(\bm{0},\mathcal{I}). This is passed into the decoder D(𝒫|𝑶^,𝒛)D(\mathcal{P}|\hat{\bm{O}},\bm{z}) along with the point cloud 𝑶^\hat{\bm{O}} to produce the grasp point prediction 𝒫\mathcal{P}.

3.4 Computing Grasps with Physical Feasibility Guarantees using Bilevel Optimization (BO)

While the grasps in the training dataset are physically feasible by construction, there is no guarantee that the CVAE output, 𝒫\mathcal{P}, is physically feasible. Additionally, 𝒫\mathcal{P} only specifies the finger placement instead of the full hand configuration. We propose a BO to obtain a physically feasible grasping pose 𝒒\bm{q} given 𝒫\mathcal{P}. To seed the BO with 𝒫\mathcal{P}, we first obtain an Euclidean projection of 𝒫\mathcal{P} onto 𝑶\partial\bm{O}, denoted as 𝒫{𝒑i|𝒑i=argmin𝒑o𝑶𝒑o𝒑i2,i{1,2,3}}.\mathcal{P}^{\prime}\triangleq\left\{\bm{p}^{\prime}_{i}\;\middle|\;\bm{p}^{\prime}_{i}=\operatorname*{arg\,min}_{\bm{p}_{o}\in\partial\bm{O}}{\|\bm{p}_{o}-\bm{p}_{i}\|_{2}},\;i\in\{1,2,3\}\right\}. Next, we solve an inverse kinematics (IK) problem for a grasp configuration finger placement problem specified by 𝒫\mathcal{P}^{\prime} up to a numerical tolerance ϵ\epsilon, i.e. find 𝒒:Ki(𝒒)𝒑i2ϵ,Ki(𝒒)𝑶,i{1,2,3}.\textrm{find }\bm{q}^{\prime}:\|K_{i}(\bm{q}^{\prime})-\bm{p}^{\prime}_{i}\|_{2}\leq\epsilon,\;K_{i}(\bm{q}^{\prime})\in\partial\bm{O},\;\forall i\in\{1,2,3\}. Ki:223K_{i}:\mathbb{R}^{22}\mapsto\mathbb{R}^{3} is the forward kinematics function of the iith finger. Ki(𝒒)𝑶K_{i}(\bm{q}^{\prime})\in\partial\bm{O} is implemented as a constraint on the finger-object signed distance D(𝒑,𝑶)[dmin,dmax]D(\bm{p},\bm{O})\in\left[d_{min},d_{max}\right]. The IK solution 𝒒\bm{q}^{\prime} serves as the initial guess for the BO.

3.4.1 Formulating the Bilevel Optimization Grasp Refinement

Accounting for the kinematic and dynamic constraints, the naïve formulation of the grasp optimization problem is given in (4).

min𝒒,𝒇1,𝒇2,𝒇30subject toKi(𝒒)𝑶,𝑳(𝒒){𝑶𝑬}=,i=13𝒇i2=0,i=13Ki(𝒒)×𝒇i2=0,fmin𝒇i𝒏^i,|𝒇i𝒕^i,j|μ𝒇i𝒏^i,i{1,2,3},j{1,2}.\displaystyle\begin{split}\min_{\bm{q},\bm{f}_{1},\bm{f}_{2},\bm{f}_{3}}\;\;\;&0\\ \textrm{subject to}\;\;\;&K_{i}(\bm{q})\in\partial{\bm{O}},\;\bm{L}(\bm{q})\cap\left\{\bm{O}\cup\bm{E}\right\}=\emptyset,\;\|\sum_{i=1}^{3}\bm{f}_{i}\|_{2}=0,\;\|\sum_{i=1}^{3}K_{i}(\bm{q})\times\bm{f}_{i}\|_{2}=0,\\ &f_{min}\leq-\bm{f}_{i}\cdot\hat{\bm{n}}_{i},\;\;|\bm{f}_{i}\cdot\hat{\bm{t}}_{i,j}|\leq-\mu\bm{f}_{i}\cdot\hat{\bm{n}}_{i},\;\forall i\in\{1,2,3\},\forall j\in\{1,2\}.\end{split} (4)

Equation (4) does not solve well in practice due to its complexity and nonconvexity. Additionally, while 𝒒\bm{q}^{\prime} may serve as an initial guess for 𝒒\bm{q}, the initial guess for 𝒇\bm{f} is not obvious. Naively relaxing Equation (4), e.g., replacing the objective with min𝒒,𝒇1,𝒇2,𝒇3i=13𝒇i22+i=13Ki(𝒒)×𝒇i22\min_{\bm{q},\bm{f}_{1},\bm{f}_{2},\bm{f}_{3}}\;\|\sum_{i=1}^{3}\bm{f}_{i}\|^{2}_{2}+\|\sum_{i=1}^{3}K_{i}(\bm{q})\times\bm{f}_{i}\|^{2}_{2}, may produce suboptimal 𝒇\bm{f} and, consequently, incorrect conclusion on grasp’s dynamic feasibility.

We propose leveraging bilevel optimization [29, 32] to offload the dynamic feasibility computation from the main optimization. Define J(𝒒)J(\bm{q}) as the minimum objective value of the dynamic constraint QP in Equation (2) with pi=Ki(𝒒)p_{i}=K_{i}(\bm{q}):

J(𝒒)min𝒇1,𝒇2,𝒇3i=13𝒇i22+i=13𝒑i×𝒇i22 subject to  0<fmin𝒇i𝒏^i and (1).J(\bm{q})\triangleq\min_{\bm{f}_{1},\bm{f}_{2},\bm{f}_{3}\;\;\;}\|\sum_{i=1}^{3}\bm{f}_{i}\|^{2}_{2}+\|\sum_{i=1}^{3}\bm{p}_{i}\times\bm{f}_{i}\|^{2}_{2}\;\;\;\textrm{ subject to}\;\;0<{f}_{min}\leq-\bm{f}_{i}\cdot\hat{\bm{n}}_{i}\textrm{ and~{}\eqref{eqn:friction_cone_approx}}. (5)

We observe that the force closure constraint J(𝒒)=0J(\bm{q})=0 can be abstracted away to form a “lower-level” QP. J(𝒒)=0J(\bm{q})=0 can be solved to optimality with exiting QP solvers without reliance on a good initial guess for 𝒇i\bm{f}_{i}.

Applying this abstraction to Equation (4) yields the upper-level problem:

min𝒒   0subject toJ(𝒒)=0,Ki(𝒒)𝑶,𝑳(𝒒){𝑶𝑬}=.\displaystyle\begin{split}\min_{\bm{q}}\;\;\;0\;\;\;\textrm{subject to}\;\;\;J(\bm{q})=0,\ K_{i}(\bm{q})\in\partial{\bm{O}},\ \bm{L}(\bm{q})\cap\left\{\bm{O}\cup\bm{E}\right\}=\emptyset.\end{split} (6)

Equation (6) is a bilevel optimization as J(𝒒)=0J(\bm{q})=0 is a constraint on the minimizer of another optimization problem (Equation (6)). While Equation (6) still defines a nonconvex optimization, the choice for 𝒇i\bm{f}_{i} has been abstracted away entirely to the lower level QP solver. By construction, a valid solution to Equation (6), denoted as 𝒒\bm{q}^{*}, defines a physically feasible grasp.

3.4.2 Solving the Bilevel Optimization Grasp Refinement

In practice, to solve Equation (6) with a nonconvex optimizer (e.g., SNOPT [40]), one needs to obtain the gradient of each of the constraints with respect to 𝒒\bm{q}. The gradients of the kinematic constraints can be obtained from simulators with automatic differentiation capabilities (e.g., Drake [41]). The gradient of J(𝒒)J(\bm{q}), 𝒒J(𝒒)\nabla_{\bm{q}}J(\bm{q}), can obtained using differentiable optimization [42]. At a high level, differentiable optimization computes 𝒒J(𝒒)\nabla_{\bm{q}}J(\bm{q}) by taking the matrix differentials of the KKT conditions of the optimization problem at its solution.

We emphasize that the bilevel optimization in Equation (6) is still a highly nonconvex optimization, and the optimizer will likely return a locally optimal 𝒒\bm{q}^{*} in the vicinity of 𝒒\bm{q}^{\prime}. Intuitively, this could be interpreted as optimizing within the grasp family of the CVAE prediction (e.g., “left side grasp” or “top-down grasp”), which implies the necessity of CVAE in the process. This is supported empirically by our ablation studies in Section 4. In practice, the bilevel optimization provides a certification of physical feasibility for the CVAE prediction. A poor 𝒒\bm{q}^{\prime} choice will likely results in the nonconvex optimizer returning infeasibility. This serves as the condition to reject 𝒫\mathcal{P} and generate a different CVAE grasp prediction with another latent variable sample.

4 Experiments

Our trained CVAE achieved a test reconstruction error of 0.50.5cm. More training details are available in the supplementary material. The effects of applying BO is shown in Figures 3(a) and 3(b). The finger placement prior to BO cannot form force closure as 𝒏^i𝒏^j>0,i,j{1,2,3}\hat{\bm{n}}_{i}\cdot\hat{\bm{n}}_{j}>0,\forall i,j\in\{1,2,3\}. BO shifted the thumb and middle fingers to an antipodal configuration, which allows for force closure. This validates our approach’s ability to make an initially infeasible grasp prediction physically feasible. The rest of this section focuses on hardware experiments and evaluations.

Refer to caption
(a) Before BO (𝒒\bm{q}^{\prime}).
Refer to caption
(b) After BO (𝒒\bm{q}^{*}).
Refer to caption
(c) Hardware setup.
Refer to caption
(d) Planned and executed grasps.
Figure 3: 3(a) and 3(b): Before and after BO, on a bottom-up view of a mustard bottle grasp. Using the color encoding of red: thumb, green: index, and blue: middle, we show 𝒫\mathcal{P} (solid spheres), 𝒫\mathcal{P}^{\prime} (transparent spheres) and 𝒏^i\hat{\bm{n}}_{i} (colored lines). The direction of the computed contact forces 𝒇i\bm{f}_{i} are shown at the respective fingertip with yellow lines. The mismatch between K(𝒒)K(\bm{q}^{\prime}) and 𝒫\mathcal{P}^{\prime} is due to the numerical tolerance ϵ\epsilon. 3(c): Hardware setup. See Section 4.1 for more details. 3(d): Grasp sim-to-real. Correspondence between the planned grasp in simulation and execution on hardware.

4.1 Hardware Setup

The hardware setup and evaluated objects are shown in Figure 3(c).
Hardware. A 16-DoF Allegro v4.0 right hand was used for hardware grasping experiments. The Allegro hand was mounted on a 7-DoF Franka Emika Panda arm to realize the planned wrist pose. The object point cloud was captured by 4 stationary Intel RealSense D435 depth cameras. The implementation details are available in the supplementary material.
Evaluated Objects. We evaluated our method on 20 rigid household objects resting on the table. We categorize our objects into three sets:
\bullet\;\,3 Seen objects. (Leftmost column) mustard bottle, soup can, and sugar box from the training set.
\bullet 4 Familiar objects. (Second-from-left column) Unseen objects with geometries similar to training objects. Includes boxes (webcam box, mask box) and cylinders (chip can, soda can).
\bullet 13 Novel objects. Objects whose geometries are distinct from that of any object in the training set. From left to right, front to back: tetrahedron, massage ball, castle, pill bottle, glasses case, condiment bottle, hairspray bottle, ramen, lego, sandwich box (side), pear, sandwich box (upright), alcohol bottle.

4.2 Experiment Procedure

At the start of each trial, the object is placed in a specified pose in front of the robot. After observing the point cloud and computing a grasp, the Franka arm first brings the hand to the desired wrist pose using an ad hoc trajectory planner. The finger joint angles 𝜽16\bm{\theta}^{*}\in\mathbb{R}^{16} from 𝒒\bm{q}^{*} is then executed on the Allegro hand. Contact forces are provided by squeezing the fingertips according to the planned contact forces: 𝜽𝜽+k(𝜽𝒑i)T𝒇i\bm{\theta}^{*}\leftarrow\bm{\theta}^{*}+k(\nabla_{\bm{\theta}}\bm{p}_{i})^{T}\bm{f}_{i}. Here 𝜽𝒑i\nabla_{\bm{\theta}}\bm{p}_{i} is the Jacobian of the fingertip locations with respect to 𝜽\bm{\theta}, and kk is a fixed “stiffness” constant as motivated by impedance control. The Franka then attempts to lift the object to a fixed height approximately 4343cm above the table. A trial is considered successful if the object is lifted and all three fingers remain in contact with the object. 3 repeats are performed for hardcoded policy trials as they are deterministic up to the object point cloud observation. All other trials are repeated 6 times. Our hardware pipeline was implemented with the intent to execute our planned grasp as accurately as possible. Figure 3(d) illustrates the simulation-hardware grasp correspondence.

4.3 Baseline and Ablation Studies

We compare our method against the following approaches. We excluded “BO only” in our ablation studies as it does not return a result without CVAE in practice. BO seldom returns a solution if seeded with a kinematically infeasible solution (e.g., open hand).
\bullet Hardcoded grasp baseline. Using the set of seen objects, we designed a top-down tripod grasp policy that chooses the wrist pose based on the point cloud and executes a fixed grasp.
\bullet CVAE-only as an ablation study. We solved a collision-free inverse kinematics problem that attempts to match the CVAE predicted fingertip positions 𝒫\mathcal{P}. This mimics a “learning only” approach.
\bullet CVAE-kinematic as an ablation study. We ablated the “lower level” part of the optimization (Equation (5)) away and use 𝒒\bm{q}^{\prime} directly without bilevel optimization. This ablation considers kinematic constraints but not dynamic constraints.

4.4 Grasp Trial Results

Our method achieved an overall success rate of 86.7% over 120 grasp trials on 20 objects. This is superior to the hardcoded baseline, which achieved an overall success rate of 53.3%53.3\%. On seen objects, our CVAE-only ablation achieved a 38.9%38.9\% success rate and our CVAE-kinematic ablation achieved an 88.9%88.9\% success rate. The results are summarized in Table 1, and the details are available in the supplementary material. A video summary is available at youtu.be/9DTrImbN99I.

Our method can grasp challenging novel objects. This includes objects that are difficult for parallel grippers and suction cups to grasp. We discuss the challenges of some of our test objects in the supplementary material. The hardcoded baseline nearly always succeeds on objects that allow top-down grasps and are similar in size to the seen object used for tuning. However, the baseline fails on objects that either require a different grasp type or have significantly different size.

Our method can generate diverse grasps through sampling different latent z\bm{z}’s. Figure 4 shows 10 distinct successful grasps performed on the upright sugar box.

The median time to produce a grasp with our method is 14.4 seconds. The median repeats for each component of our method when generating a grasp is two CVAE 𝒛\bm{z} samples, two IK solves for 𝒒\bm{q}^{\prime}, and two bilevel optimizations for 𝒒\bm{q}^{*}. Detailed timing and repetition results are available in the supplementary material.

Table 1: Object grasping statistics. Our method achieved superior grasp success rate compared to the ablations and the hardcoded baseline.
Category (object count) Seen (3) Familiar (4) Novel (13) Overall (20)
Ours 17/18 94.4% 23/24 95.8% 64/78 82.1% 104/120 86.7%
CVAE only 7/18 38.9% - - - - - -
CVAE+kinematics 16/18 88.9% - - - - - -
Hardcoded 8/9 88.9% 9/12 75.0% 15/39 38.5% 32/60 53.3%
Refer to caption
Figure 4: Diverse grasps generated by our method. Each image is a unique grasp generated from different sampled latent variables. All grasps were computed for the same initial object rest pose.

4.5 Quantitative evaluation of physical constraint enforcement

Table 2 summarizes the quantitative evaluation of physical constraints.
Kinematic constraints. The maximum finger-object distance constraint violation in a grasp planned by our method is 0.08cm, which is negligible in practice as it is smaller than other error sources such as camera observation error. The CVAE-kinematics ablation, which only considers kinematic constraints, achieved comparable results. The CVAE-only ablation has a maximum violation of 2.032cm, confirming that while CVAE alone can produce qualitatively correct grasp, it cannot enforce kinematic constraints precisely.
Dynamic constraints. To evaluate dynamic constraint satisfaction, we chose “force (torque) ratio” as our metric, i.e. i=1,2,3𝒇i213i=1,2,3𝒇i2×100%,\frac{\|\sum_{i=1,2,3}\bm{f}_{i}\|_{2}}{\frac{1}{3}\sum_{i=1,2,3}\|\bm{f}_{i}\|_{2}}\times 100\%, and i=1,2,3𝒑i×𝒇i213i=1,2,3𝒑i×𝒇i2×100%.\frac{\|\sum_{i=1,2,3}\bm{p}_{i}\times\bm{f}_{i}\|_{2}}{\frac{1}{3}\sum_{i=1,2,3}\|\bm{p}_{i}\times\bm{f}_{i}\|_{2}}\times 100\%. If wrench closure is achieved, both ratios should be zero. All ratios are reported as median, (25th percentile, 75th percentile). We computed these ratios for grasps planned with our method and CVAE-kinematics ablations. These ratios are not computed for CVAE-only because there are often fewer than 3 finger-object contacts. Our method achieved a median of 0.01%\leq 0.01\% on both ratios, showing that BO is effective in enforcing wrench closure. We note that while the CVAE+kinematics ablation does not explicitly consider external wrench, it still produced many grasps that achieve wrench closure by coincidence. This reflects CVAE’s ability to produce qualitatively correct grasps.

Table 2: Evaluation of physical constraint on various grasps. A physically feasible grasp should satisfy D(𝒑,𝑶)[dmin,dmax]=[0.68,0.32]D(\bm{p},\bm{O})\in\left[d_{min},d_{max}\right]=\left[-0.68,-0.32\right] and zero force and torque ratios.
Median DD Min DD Max DD Max violation Force ratio Torque ratio
Ours (all objects) -0.32 -0.68 -0.24 0.08 (0.00, 0.02) 0.01 (0.00, 2.68)
CVAE-kinematics (seen objects) -0.32 -0.64 -0.32 0.04 0.00 (0.00, 0.12) 0.01 (0.00, 19.54)
CVAE-only (seen objects) 0.33 -1.10 1.71 2.03 - -
Table 3: Correlation between dynamic feasibility and hardware success. Only the grasps plans with wrench closure resulted in successful hardware execution.
Successful Failed
Force ratio Torque ratio Force ratio Torque ratio
BO rejected (12 of 12 failed) - - 61.81 (25.51, 91.65) 48.53 (39.89, 78.68)
CVAE-kinematics (7 of 12 failed) 0.00 (0.00, 0.00) 0.00 (0.00, 0.01) 50.71 (19.13, 128.85) 36.75 (30.48, 131.82)

To show that physical feasibility is a necessary condition for a successful grasp, we examined the CVAE+kinematic trials, which may not satisfy dynamic constraints, on “ramen” and “mustard bottle.” We also executed grasp plans that are reported to be infeasible by BO. Table 3 summarizes the results. There is a strong correlation between hardware success and wrench closure. Moreover, all grasps reported to be infeasible by BO failed on hardware. This confirms that grasp refinement derived from rigorous physics-inspired metrics can significantly improve the final grasp plan quality.

5 Limitations and Conclusion

Limitations.

Our method requires an observation of the full object point cloud. As our choice of physical constraint formulation does not include explicit robustness margins, estimation errors from the vision pipeline is a major source of grasp failures. This may be addressed by introducing point cloud completion (e.g., [43, 44, 45]) or grasp robustness metrics (e.g., [17]). Additionally, it is currently not possible to specify which type grasps to generate (e.g., “top grasp” or “side grasp”) with our method. Finally, our method currently assumes that the object to grasp is placed on a flat surface and that there are no other objects in the scene.

Conclusion.

This work presents a novel pipeline that combines learning-based grasp generation with bilevel optimization to produce diverse and physically feasible dexterous grasps. Our method achieved a grasp success rate of 86.7%86.7\% on 20 challenging real-world objects. Ablation studies demonstrated that an integrative approach combining learned models and rigorous physics-inspired metrics can achieve superior grasp output quality. Grasps initially generated by CVAE may not satisfy all physical constraints. However, by incorporating bilevel optimization for grasp refinement and rejection sampling, poor grasp predictions can be removed. This paradigm of combining learning and physics-inspired bilevel optimization may be applied to other robotic manipulation tasks.

Acknowledgments

We would like to thank Oussama Khatib, Jeannette Bohg, Dorsa Sadigh, Samuel Clarke, Elena Galbally Herrero, Wesley Guo, and Yanchao Yang for their assistance on setting up the hardware experiments. We would also like to thank Chen Wang and Jiaman Li for advice on training the CVAE model. Our research is supported by NSF-NRI-2024247, NSF-FRR-2153854, Stanford-HAI-203112, and the Facebook Fellowship.

References

  • Huang et al. [2013] B. Huang, S. El-Khoury, M. Li, J. J. Bryson, and A. Billard. Learning a real time grasping strategy. In 2013 IEEE International Conference on Robotics and Automation, pages 593–600. IEEE, 2013.
  • Lundell et al. [2021] J. Lundell, F. Verdoja, and V. Kyrki. Ddgc: Generative deep dexterous grasping in clutter. IEEE Robotics and Automation Letters, 6(4):6899–6906, 2021.
  • Shao et al. [2020] L. Shao, F. Ferreira, M. Jorda, V. Nambiar, J. Luo, E. Solowjow, J. A. Ojea, O. Khatib, and J. Bohg. Unigrasp: Learning a unified model to grasp with multifingered robotic hands. IEEE Robotics and Automation Letters, 5(2):2286–2293, 2020.
  • Wei et al. [2022] W. Wei, D. Li, P. Wang, Y. Li, W. Li, Y. Luo, and J. Zhong. Dvgg: Deep variational grasp generation for dextrous manipulation. IEEE Robotics and Automation Letters, 7(2):1659–1666, 2022.
  • Lu et al. [2020] Q. Lu, K. Chenna, B. Sundaralingam, and T. Hermans. Planning multi-fingered grasps as probabilistic inference in a learned deep network. In Robotics Research, pages 455–472. Springer, 2020.
  • Mandikal and Grauman [2021] P. Mandikal and K. Grauman. Learning dexterous grasping with object-centric visual affordances. In 2021 IEEE International Conference on Robotics and Automation (ICRA), pages 6169–6176. IEEE, 2021.
  • Pfrommer et al. [2020] S. Pfrommer, M. Halm, and M. Posa. Contactnets: Learning discontinuous contact dynamics with smooth, implicit representations. arXiv preprint arXiv:2009.11193, 2020.
  • Jiang et al. [2022] Y. Jiang, J. Sun, and C. K. Liu. Data-augmented contact model for rigid body simulation. In Learning for Dynamics and Control Conference, pages 378–390. PMLR, 2022.
  • Zhu et al. [2021] T. Zhu, R. Wu, X. Lin, and Y. Sun. Toward human-like grasp: Dexterous grasping via semantic representation of object-hand. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 15741–15751, 2021.
  • Brahmbhatt et al. [2020] S. Brahmbhatt, C. Tang, C. D. Twigg, C. C. Kemp, and J. Hays. Contactpose: A dataset of grasps with object contact and hand pose. In European Conference on Computer Vision, pages 361–378. Springer, 2020.
  • Taheri et al. [2020] O. Taheri, N. Ghorbani, M. J. Black, and D. Tzionas. Grab: A dataset of whole-body human grasping of objects. In European conference on computer vision, pages 581–600. Springer, 2020.
  • Jiang et al. [2021] H. Jiang, S. Liu, J. Wang, and X. Wang. Hand-object contact consistency reasoning for human grasps generation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 11107–11116, 2021.
  • Yang et al. [2021] L. Yang, X. Zhan, K. Li, W. Xu, J. Li, and C. Lu. Cpf: Learning a contact potential field to model the hand-object interaction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 11097–11106, 2021.
  • Corona et al. [2020] E. Corona, A. Pumarola, G. Alenya, F. Moreno-Noguer, and G. Rogez. Ganhand: Predicting human grasp affordances in multi-object scenes. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 5031–5041, 2020.
  • Ciocarlie and Allen [2009] M. T. Ciocarlie and P. K. Allen. Hand posture subspaces for dexterous robotic grasping. The International Journal of Robotics Research, 28(7):851–867, 2009.
  • Pokorny and Kragic [2013] F. T. Pokorny and D. Kragic. Classical grasp quality evaluation: New algorithms and theory. In 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 3493–3500. IEEE, 2013.
  • Ferrari and Canny [1992] C. Ferrari and J. F. Canny. Planning optimal grasps. In ICRA, volume 3, page 6, 1992.
  • Liu et al. [2021] T. Liu, Z. Liu, Z. Jiao, Y. Zhu, and S.-C. Zhu. Synthesizing diverse and physically stable grasps with arbitrary hand structures using differentiable force closure estimator. IEEE Robotics and Automation Letters, 7(1):470–477, 2021.
  • Roa and Suárez [2015] M. A. Roa and R. Suárez. Grasp quality measures: review and performance. Autonomous robots, 38(1):65–88, 2015.
  • Bohg et al. [2013] J. Bohg, A. Morales, T. Asfour, and D. Kragic. Data-driven grasp synthesis—a survey. IEEE Transactions on robotics, 30(2):289–309, 2013.
  • Liu et al. [2019] M. Liu, Z. Pan, K. Xu, K. Ganguly, and D. Manocha. Generating grasp poses for a high-dof gripper using neural networks. In 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 1518–1525. IEEE, 2019.
  • Berenson et al. [2007] D. Berenson, R. Diankov, K. Nishiwaki, S. Kagami, and J. Kuffner. Grasp planning in complex scenes. In 2007 7th IEEE-RAS International Conference on Humanoid Robots, pages 42–48. IEEE, 2007.
  • Saxena et al. [2006] A. Saxena, J. Driemeyer, J. Kearns, and A. Ng. Robotic grasping of novel objects. Advances in neural information processing systems, 19, 2006.
  • Mousavian et al. [2019] A. Mousavian, C. Eppner, and D. Fox. 6-dof graspnet: Variational grasp generation for object manipulation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2901–2910, 2019.
  • Mahler et al. [2017] J. Mahler, J. Liang, S. Niyaz, M. Laskey, R. Doan, X. Liu, J. A. Ojea, and K. Goldberg. Dex-net 2.0: Deep learning to plan robust grasps with synthetic point clouds and analytic grasp metrics. arXiv preprint arXiv:1703.09312, 2017.
  • Du et al. [2021] G. Du, K. Wang, S. Lian, and K. Zhao. Vision-based robotic grasping from object localization, object pose estimation to grasp estimation for parallel grippers: a review. Artificial Intelligence Review, 54(3):1677–1734, 2021.
  • Caldera et al. [2018] S. Caldera, A. Rassau, and D. Chai. Review of deep learning methods in robotic grasp detection. Multimodal Technologies and Interaction, 2(3):57, 2018.
  • Kleeberger et al. [2020] K. Kleeberger, R. Bormann, W. Kraus, and M. F. Huber. A survey on learning-based robotic grasping. Current Robotics Reports, 1(4):239–249, 2020.
  • Sinha et al. [2017] A. Sinha, P. Malo, and K. Deb. A review on bilevel optimization: from classical to evolutionary approaches and applications. IEEE Transactions on Evolutionary Computation, 22(2):276–295, 2017.
  • Zimmermann et al. [2020] S. Zimmermann, G. Hakimifard, M. Zamora, R. Poranne, and S. Coros. A multi-level optimization framework for simultaneous grasping and motion planning. IEEE Robotics and Automation Letters, 5(2):2966–2972, 2020.
  • Zhu et al. [2021] Y. Zhu, Z. Pan, and K. Hauser. Contact-implicit trajectory optimization with learned deformable contacts using bilevel optimization. In 2021 IEEE International Conference on Robotics and Automation (ICRA), pages 9921–9927. IEEE, 2021.
  • Landry et al. [2019] B. Landry, J. Lorenzetti, Z. Manchester, and M. Pavone. Bilevel optimization for planning through contact: A semidirect method. arXiv preprint arXiv:1906.04292, 2019.
  • Shirai et al. [2022] Y. Shirai, D. K. Jha, A. U. Raghunathan, and D. Romeres. Robust pivoting: Exploiting frictional stability using bilevel optimization. In 2022 International Conference on Robotics and Automation (ICRA), pages 992–998. IEEE, 2022.
  • Stouraitis et al. [2020] T. Stouraitis, I. Chatzinikolaidis, M. Gienger, and S. Vijayakumar. Online hybrid motion planning for dyadic collaborative manipulation via bilevel optimization. IEEE Transactions on Robotics, 36(5):1452–1471, 2020.
  • Zhao et al. [2021] Z. Zhao, Z. Zhou, M. Park, and Y. Zhao. Sydebo: Symbolic-decision-embedded bilevel optimization for long-horizon manipulation in dynamic environments. IEEE Access, 9:128817–128826, 2021.
  • Stewart and Trinkle [1996] D. E. Stewart and J. C. Trinkle. An implicit time-stepping scheme for rigid body dynamics with inelastic collisions and coulomb friction. International Journal for Numerical Methods in Engineering, 39(15):2673–2691, 1996.
  • Calli et al. [2017] B. Calli, A. Singh, J. Bruce, A. Walsman, K. Konolige, S. Srinivasa, P. Abbeel, and A. M. Dollar. Yale-cmu-berkeley dataset for robotic manipulation research. The International Journal of Robotics Research, 36(3):261–268, 2017.
  • Kingma and Welling [2013] D. P. Kingma and M. Welling. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114, 2013.
  • Qi et al. [2017] C. R. Qi, L. Yi, H. Su, and L. J. Guibas. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Advances in neural information processing systems, 30, 2017.
  • Gill et al. [2005] P. E. Gill, W. Murray, and M. A. Saunders. Snopt: An sqp algorithm for large-scale constrained optimization. SIAM review, 47(1):99–131, 2005.
  • Tedrake and the Drake Development Team [2019] R. Tedrake and the Drake Development Team. Drake: Model-based design and verification for robotics, 2019. URL https://drake.mit.edu.
  • Amos and Kolter [2017] B. Amos and J. Z. Kolter. Optnet: Differentiable optimization as a layer in neural networks. In International Conference on Machine Learning, pages 136–145. PMLR, 2017.
  • Yuan et al. [2018] W. Yuan, T. Khot, D. Held, C. Mertz, and M. Hebert. Pcn: Point completion network. In 2018 International Conference on 3D Vision (3DV), pages 728–737. IEEE, 2018.
  • Sarmad et al. [2019] M. Sarmad, H. J. Lee, and Y. M. Kim. Rl-gan-net: A reinforcement learning agent controlled gan network for real-time point cloud shape completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5898–5907, 2019.
  • Han et al. [2017] X. Han, Z. Li, H. Huang, E. Kalogerakis, and Y. Yu. High-resolution shape completion using deep neural networks for global structure and local geometry inference. In Proceedings of the IEEE international conference on computer vision, pages 85–93, 2017.