This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Stable Object Placing using Curl and Diff Features
of Vision-based Tactile Sensors

Kuniyuki Takahashi1, Shimpei Masuda1, Tadahiro Taniguchi2 1K.Takahashi and S. Masuda are with Preferred Networks, Inc. {takahashi, masuda}@preferred.jp2T.Taniguchi is with Ritsumeikan University, College of Information Science and Engineering. [email protected]
Abstract

Ensuring stable object placement is crucial to prevent objects from toppling over, breaking, or causing spills. When an object makes initial contact to a surface, and some force is exerted, the moment of rotation caused by the instability of the object’s placing can cause the object to rotate in a certain direction (henceforth referred to as direction of corrective rotation). Existing methods often employ a Force/Torque (F/T) sensor to estimate the direction of corrective rotation by detecting the moment of rotation as a torque. However, its effectiveness may be hampered by sensor noise and the tension of the external wiring of robot cables. To address these issues, we propose a method for stable object placing using GelSights, vision-based tactile sensors, as an alternative to F/T sensors. Our method estimates the direction of corrective rotation of objects using the displacement of the black dot pattern on the elastomeric surface of GelSight. We calculate the Curl from vector analysis, indicative of the rotational field magnitude and direction of the displacement of the black dots pattern. Simultaneously, we calculate the difference (Diff) of displacement between the left and right fingers’ GelSight’s black dots. Then, the robot can manipulate the objects’ pose using Curl and Diff features, facilitating stable placing. Across experiments, handling 18 differently characterized objects, our method achieves precise placing accuracy (less than 1-degree error) in nearly 100% of cases. 333An accompanying video is available at the following link:
https://youtu.be/fQbmCksVHlU

Refer to caption
Figure 1: Concept of the proposed method: The displacement of the black dots on GelSight is analyzed in two patterns corresponding to an object’s directions of corrective rotation: roll and pitch. By calculating the Curl and Diff of the displacement of the black dot to estimate the object’s direction of corrective rotation and manipulate the object accordingly, the robot ensures its stable placement.

I Introduction

Stable placing of grasped objects is important for many reasons, such as 1) preventing objects from toppling over, 2) preventing damage to the objects themselves or their surroundings, and 3) keeping their contents, such as liquids or small particles, from spilling. Consider settings like chemistry and biology laboratories, for instance, where beakers and flasks are handled routinely. With the growing demand for laboratory automation, the risk of these objects toppling over with resulting breakage or spillage of their contents has become a significant concern.

We define stable placing as a state where a flat-bottomed object’s bottom surface is parallel to a table or a non-flat object contacts the table at three or more points. Conversely, unstable placing is defined as the state where a flat-bottom object’s bottom surface isn’t parallel to the table or a non-flat object contacts the table at less than three points. Unstable placement can result from many factors, such as inaccurate object pose estimation.

The moment of rotation caused by the instability of the object’s placing can cause the object to rotate in a certain direction (henceforth referred to as direction of corrective rotation). The existing method for stable object placing is to use a Force/Torque (F/T) sensor mounted on the robot’s wrist by estimating the object’s direction of corrective rotation from the torque applied to the grasped object [1]. However, several issues when using F/T sensors make this estimation uncertain. 1) Tension of external wiring of robot cables: The F/T sensor needs to be calibrated to compensate for the weight of the robot’s wrist or gripper. However, changes in the robot’s pose can cause variable tension in the cables, thereby complicating accurate calibration and preventing precise force and torque detection. 2) Sensor noise: The F/T sensor contains noise, which makes the measured sensor values uncertain. The F/T sensors attached to the wrist often require a larger rated capacity. In general, increasing the sensor’s rated capacity sacrifices the ability to measure changes in a small range.

To address these issues, typical approaches include mounting a F/T or tactile sensor on the fingertip. As precision manipulation needs detailed information like fingertip contact area, interest in the use of tactile sensors, particularly vision-based ones like GelSight [2], are growing [3, 4, 5]. In this context, this study uses GelSight tactile sensors. Unlike F/T sensors attached to the wrist, tactile sensors are attached to the end-effector’s fingertip and are not affected by cable tension from other sensors (such as a camera’s USB cable). Moreover, since detecting small magnitudes of force and torque is required, they are typically designed to minimize the impact of noise.

GelSight uses a camera to measure the deformation of an attached elastomer during contact with a surface. Some GelSight devices feature black dots on their surface to track their displacement when they make contact with a surface. The displacement of GelSight’s black dots provides significant information, and some previous studies used it for a learning-based approach in peg-insertion task [6, 5].

Our work proposes utilizing these displacements as an alternative to torque measurements traditionally given by F/T sensors to achieve stable object placing. Our observations of how tactile sensors respond to various objects during unstable placement suggest that the displacement patterns of GelSight’s black dots can be categorized into two types with two directions of corrective rotation: roll and pitch (Fig. 1). After analyzing each classification, the displacement of GelSight’s black dots can be expressed as a one-dimensional feature for each direction of corrective rotation. The direction of corrective rotation of the object is estimated by calculating the Curl from vector analysis for the roll axis, indicative of the rotational field magnitude and direction of the displacement of the black dots pattern, and the difference (Diff) of displacement for the pitch axis between the left and right fingers’ GelSight’s black dots. These features allow accurate stable object placing on the robot. The novelty and contribution of this method are as follows:

  • Estimation of the direction of corrective rotation of objects for stable placing using a training-free method employing Curl and Diff features based on the GelSight’s black dots displacement. To the best of our knowledge, achieving stable placement with proposed Curl and Diff is a novel and non-trivial achievement.

  • The use of these features yielded a nearly 100% success rate of stable placing, achieving a consistently high rate across various objects with less than a 1-degree error. Despite the apparent simplicity of our proposed method, it is versatile enough to be applied to many objects.

II Related Works

II-A Object Pose Estimation

When considering a placing task, a typical first step is the use of object pose estimation methods. Object pose estimation is a well-studied problem in computer vision and is important for robotic tasks. While many approaches are image-based, they often face issues associated with occlusion and coarse accuracy [7, 8]. It is worth noting that there exist methods that reduce uncertainty and increase object pose estimation accuracy by having the grasped object contact other objects [9, 10, 11]. However, these require prior knowledge of the object’s shape, limiting their use for arbitrary objects.

Research utilizing tactile sensors has demonstrated their ability to estimate pose with greater accuracy [3, 12]. In particular, vision-based tactile sensors, such as GelSight, have been used in recent research, and some have achieved success in tasks with milli-order to sub-milli-order accuracy [3, 5, 4, 13, 14]. However, these methods only estimate an object’s relative or absolute pose and do not consider whether the object can be placed stably on a desk or other surface. If there is an error in the estimation result, the object may be unstable and may topple over. Therefore, the focus of this paper is the stable placing of objects.

II-B Object Placing

Object placing can be approached from both hardware and software perspectives. A well-known method for the hardware approach is to incorporate passive compliance in the robot’s wrist [15]. This mechanical compliance allows the robot to adjust to its environment and achieve stable placing without intricate control manipulations. However, hardware approaches are often specialized for specific objects, making them less flexible in accommodating a variety of objects.

As for the software approach, conventional methods involve control of applied force using robot joint torque or F/T sensors [1]. Yet, as discussed in the introduction, these methods often require substantial engineering effort to address problems like cable tension and sensor noise.

Several studies have proposed the use of tactile sensors, either solely or combined with F/T sensors, to estimate and control the force and torque on an object [14, 16, 5, 6]. These studies can be broadly divided into learning-based and analytical approaches. The learning-based approach, such as [14, 16], predicts where and how much external forces are applied to an object and performs object manipulation based on these predictions. In [5], a policy for controlling the object pose directly from raw images of GelSight is proposed. These studies indicate that tactile sensors can be used as F/T sensors, either directly or indirectly, by learning from raw images of vision-based tactile sensors and flow images showing tactile sensor displacement. While these approaches can generally adapt to various objects, the prerequisite of extensive training can be time-consuming and may result in out-of-distribution problems if data is insufficient.

In an analytical approach, [6] successfully measured the force and torque applied to a single GelSight sensor by using the Helmholtz-Hodge Decomposition to calculate the relationship between multiple forces and torques applied to different objects and GelSight’s black dots displacement. This method does not require the collection of large-scale data on various objects. However, to fully capture the potential of these approaches, a conversion to force and torque is attempted, which may necessitate calibration with dedicated equipment. In our work, we propose an analytical approach that avoids the need for calibration and shows promise in achieving stable object placing across various objects.

Refer to caption
Figure 2: Robot setup and coordinate system used for the experiment

III Preliminaries

As Preliminaries, this section describes the estimation of the direction of corrective rotation of an object when the object is placed unstably by pressing the object against a desk. Then, the robot control method for stable placing based on sensor information is described.

Three coordinate systems used in this study are described below (Fig. 2). The first one is the robot coordinate system, which is set at the base of the robot arm. The second is the end-effector (EE) coordinate system, which is set at the center of the two robot gripper’s fingertips. The robot manipulates objects by applying force and torque to them for stable placing. The point where the robot arm generates force is the origin of the EE coordinate system. In this paper, ‘the point where the robot arm generates force’ is referred to as the force application point. The third is the tactile sensor coordinate system, in which the robot coordinate system is rotated around the z-axis so that the x-axis direction of the tactile sensor coordinate system is parallel to the x-axis direction of the EE coordinate system.

Refer to caption
Figure 3: The estimation of the direction of corrective rotation of an object when the object is pressed against a desk.

III-A Estimation of Direction of Corrective Rotation

The idea behind the estimation of the direction of corrective rotation can be explained by considering the moment around the point of contact between the object and the environment [17]. The direction of corrective rotation is determined based on whether the force application point is to the right or left of the vertical line from the object’s contact point (Fig 3 (1)). The direction of corrective rotation is clockwise when the force application point is on the right side of the vertical line and counterclockwise when the force application point is on the left. If the force application point is on the vertical line, the object does not rotate.

The sign of the torque at the force application point indicates the direction of corrective rotation. The torque is positive, indicating clockwise rotation, if the force application point is to the right side of the vertical line; negative, for counterclockwise rotation, if on the left. We note that depending on the convention used, positive torque can indicate either clockwise or counterclockwise rotation but should stay consistent across future experiments. If the torque is zero, the force application point is on the line, indicating that the object does not rotate. Based on the sign of this torque detected by sensors, the robot is controlled to perform stable placing of the object. This control method of the robot is described in Section III-B.

Sensors measure these torques to estimate the direction of corrective rotation. While the method using the F/T sensor (referred to in this paper as the F/T sensor-based method, which is the baseline for this paper) can estimate the direction of corrective rotation, it is prone to inaccuracies primarily due to the cable tension and sensor noise. Especially near the vertical line, misrecognition of the direction of corrective rotation may occur as the measured sign of torque is likely to invert. Our proposed method uses tactile sensors to overcome these issues, providing a more accurate estimation of the direction of corrective rotation. Although tactile sensors measure different physical quantities than torque sensors, their measurements can still be interpreted as positive or negative with respect to the vertical line, providing a consistent approach to the estimation of the direction of corrective rotation.

III-B Robot Control for Stable Placing

The previous section described that the direction of corrective rotation of the object can be estimated based on the sign of the torque. This section describes how to control the robot for stable placing based on the torque.

If the torque measured at the force application point is not zero, the object, which is in an unstable state, can be rotated in the direction of corrective rotation. Stable placing can be performed by controlling the object by applying torque to the force application point so that the torque becomes 0. We make assumptions for our model, which are detailed below: As shown in Fig. 3 (2), the force application point is in the area where the object can be placed stably. We note that the area where this stable placement is feasible is not constant but varies depending on the position of the vertical line from the object’s contact point at the time the object is unstably placed. Under this assumption, manipulating the object so that the detected torque is zero results in a stable placing.

To realize stable placing, the robot arm is controlled by admittance control so that force Ftarget=(Ftargetx,Ftargety,Ftargetz)=(0,0,Ftz)F_{target}=(F_{target-x},F_{target-y},F_{target-z})=(0,0,F_{tz}) and torque τtarget=(τtargetx,τtargety,τtargetz)=(0,0,0)\tau_{target}=(\tau_{target-x},\tau_{target-y},\tau_{target-z})=(0,0,0) for the force application point exerts the force and torque in the robot coordinate system. The combination of force and torque is called a wrench. Let wtarget=(Ftarget,τtarget)w_{target}=(F_{target},\tau_{target}) be the target wrench and wcurrent=(F,τ)w_{current}=(F,\tau) be the current wrench. For this target wrench, when the object is pushed against the desk, the reaction force from the desk is FtzF_{tz}, and the other forces and torques are 0. In this experiment, Ftz=5NF_{tz}=5\mathrm{\,N}.

The end-effector velocity of the robot arm, veev_{ee} from the EE coordinate system, is calculated as follows to achieve the target wrench wtargetw_{target}:

vee=k(wcurrentwtarget)v_{ee}=k(w_{current}-w_{target}) (1)

where kk is a gain. In this paper, the same value for this gain is used for all objects.

In the case of the F/T sensor-based approach, the current wrench wcurrentw_{current} uses an F/T sensor mounted on the wrist. Let’s denote the force measured by the F/T sensor on the wrist as FF^{\prime} and the torque as τ\tau^{\prime}. The force FF and torque τ=(τx,τy,τz)\tau=(\tau_{x},\tau_{y},\tau_{z}) at the force application point can be calculated as follows:

F\displaystyle F =RF\displaystyle=R\cdot F^{\prime}
τ\displaystyle\tau =Rτ+d×(RF)\displaystyle=R\cdot\tau^{\prime}+d\times(R\cdot F^{\prime}) (2)

Here, RR is a 3x3 rotation matrix, and dd is a 3x1 displacement vector from the F/T sensor coordinate to the EE coordinate in the F/T sensor coordinate.

In our method, we use the Curl and Diff features values calculated by GelSight, instead of the τx\tau_{x} and τy\tau_{y} values for roll and pitch that are calculated from the F/T sensor.

IV Method: Tactile-based Placing

The previous section described that the direction of corrective rotation can be estimated based on the sign of the torque. This section describes how to use tactile sensors as an alternative to F/T sensors to estimate the direction of corrective rotation.

Our observations of how tactile sensors respond to various objects during unstable placement suggest that the displacement patterns of GelSight’s black dots can be categorized into two types with two directions of corrective rotation: roll of the tactile sensor coordinate system and pitch of the EE coordinate system (Fig. 1). The following section describes how to calculate the rotations for pitch and roll based on the displacement of the black dots on the GelSight.

Refer to caption
Figure 4: Schematic of torque applied to fingertip

IV-A Calculation of Rotation in Pitch Direction by Curl

We describe a method where the Curl features calculated from the displacement of the black point of the tactile sensor, used as an alternative to τy\tau_{y} (pitch rotation at the force application point) calculated from the F/T sensor by eq. 2.

When the object is pressed against the desk, a torque is generated that attempts to rotate the object (Fig. 4 (1)). This torque is transmitted to the GelSight and results in a shift in the position of the black dots on the GelSight, which represents the applied torque. This torque at the contact point on GelSight can be observed as the displacement of the black dots on GelSight (Fig. 1 (1) Pitch rotation). From the red arrows in Fig. 1 (1), when the left side of the object contacts the desk (with the gripper considered as the center), we observe a displacement of the arrows representing the black dots on the GelSight moving in a specific rotational direction. If the right side of the object is in contact instead, the arrows exhibit the opposite rotational movement. Specifically, in the former case, the arrows move in a clockwise direction, whereas in the latter, they move in a counterclockwise direction.

When the displacement of these arrows, which represent the movement of the black dots on the GelSight, becomes zero, the object is in a stable placing state with respect to the pitch rotation. To achieve a stable placement of the object by minimizing the torque, the robot should rotate around the force application point in the direction indicated by these arrows, which in turn reduces the displacement of the black dots. We note that the contact point is the fingertip, and the force application point is the middle of the contact point between the two fingers. Thus, this rotation center is considered the origin of the EE coordinate system.

From the arrows, the direction of rotation and its degree of rotation are calculated by Curl of vector analysis. The Curl is indicative of the rotational field magnitude and direction of the displacement of the black dots pattern. The displacements of the black dots on the GelSight give a vector field Atactile=(Ax,Az)A_{tactile}=(A_{x},A_{z}). In the EE coordinate system, the Curl of this field can be computed as follows:

CurlAtactile=Az/xAx/z\text{Curl}\>A_{tactile}=\partial A_{z}/\partial x-\partial A_{x}/\partial z (3)

We use the mean of the calculated Curl of all black dots as an alternative to τy\tau_{y} calculated from the F/T sensor by eq. 2, the torque at the force application point. The goal is to control the end-effector speed by eq. 1 of the robot so that this value becomes 0.

IV-B Calculation of Rotation in Roll Direction by Diff

We describe a method where the Diff features calculated from the displacement for the pitch axis between the left and right fingers’ GelSight’s black dots, used as an alternative to τx\tau_{x} (roll rotation at the force application point) calculated from the F/T sensor by eq. 2.

When the object is pressed against the desk, a torque is generated that attempts to rotate the object (Fig. 4 (2)). This torque is transmitted to the GelSight, and is exerted as shear forces in opposite directions on the right and left GelSight. This makes a difference in the displacement of the black dots in GelSights between the left and right fingers. The shear force arising at the contact point on GelSights can be observed by the displacement of the black dots on the GelSights (Fig. 1 (2) Roll rotation). Additionally, there are also forces generated by the action of pushing the object against the table that is exerted on the GelSights, whose magnitudes may change the overall direction of displacement. Fig. 1 (2) demonstrates the displacement of the black dot of GelSight when the tilted object makes contact with the desk. On the side where the object is in contact, the displacement of the black dot is represented as an arrow pointing vertically upward (in the z-axis direction of the tactile sensor coordinate system). This displacement is noticeably larger compared to the side where the object is not making contact.

When there is no longer any imbalance between the left and right of the displacement of the black dot of GelSight, the object is in a state of stable placing with respect to the direction of rotation of the roll. In order to reduce the imbalance of the left and right of the displacement of the black dot of GelSight, the robot should rotate around the force application point with respect to the direction of the object’s rotation. We note that the contact point is the fingertip, and the force application point is the middle of the contact point between the two fingers. Thus, this rotation center is considered the origin of the EE coordinate system.

In this study, we use two GelSight sensors, GelSight1 and GelSight2. Considering the black dots’ displacement in the z-axis direction of the tactile sensor coordinate system (normal facing upwards to the desk) helps calculate the GelSights’ displacement affecting the roll axis. This makes it possible to be irrespective of the pose of the robot and GelSights. We compute the average displacement values of the black dots on each of the two GelSights along the z-axis of the tactile sensor coordinate system. We denote these values as A¯z1\bar{A}_{z1} and A¯z2\bar{A}_{z2} respectively. We define a metric Diff as the difference between A¯z1\bar{A}_{z1} and A¯z2\bar{A}_{z2}:

Diff=A¯z1A¯z2Diff=\bar{A}_{z1}-\bar{A}_{z2} (4)

Diff serves as the torque at the force application point in the tactile sensor coordinate system. The goal is to control the robot’s end-effector speed to reduce Diff to zero. We note that the end-effector speed calculated from Diff corresponds to the speed measured from the tactile sensor coordinate system. However, as τy\tau_{y} in eq. 2 is calculated in the EE coordinate system, we must convert the value of Diff from the tactile sensor to the EE coordinate system. This can be achieved using standard coordinate transformation methods.

V Experimental Setup

V-A Robot Setup

Our robotic system, as shown in Fig. 2, features a 7-DoF Franka Emika Panda Arm. The end-effector command to the robot is updated at 200Hz200\mathrm{\,Hz}. This robotic arm is equipped with a custom-designed parallel gripper. The gripper features vision-based tactile sensors, known as GelSight Mini [18]. The sampling rate of GelSights is 10Hz10\mathrm{\,Hz}. A Leptrino force/torque sensor (FFS055YA501U6) is mounted between the robot arm and the gripper. The rated capacity of force of F/T sensor is Fxyz=±500NF_{xyz}=\pm 500\mathrm{\,N}, torque is Txyz=±4NmT{xyz}=\pm 4\mathrm{\,Nm}, resolution is ±1/2000\pm 1/2000, and sampling rate is 200Hz200\mathrm{\,Hz}. An Intel RealSense Depth Camera D435i is used to capture the pose of the arUco marker [19]. The system runs on a PC equipped with 32 GB RAM and an Intel Core i7-8700 CPU, running Ubuntu 20.04.6 LTS with ROS Noetic.

Refer to caption
Figure 5: Experiment process for object placing
Refer to caption
Figure 6: Target objects used in the experiment and GelSight images while grasping the target objects

V-B Task Setup and Process of Object Placing

We conduct experiments where a robot grasps objects and places them on a desk to evaluate our proposed system. To aid evaluation, an arUco marker is placed on the object, and the object’s pose is measured when the object is placed on the desk. The displacement is then calculated by comparing the object’s pose initially in its manually assigned stable state on the table with the final pose achieved through the robot’s stable placing process. The detailed steps of the experimental procedure are graphically illustrated in Fig. 5.

This experiment begins by manually placing an object with an attached arUco marker of size 25mm25~{}\mathrm{\,mm} onto a desk. We note that the experiment is repeated 10 times for each of the proposed and baseline methods but for the same object placed in the same location. The robot moves to a predetermined pre-grasping position (Fig. 5 (1)). The pre-grasping positions are specific to each object.

The robot grasps the object, and then, the camera (RealSense) in front of the robot (Fig. 2), with a sampling rate of 30Hz30\mathrm{\,Hz}, is used to capture the pose of the arUco marker (Fig. 5 (2)). The initially placed object is defined as being in a stably placed state in this experiment. To ensure accuracy and reduce potential noise during marker recognition, we use the average pose from the marker across 90 static images captured by the camera over three seconds.

After capturing the arUco marker, the robot lifts it off the ground by moving along the z-axis, and tilts it (Fig. 5 (3)-(4)). To evaluate the stable placing, the grasped object is tilted (either fixed or randomly sampled) in this pre-placing process. The way the tilt is specified is different for each experiment and will be described in the section VI. In the tilted state, the value read from the F/T sensor is recorded as an offset. During control procedures, we will use the value derived from the F/T sensor after removing this offset. The displacement of the black dots of the tactile sensor is calculated from the position of the black dots of the tactile sensor at this stage.

The robot then moves under velocity control at 0.005m/s0.005~{}\mathrm{\,m/s} in the -z-axis direction in the robot coordinate system (Fig. 5 (5)). When the force value of the F/T sensor along the z-axis of the robot coordinate system exceeds 1N1\mathrm{\,N}, the process of stable placing begins using both the proposed and baseline methods (Fig. 5 (5)).

After sufficient time has passed to ensure the completion of stable placing, the placing process is stopped, and the pose of the arUco marker is captured again (Fig. 5 (6)). As with the acquisition of the arUco marker taken in the beginning, the mean value of the 6d pose of the arUco marker was calculated from 90 static images captured over three seconds. Under the control of stable placing, the displacements in the roll and yaw axes of the arUco marker are evaluated. This evaluation is done by comparing the marker’s pose before and after the stable placing procedure is completed (Fig. 2).

Finally, the robot releases the object and returns to the home position. We performed this experimental process ten times for each object and calculated the mean and variance of the displacements of the arUco marker at that time.

V-C Target Objects

For the evaluation, we chose 18 objects considering various factors such as size (lengths of the short and long axes and thickness), softness, surface friction, texture, left-right asymmetrical shape and texture, and whether the object had a flat contact surface or not (Fig. 6).

VI Experiment Results

This evaluation focused on two key aspects: 1) Evaluation of stable placing, considering changes in the object’s grasp positions and the size of its support polygon - the area defined by the contact points between the object and the surface upon which it rests. 2) Comparison of the proposed method, Tactile-based method, with baseline, F/T sensor-based method across a variety of objects.

TABLE I: The accuracy of stable placing when the grasping position and the size of the support polygon of the object are different.
Object Grasp height Tactile-based (Ours) F/T-based (Baseline)
Roll [deg] Pitch [deg] All [deg] <1<1 deg <2<2 deg Roll [deg] Pitch [deg] All [deg] <1<1 deg <2<2 deg
Large rectangular 1.0cm1.0~{}\mathrm{\,cm} 0.4±0.20.4\pm 0.2 0.2±0.20.2\pm 0.2 0.3±0.20.3\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} 0.4±0.20.4\pm 0.2 0.2±0.20.2\pm 0.2 0.3±0.20.3\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10}
2.5cm2.5~{}\mathrm{\,cm} 0.4±0.30.4\pm 0.3 0.2±0.10.2\pm 0.1 0.3±0.20.3\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} 0.3±0.20.3\pm 0.2 0.3±0.20.3\pm 0.2 0.3±0.20.3\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10}
Small rectangular 1.0cm1.0~{}\mathrm{\,cm} 0.3±0.10.3\pm 0.1 0.1±0.10.1\pm 0.1 0.2±0.20.2\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 6/106/10 6/106/10
2.5cm2.5~{}\mathrm{\,cm} 0.3±0.20.3\pm 0.2 0.1±0.10.1\pm 0.1 0.2±0.20.2\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 0/100/10 0/100/10
TABLE II: The accuracy of stable placing.
Object type Object Tactile-based (Ours) F/T-based (Baseline)
Roll [deg] Pitch [deg] All [deg] <1<1 deg <2<2 deg Roll [deg] Pitch [deg] All [deg] <1<1 deg <2<2 deg
Elongated shape Small rectangular 0.4±0.20.4\pm 0.2 0.1±0.10.1\pm 0.1 0.2±0.20.2\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 2/102/10 2/102/10
Large rectangular 0.4±0.30.4\pm 0.3 0.3±0.20.3\pm 0.2 0.4±0.30.4\pm 0.3 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} 0.4±0.20.4\pm 0.2 0.4±0.30.4\pm 0.3 0.4±0.20.4\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10}
Wooden rectangular 0.4±0.20.4\pm 0.2 0.2±0.10.2\pm 0.1 0.3±0.20.3\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 7/107/10 7/107/10
Wooden cylinder 0.2±0.20.2\pm 0.2 0.1±0.10.1\pm 0.1 0.2±0.10.2\pm 0.1 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 6/106/10 7/107/10
Beaker 0.2±0.20.2\pm 0.2 0.3±0.10.3\pm 0.1 0.3±0.20.3\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 4/104/10 4/104/10
Metal rectangular 0.2±0.10.2\pm 0.1 0.2±0.10.2\pm 0.1 0.2±0.10.2\pm 0.1 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} 0.3±0.20.3\pm 0.2 0.1±0.10.1\pm 0.1 0.2±0.20.2\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10}
V shape V block 0.4±0.30.4\pm 0.3 0.2±0.10.2\pm 0.1 0.3±0.20.3\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 1/101/10 1/101/10
Joint 0.7±0.50.7\pm 0.5 0.2±0.10.2\pm 0.1 0.4±0.50.4\pm 0.5 𝟖/𝟏𝟎\bf{8/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 0/100/10 0/100/10
Asymmetry LEGO blocks 0.3±0.20.3\pm 0.2 0.9±0.10.9\pm 0.1 0.6±0.30.6\pm 0.3 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 6/106/10 6/106/10
Metal object 0.2±0.20.2\pm 0.2 0.2±0.10.2\pm 0.1 0.2±0.20.2\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 0/100/10 0/100/10
Non-surface contact Metal pedestal 0.2±0.20.2\pm 0.2 0.2±0.10.2\pm 0.1 0.2±0.10.2\pm 0.1 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} 0.2±0.20.2\pm 0.2 0.1±0.10.1\pm 0.1 0.2±0.10.2\pm 0.1 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10}
Conversion connector 0.4±0.20.4\pm 0.2 0.2±0.10.2\pm 0.1 0.3±0.20.3\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 2/102/10 2/102/10
Soft material Wood bond tube 0.3±0.20.3\pm 0.2 0.2±0.10.2\pm 0.1 0.2±0.20.2\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 1/101/10 1/101/10
Hand cream tube 0.4±0.30.4\pm 0.3 0.4±0.20.4\pm 0.2 0.4±0.20.4\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 4/104/10 4/104/10
Fewer contact Test tube stand 0.1±0.10.1\pm 0.1 0.6±0.20.6\pm 0.2 0.3±0.30.3\pm 0.3 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} 0.2±0.20.2\pm 0.2 0.5±0.20.5\pm 0.2 0.3±0.20.3\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10}
Metal part 0.4±0.20.4\pm 0.2 0.1±0.10.1\pm 0.1 0.3±0.20.3\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} 0.3±0.20.3\pm 0.2 0.1±0.10.1\pm 0.1 0.2±0.20.2\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10}
Round shape Round bottom flask 0.4±0.20.4\pm 0.2 0.2±0.20.2\pm 0.2 0.3±0.20.3\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 0/100/10 0/100/10
Liquid Bottle with liquid 0.4±0.30.4\pm 0.3 0.2±0.20.2\pm 0.2 0.3±0.20.3\pm 0.2 𝟏𝟎/𝟏𝟎\bf{10/10} 𝟏𝟎/𝟏𝟎\bf{10/10} - - - 4/104/10 5/105/10

VI-A Different Grasp Positions and Support Polygon Size

We evaluate the influence of the object’s grasping position and the size of the object’s support polygon on stable placing. To investigate these effects, we utilize two rectangular objects, small rectangular and large rectangular, which have the same height and different sizes of support polygon (Fig. 6). The dimensions of small and large rectangular are, respectively, 30mm×30mm×70mm30~{}\mathrm{\,mm}\times 30~{}\mathrm{\,mm}\times 70~{}\mathrm{\,mm}, 50mm×50mm×70mm50~{}\mathrm{\,mm}\times 50~{}\mathrm{\,mm}\times 70~{}\mathrm{\,mm} in terms of width, depth, and height. These two objects were made by 3D printing, and they both weigh 142g142\mathrm{\,g}. The two grasping positions are at heights of 1.0cm1.0~{}\mathrm{\,cm} and 2.5cm2.5~{}\mathrm{\,cm} from the desk. To compare under the same noise conditions, the stable placing experiment, following the process outlined in section V-B, was conducted 10 times with the object tilted 10 degrees to the roll and pitch in the tactile-sensor coordinate in the pre-placing process (Fig. 5 (4)). Table I presents the mean and standard deviation of the roll and pitch misalignment and their combined values from ten iterations of stable placing experiments. However, when the error magnitude exceeds 2 degrees, we do not compute the mean and standard deviation as the means appear exceedingly large when the object topples over. We also report the number of times when roll and pitch errors were below 1 and 2 degrees. We chose the thresholds of 1 degree and 2 degrees based on our experimental considerations. Specifically, the 1-degree threshold was selected due to the detection accuracy of the arUco marker. The 2-degree threshold was chosen because our experimental experience suggested that exceeding this value often leads to the object toppling over.

Using our proposed tactile-based method, stable placing was achieved under all conditions regardless of the grasping position and the size of the support polygon. In the case of the baseline F/T-based method, for the large rectangular object, stable placing was achieved at both grasp heights. For the small rectangular object, while there were six successful outcomes with a low grasping height (1.0cm1.0~{}\mathrm{\,cm}), there were no successes when the grasping height was high (2.5cm2.5~{}\mathrm{\,cm}).

The reasons for this are explained below. The direction of corrective rotation can be estimated through the sign of the torque. For a rectangular object, as illustrated in Fig. 3 (2), bias in estimating the direction of corrective rotation depends on the grasping position and the size of the supporting polygon. A low grasping position or large support polygon may still enable accurate estimation despite sensor inaccuracies since the area in which stable placement can be achieved is larger. Conversely, a high grasping position or small support polygon decreases the area in which stable placing can be feasible and increases the risk of sensor inaccuracies negatively affecting the direction of corrective rotation estimation. F/T sensors’ measured values have low accuracy due to their noisy nature. We note that a low pass filter is used to remove noise from the F/T sensor, but noise that cannot be removed still has an effect. Therefore, even under the same conditions of conducting 10 stable placings with a 10-degree tilt in roll and pitch, failures may occasionally occur due to the inherent uncertainty of the F/T sensor. Additionally, cable tension can cause a shift in the torque value, which could potentially lead to misinterpretations of the object’s direction of corrective rotation.

Refer to caption
Figure 7: Examples of stable placing. For the sake of visibility, the length of the arrows is eight times the actual displacement.

VI-B Evaluation across Diverse Objects

We evaluate the proposed method and a baseline F/T-based method for the stable placing of 18 different objects. This stable placing experiment, which follows the process outlined in section V-B, was conducted 10 times with the object tilted. The tilt in the pre-placing process (Fig. 5 (4)) is calculated from a normal distribution with a mean of ±10\pm 10 degrees (and the direction of the tilt is uniformly sampled to be positive or negative) and a variance of 1 in both roll and pitch directions in the tactile-sensor coordinate. For objects prone to toppling over, we specify their tilts as follows: The conversion connector, which is tilted even in the stable placing state, has a normally distributed noise with a mean of ±10\pm 10 degrees in the roll direction and a mean of ±3\pm 3 degrees in the pitch direction, while the Joint object has a mean of -5 degrees for noise in the pitch direction and a mean of 10 degrees in the roll direction. For hand cream, which has a small support polygon, a mean of ±5\pm 5 degrees. The variance of tilt for all items tested is 1. The Table II shows the results.

In the Tactile-based method, stable placing with an error of less than 1 degree was achieved for all objects except for the Joint object. The method was successful for objects with asymmetrical shapes and textures, objects with a point of contact with the desk instead of a surface, soft objects, objects with a small size of support polygon, and objects with a changing center of gravity, such as liquids. Fig. 7 shows examples of the stable placing process using the proposed method. The Curl and Diff values are nearly zero for a variety of objects, and stable placement is achieved.

Additionally, the method was also robust against cracks in the sensor of the surface and manufacturing differences in the size and shape of the black dots. The image of the tactile sensor in Fig. 6 shows that the surface of the tactile sensor is cracked. The tactile sensor was completely damaged and was replaced by another GelSight, which was also handled with a large black dot due to individual differences in the GelSight (Metal object and Test tube stand in Fig. 7).

For the Joint object, while it did not topple over, an error of less than 2 degrees remained. We analyze this occurrence as follows: The displacement of the black dots on GelSight is calculated based on their positions before stable placing when noise is added to the object’s pose. In the Joint object experiment, the object was grasped at the tip of the V-shape, which was far from the center of gravity. As gravity causes the object to rotate, the black dots on GelSight deviate from their positions in the stable state. Even when the Joint object reaches a stable placement, some displacement of the black point remains present. Thus, the robot arm is controlled to apply force to the force application point to reduce the displacement of the black point. Unlike objects with a larger support polygon that retain their pose, the Joint object, with its smaller support polygon, appears to tilt due to the force intended for its rotation. However, the tilt is small, less than 2 degrees, and the risk of toppling over is small.

The Tactile-based method achieved high accuracy for a variety of objects, while the F/T-based method often failed. A common feature of the objects with a 100% success rate in the F/T-based method is that the support polygon is large, such as large rectangular, metal rectangular, metal pedestal, test tube stand, and metal part. Objects with small support polygon failed with probability or did not succeed at all. As explained in Section VI-A, the F/T sensor is affected by sensor noise and the cable tension, causing uncertainty in sensor values and misrecognizing the direction of rotation of the object. As a result, there were many failures that resulted in the object toppling over in the F/T sensor-based approach. There were also failures that the object’s tilt was kept after contact between the object and the desk. Moreover, some were seen where the robot arm applied force in the direction that the object was to be upright, but the object passed through the upright state and toppled over.

In summary, the F/T sensor method fails in stable placing due to sensor noise and the cable tension, causing deviations in the measured torque values, which result in misrecognition of the object’s direction of corrective rotation. On the other hand, the Tactile-based method succeeds in stable placing with high accuracy as a result of its precise estimation of the object’s direction of corrective rotation.

VII Conclusion

We propose a method for stable object placing using tactile sensors. Under unstable object placing conditions, we observed that the displacement pattern of the black dots in GelSight aligns with the two possible directions of corrective rotation. By estimating these two object directions of corrective rotation from Curl and Diff, the robot can manipulate the object’s pose for stable placing. This method of estimating the direction of corrective rotation is robust against sensor failures such as cracks in the surface, and manufacturing differences in the size and shape of the black dots. In experiments with 18 objects with different characteristics, the F/T sensor-based method fails because it misrecognizes the direction of rotation for objects other than those with a large support polygon due to sensor noise and cable tension. In contrast, the proposed Tactile sensor-based approach can achieve stable placing with high accuracy (less than 1-degree error) in nearly 100% of cases across all objects, highlighting its potential as an effective F/T sensor alternative.

ACKNOWLEDGMENT

The authors thank Avinash Ummadisingu and Dr. Naoki Fukaya for the many discussions about this research. This work was supported by JST [Moonshot R&D][Grant Number JPMJMS2033].

References

  • [1] M. H. Raibert and J. J. Craig, “Hybrid position/force control of manipulators,” 1981.
  • [2] W. Yuan, et al., “GelSight: High-Resolution Robot Tactile Sensors for Estimating Geometry and Force,” Sensors, vol. 17, no. 12, p. 2762, 2017.
  • [3] T. Anzai and K. Takahashi, “Deep gated multi-modal learning: In-hand object pose changes estimation using tactile and image data,” in 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2020, pp. 9361–9368.
  • [4] M. Bauza, et al., “Tac2pose: Tactile object pose estimation from the first touch,” The International Journal of Robotics Research, vol. 42, no. 13, pp. 1185–1209, 2023.
  • [5] S. Dong, et al., “Tactile-rl for insertion: Generalization to objects of unknown geometry,” in 2021 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2021, pp. 6437–6443.
  • [6] Y. Zhang, et al., “Effective estimation of contact force and torque for vision-based tactile sensors with helmholtz–hodge decomposition,” IEEE Robotics and Automation Letters, vol. 4, no. 4, pp. 4094–4101, 2019.
  • [7] J. Sun, et al., “Onepose: One-shot object pose estimation without cad models,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 6825–6834.
  • [8] B. Li, et al., “VoxDet: Voxel Learning for Novel Instance Detection,” in Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), 2023.
  • [9] N. Chavan-Dafle and A. Rodriguez, “Stable prehensile pushing: In-hand manipulation with alternating sticking contacts,” in 2018 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2018, pp. 254–261.
  • [10] F. Von Drigalski, et al., “Contact-based in-hand pose estimation using bayesian state estimation and particle filtering,” in 2020 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2020, pp. 7294–7299.
  • [11] J. Pankert and M. Hutter, “Learning Contact-Based State Estimation for Assembly Tasks,” in 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2023, pp. 5087–5094.
  • [12] L. Lach, et al., “Placing by Touching: An empirical study on the importance of tactile sensing for precise object placing,” in 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2023, pp. 8964–8971.
  • [13] S. Pai, et al., “Laboratory Automation: Precision Insertion with Adaptive Fingers utilizing Contact through Sliding with Tactile-based Pose Estimation,” arXiv preprint arXiv:2309.16170, 2023.
  • [14] K. Ota, et al., “Tactile Estimation of Extrinsic Contact Patch for Stable Placement,” arXiv preprint arXiv:2309.14552, 2023.
  • [15] D. E. Whitney et al., “Quasi-static assembly of compliantly supported rigid parts,” Journal of Dynamic Systems, Measurement, and Control, vol. 104, no. 1, pp. 65–77, 1982.
  • [16] C. Higuera, et al., “Neural contact fields: Tracking extrinsic contact with tactile sensing,” in 2023 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2023, pp. 12 576–12 582.
  • [17] M. Shimizu and K. Kosuge, “Spatial parts mating with fiction using structured compliance with compliance center,” in IEEE/RSJ International Conference on Intelligent Robots and Systems, vol. 2.   IEEE, 2002, pp. 1585–1590.
  • [18] “Gelsight mini,” https://www.gelsight.com/gelsightmini/, 2024.
  • [19] S. Garrido-Jurado, et al., “Automatic generation and detection of highly reliable fiducial markers under occlusion,” Pattern Recognition, vol. 47, no. 6, pp. 2280–2292, 2014.