Next Article in Journal
A Review on Concrete Structural Properties and Damage Evolution Monitoring Techniques
Next Article in Special Issue
Hardware Schemes for Smarter Indoor Robotics to Prevent the Backing Crash Framework Using Field Programmable Gate Array-Based Multi-Robots
Previous Article in Journal
A 3D Point Cloud Classification Method Based on Adaptive Graph Convolution and Global Attention
Previous Article in Special Issue
Hardware-Efficient Scheme for Trailer Robot Parking by Truck Robot in an Indoor Environment with Rendezvous
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robot Grasp Planning: A Learning from Demonstration-Based Approach †

1
FANUC Advanced Research Laboratory, FANUC America Corporation, Union City, CA 94587, USA
2
Department of Precision Engineering, The University of Tokyo, Tokyo 113-8654, Japan
*
Author to whom correspondence should be addressed.
This paper is an extension version of the conference paper: Wang, K.; Fan, Y.; Sakuma, I. Robot Grasp Planning from Human Demonstration. In Proceedings of the 2023 15th International Conference on Computer and Automation Engineering (ICCAE), Sydney, Australia, 3–5 March 2023.
Submission received: 3 December 2023 / Revised: 31 December 2023 / Accepted: 15 January 2024 / Published: 18 January 2024
(This article belongs to the Special Issue Novel Sensors and Algorithms for Outdoor Mobile Robot)

Abstract

:
Robot grasping constitutes an essential capability in fulfilling the complexities of advanced industrial operations. This field has been extensively investigated to address a range of practical applications. However, the generation of a stable grasp remains challenging, principally due to the constraints imposed by object geometries and the diverse objectives of the tasks. In this work, we propose a novel learning from demonstration-based grasp-planning framework. This framework is designed to extract crucial human grasp skills, namely the contact region and approach direction, from a single demonstration. Then, it formulates an optimization problem that integrates the extracted skills to generate a stable grasp. Distinct from conventional methods that rely on learning implicit synergies through human demonstration or on mapping the dissimilar kinematics between human hands and robot grippers, our approach focuses on learning the intuitive human intent that involves the potential contact regions and the grasping approach direction. Furthermore, our optimization formulation is capable of identifying the optimal grasp by minimizing the surface fitting error between the demonstrated contact regions on the object and the gripper finger surface and imposing a penalty for any misalignment between the demonstrated and the gripper’s approach directions. A series of experiments is conducted to verify the effectiveness of the proposed algorithm through both simulations and real-world scenarios.

1. Introduction

In recent years, there has been a notable escalation in the demand for industrial robots, driven by the need to enhance productivity and boost product quality. With the continuous advancements in robotics technology, the trend toward increased robot installations has been significantly pronounced in sectors such as logistics and assembly production. Among the broad spectrum of robotics research, the capability of robot grasping stands out as a fundamental skill that is crucial for the execution of complex tasks in industrial environments [1]. Despite being a subject of research for numerous decades, robot grasp planning in the context of industrial automation presents persistent challenges due to the lack of understanding of the constraints that are inherent to the downstream processes in industrial operations [2].
To address this challenge, learning from demonstration (LfD) has emerged as a promising approach that synthesizes robot grasping by observing human demonstrations. LfD offers an intuitive and straightforward method for non-experts in robotics to program a new task on a robot. Despite significant research efforts dedicated to efficiently transferring human grasping skills to robots, substantial challenges remain due to the low complexity and dexterity of the robot gripper compared to the human hand, as well as the diverse array of object shapes encountered in practical applications. Recent advancements in data-driven methods have made remarkable progress in the extraction of grasping skills by leveraging deep learning techniques. However, these methods typically require a large amount of data for training, leading to time-consuming and labor-intensive processes [3]. On the other hand, traditional analytic approaches formulate grasp planning as an optimization problem, focusing on achieving a stable grasp. The combination of human intention with optimization techniques enables robots to execute functional grasps while being mindful of constraints emanating from downstream processes. Existing methods typically represent human grasping skills through a contact model and then search for an optimal grasp that aligns with the corresponding contact model based on prior knowledge of object shapes and gripper configurations [4]. However, a notable limitation of many current approaches is their reliance on pre-defined grasping taxonomies, which aim to bridge the gap in configuration differences between human hands and robot grippers. This pre-defined grasping taxonomy hinders the ability to generalize across a broad spectrum of industrial tasks.
In this work, we propose a novel LfD-based grasp-planning approach, which incorporates human grasp skills into an optimization framework to synthesize a stable grasp. This method revolves around extracting key elements of human grasp skills, namely the contact region and the approach direction, from a single human demonstration. The key idea of our approach is to account for high-level task knowledge while addressing the difference in configuration between human hands and robot grippers. More specifically, the proposed method utilizes contact regions as a foundational element and applies optimization to reduce the surface fitting error between the robot gripper fingers and the object. To achieve a functional grasp, both the contact region and the approach direction are crucial, as often observed in human grasp skills [5]. To extract this information from human demonstrations, we propose an intuitive and straightforward method that involves passing a plane formed by the thumb and index finger through the object. The obtained grasp skills are then fed into an optimization algorithm to generate a grasp that closely approximates the human demonstration. In summary, our main contributions are:
  • We propose a novel LfD-based grasp-planning framework that utilizes both the contact regions and approach direction as key skills derived from a single human demonstration. This approach effectively harmonizes human intention with the environmental constraints of both the objects and the robot gripper.
  • We develop an intuitive and straightforward method for detecting the contact regions and approach direction by employing a specific hand plane formed by the thumb and index fingers using a single RGB-D camera.
  • We integrate human grasp skills, encompassing both the contact regions and approach direction, into an optimization problem. This integration aims to generate a stable functional grasp that aligns with human intention while considering environmental constraints.
We note that a short conference version of this work was presented in [6]. Our initial conference paper did not delve into specific challenges such as collision avoidance and internal grasp complexities. This manuscript addresses these critical aspects, supplemented by additional analytical experiments to provide deeper insights.

2. Related Works

Data-driven methods: The data-driven approach in robot grasping has gained considerable traction due to its ability to learn grasp features from a large amount of data by leveraging deep learning techniques [7]. Benefiting from robustness to perception errors and the capacity to generalize to unseen objects, this approach has been successfully deployed in various applications, notably in bin-picking scenarios [8]. However, the collection of large volumes of training data is a costly and resource-intensive endeavor. Additionally, generating a feasible functional grasp in complex industrial settings often requires an understanding of environmental constraints and downstream procedures, an aspect that data-driven methods may struggle to incorporate effectively. In response to the challenges associated with data collection, Reinforcement Learning (RL) methods have emerged as a potential solution. RL methods enable robots to automatically learn grasping policies through self-supervised interaction with the environment [9]. The advent of advanced computer graphics has further facilitated this process, allowing researchers to train grasping policies in simulation environments and then transfer them to real-world applications [10]. However, setting up training simulations can be daunting for novice users. Moreover, due to the difficulty in acquiring general skills applicable across a variety of tasks, these methods often require extensive retraining for different task configurations. In contrast, the proposed method synthesizes the functional grasp by observing human demonstration once, without requiring extensive data collection and time-consuming training processes.
Analytic methods: Analytic approaches in robot grasping search for a stable grasp by formulating and solving an optimization problem, utilizing prior knowledge of object shapes and gripper configurations [11]. These methods are suitable for well-structured industrial environments since they are low data-dependent and eliminate the need for tedious supervised or unsupervised training. Some analytic strategies represent a grasp as a set of contact points, sampling gripper poses in the configuration space to seek an optimal grasp. This process often employs various metrics, such as contact surface matching [12] or gripper configuration [13]. However, these approaches do not always guarantee that the generated grasps can successfully perform specific post-grasp motions in line with the downstream processes. Alternative approaches focus on developing a taxonomy of grasp types tailored for different manipulation purposes. This type of approach can initially assign a feasible functional grasp and enhance sample efficiency by reducing the complexity of the search space [14]. An in-depth analysis of grasp taxonomy for continuum robots was presented in [15]. Based on this grasp taxonomy, the study in [16] introduced an analytical grasp synthesis approach, allowing continuum robots to adapt grasping strategies to diverse objects and tasks. In [17], a comprehensive taxonomy of human grasp types was presented through an analysis of the data recorded from 40 healthy subjects performing 20 unique hand grasps. Extending this concept, ref. [18] proposed an approach to predict the best grasp type from a taxonomy of 33 grasp classes in multi-object scenarios using just a single RGB image. However, the taxonomy-based approach is heavily dependent on predefined shape priors of both grippers and objects, which can limit its generalization capabilities for unseen objects. In contrast, our approach directly transfers human grasp skills to the robot through observations of human demonstrations, without the need for a predefined taxonomy of grasp types, which enables a more flexible and adaptive approach to robot grasping.
Grasp synthesis by learning from demonstration: The remarkable grasping skills of humans, capable of handling a diverse range of objects in complex task contexts, have sparked significant interest in the field of robotics, particularly in the transfer of these skills to robots, as surveyed in [19]. To recognize the human grasping pose, many different methods have been discussed. Early research in the field predominantly relied on a data glove for the precise detection of human finger positions, which were then replicated by robotic grippers. A teleoperation system utilizing a data glove to control a multi-fingered robotic hand was developed in [20]. The integration of tactile sensors into this glove-based framework plays a crucial role in understanding how the hand interacts with different objects. In [21], a data glove equipped with multiple Inertial Measurement Units (IMUs) and tactile sensors was designed to capture the complex dynamics of hand–object interactions in real time. However, the complexity of data gloves, especially in integrating various sensors, presents significant challenges in terms of cost and maintenance. To simplify the system of human grasping demonstration, the study in [22] proposed the use of thermal sensors to trace contact areas. With the advent of deep learning technologies, more recent studies have begun to employ simple cameras to track 3D hand poses [23]. A learning from demonstration pipeline designed to infer the poses of both the human hand and the object and then transfer these relative poses into robot grasping instructions is discussed in [24]. The integration of human demonstrations into the initialization of grasping policies has been shown to significantly improve sample efficiency for both reinforcement learning and heuristic-based methods. This efficiency is achieved by limiting the search space to a more constrained domain. A critical element in learning from demonstration (LfD) approaches is the contact model between the human hand and the object, which was extensively explored in [25]. Some researchers measure accurate contact points between the fingertips and objects through the use of a specific data glove and then map the contact points onto the corresponding robot gripper fingers [26,27]. Although these methods provide an intuitive way to transfer human grasping poses to robots, the reliance on single-point approximations imposes rigid mapping constraints and sacrifices flexibility in accounting for the configuration differences between human hands and robot grippers. To relax the mapping constraints, recent research has shifted its focus toward contact regions, recognizing that human grasps typically involve surface matching between the hand and the object [28]. However, most of these approaches employ specialized equipment (e.g., thermal cameras) or manual labeling to identify the contact regions, often overlooking the importance of the grasping approach direction, which could further narrow the search space for optimal grasping. In contrast, our method proposes a novel approach to detect both the contact regions and approach direction using a single RGB-D camera, a common device in industrial settings. In addition, the proposed optimization method formulates the surface fitting issue, aiming to minimize the distance between the demonstrated contact regions and gripper finger surface, as well as the misalignment between the demonstrated approach direction and the robot gripper, as a least-squares problem to accelerate the computation required for grasp searching.

3. Proposed Approach

The proposed method for robot grasp synthesis is structured into two stages: grasp skill recognition from human demonstration and grasp synthesis with the optimization method, as shown in Figure 1. The process starts with the identification of contact regions and approach directions from a human demonstration. This identification involves defining a hand plane formed by the thumb and one of the other four fingers [29]. The approach direction is determined by combining the directions of the thumb and index finger. The contact region on the object is then determined by intersecting this hand plane with the object’s surface and selecting proximate surface points. Following the extraction of the contact region and approach direction, the grasp optimization process, which iteratively fits the surface of the gripper fingers to the previously identified contact region and aligns the direction of the gripper approach with the human demonstration, is employed to determine the best grasping pose. The optimization is mathematically formulated as a least-squares problem, taking into account the grasp skills extracted from the human demonstration. Once a grasp pose is generated, it undergoes a critical evaluation to ensure it is collision-free. If a collision is detected, the gripper is moved to a safe pose, and the grasp optimization process is repeated to generate a new collision-free pose.

3.1. Grasp Skill Recognition

During the human demonstration phase, the movements of the human hand are tracked by leveraging deep learning techniques [30], employing a commercial camera. The research in [31] has proven that the majority of human grasp poses are formed by the thumb and some other opposing fingers. This finding has significant implications for robot gripper design, many of which mimic this fundamental grasp characteristic by incorporating a thumb finger and a set of opposing fingers. Therefore, in this work, the hand plane is defined by the thumb and one of the other four fingers. As shown in Figure 2, one hand plane is defined by the thumb and index fingers in the case of a parallel gripper, and two hand planes for a three-finger gripper are defined: one formed by the thumb and index fingers, and another formed by the thumb and middle fingers. The coordinate system of the hand plane is calculated based on the 3D positions of specified hand joints on the thumb and index fingers, as shown in Figure 3. The midpoint between the root joints of the thumb and index fingers is defined as the origin of the hand-plane coordinate system. The normal vector of the hand plane is defined as the X-axis, whereas the Y-axis is determined by the vector connecting the thumb and index root joints. The Z-axis is then established following the right-hand rule.
The approach direction, a crucial feature for functional grasp, indicates grasp reachability. Upon determining the hand plane, the Z-axis direction is selected as the approach direction, a concept similarly explored in [5]. Note that the approach direction is uniformly calculated using the hand plane formed by the thumb and index fingers, regardless of the number of fingers on the gripper.
Our method adopts a novel approach to determining the contact region for robot grasping. Instead of striving for pinpoint accuracy in detecting the positions where human fingertips contact the object surface, the hand plane is utilized to estimate a broader contact region on the object. This strategy not only simplifies the process but also enhances the robustness of our method, particularly in optimizing the grasping pose and accommodating the configuration differences between the human hand and the robot gripper.
Figure 4 exemplifies the process of identifying the contact region on an object. To calculate this contact region, we employ a two-step procedure. The first step involves passing the hand plane through the point cloud of an object. During this phase, all points on the object surface that fall within a specified tolerance distance, denoted as d 1 , from the hand plane are selected using Equation (1). These points serve as preliminary candidates for the contact region.
| A O x + B O y + C O z + D | A 2 + B 2 + C 2 d 1
where A ,   B ,   C , and D are the coefficients defining the hand plane A x + B y + C z + D = 0 . Each point within the object’s point cloud is denoted by its 3D coordinates O x , O y , and O z . The parameter d 1 is the first tolerance distance, which represents the permissible distance from any point in the object’s point cloud to the hand plane. This tolerance distance is determined based on the width of the gripper finger and subsequently guides the selection of the contact area. The results of the first step are shown in Figure 4a. The selected volume contains points from the object’s point cloud that fall within the tolerance distance d 1 from the hand plane.
In the second step, we refine the selection of potential contact points on the object that were initially identified in the first step. This refinement involves selecting only those points that are within a second tolerance distance, denoted as d 2 , from the origin of the hand plane using Equation (2). The purpose of this second tolerance distance is to ensure that the chosen contact regions on the object are not too far away from the demonstrated grasp pose.
( O x x 0 ) 2 + ( O y y 0 ) 2 + ( O z z 0 ) 2 d 2
where x 0 ,   y 0 ,   z 0 denotes the origin of the hand-plane coordinate, and the parameter d 2 is the second tolerance distance, which is determined based on the length of the gripper finger. The results of the second step are shown in Figure 4b. Only the points within the object’s point cloud, where the distance to the origin of the hand-plane frame does not exceed the specified threshold d 2 , are retained.

3.2. Grasp Synthesis

3.2.1. Grasp Optimization

The contact regions on the object that align with the human demonstration are identified and then employed in an optimization computation. It is generally observed that a larger contact surface between the gripper and the object results in greater friction, which, in turn, facilitates the formation of a more robust grasp. Leveraging this principle, an iterative surface fitting (ISF) optimization method, adapted from the Iterative Closest Point (ICP) algorithm, was proposed in [32].
Compared with the conventional ICP approach, the ISF method deforms the gripper fingertip surfaces while obeying kinematic constraints and concurrently aligns these modified surfaces with the target object’s surface. This method is more precisely formulated in Equation (3).
min R , t , δ d ( E p + λ E n ) s . t . E p ( R , t , δ d ) = i = 1 m j = 1 2 ( ( p i j q i j ) T n i j q ) 2 E n ( R ) = i = 1 m ( ( R n i p ) T n i q + 1 ) 2
where the rotation R and translation t are the transformation matrix from the robot gripper to the object, and  δ d is the displacement between the gripper fingers. The variables p i , q i , n i p , and n i q represent the points on the robot gripper surface, the corresponding point on the object surface, the normal vector at point p i on the gripper surface, and the normal vector at point q i on the object surface, respectively. The variable m is the total number of points, and  j = 1 , 2 represents the indexes of two fingers. According to [32], Equation (3) can be further formulated as a least-square problem min R , t , δ d A x b 2 .
The core objective of the ISF method is to search for an optimal transformation matrix (rotation R and translation t), as well as the displacement ( δ d ) of the gripper fingers, by minimizing the surface fitting error. In practical terms, the most intuitive approach involves initializing R , t in Equation (3) with the pose corresponding to the human grasp. This initialization forms the baseline from which the ISF algorithm iteratively computes the errors E p and E n to refine the fit to the contact surface. However, this method may diverge from the initial human grasp pose due to the difference in fingertip shape between human hands and robot grippers. To address this, we incorporate the approach direction derived from the human demonstration into the optimization process as a constraint in Equation (4). This incorporation ensures a more accurate and feasible replication of human-like grasping by the robot.
min R , t , δ d A x b 2 s . t . R n z = n a p p
where n z denotes the Z-axis direction of the gripper, and  n a p p represents the approach direction, determined from the human demonstration. The rotation matrix R is designed to align n z with n a p p .
To effectively solve this optimization problem while maintaining the integrity of the approach direction constraint, as defined by the human demonstration, Equation (4) is rewritten as follows:
min R , t , δ d A x b 2 + ω 2 ( R n z ) × n a p p 2
where ω 2 is a weight parameter to balance the surface matching accuracy and the alignment of the approach direction. Under the assumption that the rotation angle involved in each iterative step is small, it becomes feasible to approximate the rotation matrix R as follows:
R 1 δ ψ δ θ δ ψ 1 δ ϕ δ θ δ ϕ 1 = I + δ r ^
where δ r ^ s o ( 3 ) is the skew-symmetric matrix form of the rotation vector δ r ^ = [ δ ϕ , δ θ , δ ψ ] , which contains the small rotational changes in three dimensions—roll ( δ ϕ ), pitch ( δ θ ), and yaw ( δ ψ ). By substituting Equation (6), which characterizes the skew-symmetric matrix δ r ^ , into Equation (5), the second term of Equation (5), which concerns the alignment of the approach direction, can be rewritten as follows:  
ω 2 n a p p ^ · n z ^ · δ r + n z ^ · n a p p 2 = ω 2 E · δ r f 2 where E = n a p p ^ · n z ^ ,   f = n z ^ · n a p p
Consequently, by integrating Equation (7) into Equation (5), the optimal grasp pose can be calculated as
[ δ r , t ] = ( A T A + ω G T G ) 1 ( A T b + ω G T F ) where G = [ E T , 0 3 · 3 T ] T ,   F = [ f T , 0 3 T ] T
Using Equation (8), the optimization problem is formulated into a standard least-squares problem that can be solved efficiently. The optimal finger relative displacement is calculated using Equation (9), given by [32].
δ d = i = 1 m j = 1 2 K i j H i j i = 1 m j = 1 2 K i j 2 s . t . δ d [ d m i n , d m a x ]
where K i j = 0.5 ( 1 ) j 1 ( R v ) T n i j q , and  H i j = ( R p i j + t q i j ) T n i j q . d m i n and d m a x indicate the minimum and maximum distances between two fingers, and v is the direction vector of the finger movement.
The pseudo-code of the proposed LfD-based iterative surface fitting is summarized in Algorithm 1. The proposed method is designed to optimize for the optimal gripper transformation R , t with fixed δ d . Concurrently, the finger optimization focuses on determining the optimal finger displacement δ d while maintaining fixed values for R , t .
Algorithm 1 Learning from demonstration-based iterative surface fitting
1:
Input:  p i , n i p g r i p p e r , q i , n i q o b j e c t
2:
Output:  R , t , δ d
3:
Init:  R = R h a n d p l a n e , t = t h a n d p l a n e , δ d = 0 , e r r =
4:
while e r r E ( R , t , δ d ) > e r r do
5:
     R , t m i n R , t E ( R , t , δ d ) using Equation (8)
6:
     δ d m i n δ d E ( R , t , δ d ) using Equation (9)
7:
     e r r E ( R , t , δ d ) using Equation (3)
8:
end while
9:
return  R , t , δ d

3.2.2. Collision Avoidance

In the iterative surface fitting phase of robot grasp planning, it was observed that while the gripper fingers tend to align toward collision-free grasps, there remains a non-negligible risk of the gripper base colliding with the target object. To mitigate this issue and ensure that the entire gripper body avoids collision, we incorporate a collision avoidance mechanism within the grasp optimization framework.
An existing method in this domain is the Signed Distance Function (SDF), which computes the distance between each point in a three-dimensional space and the surface of its nearest obstacle. This function offers a differentiable representation of the environment, which proves advantageous for algorithms based on gradient optimization. However, a critical limitation of the SDF approach is its inability to accurately determine the exact penetration vector in scenarios involving deep collisions, as shown in Figure 5. To address collision detection and avoidance, the Gilbert–Johnson–Keerthi (GJK) algorithm coupled with the Expanding Polytope Algorithm (EPA) [33] is commonly employed. The GJK algorithm is utilized to rapidly determine whether a collision occurs through the concept of the Minkowski difference. Subsequent to this, the EPA is tasked with calculating the most efficient vector for collision escape, based on the findings of the GJK algorithm. Despite their utility, these algorithms exhibit limitations, particularly in handling objects with concave geometries. In light of these challenges, particularly prevalent in internal grasp scenarios, we propose a new approach to deep collision avoidance for grasp planning. This approach is designed to effectively navigate the intricacies of deep collision scenarios, thereby enhancing the robustness and reliability of robot grasping.
The proposed approach for determining the shortest escape vector in robot grasping scenarios, as depicted in Figure 6, is structured into three steps.
The first step involves representing both the object and the gripper as multiple spheres, as shown in Figure 7. This step enables the decomposition of any non-convex object into convex shapes, facilitating the efficient calculation of the Minkowski difference for spherical forms.
Next, the Minkowski difference between each pair of spheres—one from the object and one from the gripper—is computed. This calculation, guided by Equation (10), results in the formation of a union of spheres, collectively representing the overlapping regions between the object and the gripper.
P G = { A i B j | A i P , B j G }
The object, denoted as P, is represented as P = i = 1 : M A i , where each A i signifies an individual sphere on the object. Here, M is the total number of spheres that constitute the object. Similarly, the gripper, denoted as G, is represented as G = i = 1 : N B i , with each B j representing a sphere on the gripper, and N being the total number of spheres that constitute the gripper. The Minkowski difference between two spheres is calculated using Equation (11).
A i B j = S p h e r e ( c A i c B j , r A i + r B j )
where c A i represents the center position, and r A i denotes the radius of the sphere A i constituting the object P. Similarly, c B i is the center position, and r B i signifies the radius of the sphere B i on the gripper G.
The final step involves calculating the boundary of the union of these spheres. The shortest distance from this boundary to the origin is identified as the escape vector. To streamline this process, we introduce a grid-based method to reduce computational complexity. Each grid point is evaluated to determine whether it is inside a sphere, on the sphere’s surface, or unoccupied, according to Equation (12). The shortest escape vector is then determined by identifying the shortest distance between a point on the surface of the combined spherical union and the origin. The vector from this point to the origin represents the shortest possible path to disengage the gripper from the object without collision.
d i j = | p k c i j | r i j d i j < 0 , inside   ball d i j = 0 , on   ball   surface d i j > 0 , not   occupied
The position of the kth grid in the space is denoted by p k . c i j = c A i c B j , which represents the center of the union ball A i B j , calculated as the difference between the centers of A i and B j . r i j = r A i + r B j , which is the radius of the union ball A i B j , determined by the sum of the radius of A i and B j .

4. Experiments

In this section, both simulations and real-world experiments are conducted to verify the effectiveness of the proposed LfD-based grasp-planning algorithm. The desktop computer we used was equipped with a 3.7 GHz CPU and 32 GB of memory. An SMC parallel gripper was used in the experiments. The width d 1 of the gripper finger was 2 cm, whereas the length d 2 was 10 cm. To incorporate a margin of flexibility in the grasp planning, a scale factor α was introduced, set at 1.5 . Applying this scale factor, the parameters of the gripper were adjusted to d 1 = α · 2 = 3 cm and d 2 = α · 10 = 15 cm.

4.1. Experiments through Simulations

In the simulations, the contact regions and the approach direction were specified by the human operator through the use of a mouse, as shown in Figure 8. The method employed for evaluation incorporated three datasets: HomebrewedDB [34], YCB-Video [35], and T-less [36]. To ensure a comprehensive and varied analysis, nine objects were picked from each dataset, resulting in a diverse range of 27 objects. To evaluate the proposed method, the original iterative surface fitting (ISF) method was employed as a benchmark. The evaluation metric was the surface fitting error E t o t a l = E p + λ E n , as defined by Equation (3). This metric quantifies the average distance, in millimeters, between the corresponding points on the surfaces of the gripper tips and the object, thus providing a measure of the accuracy of the fit. In addition, the computation time for grasp planning, measured in seconds, was recorded for each method. The simulated results for the various objects are presented in Figure 9, Figure 10 and Figure 11. Furthermore, as detailed in Table 1, the simulation results revealed certain limitations of the original ISF method, particularly its tendency to converge to local optima, requiring the resampling of grasp poses. This resampling process subsequently resulted in increased computation times. In contrast, the proposed method, incorporating contact regions and approach directions specified by human operators, demonstrated improved performance. Specifically, it achieved reduced fitting errors and shorter computation times for determining the optimal grasp poses owing to the employment of closed-form solutions for grasp planning. The experimental results underscore the enhanced efficiency and accuracy of the proposed method in grasp-planning applications.
We further evaluated the effectiveness of the proposed collision avoidance method in scenarios involving the internal grasping of four distinct objects. The results are visualized in Figure 12. The process involved randomly sampling an initial grasp pose, followed by the generation of collision-free grasp poses using the surface fitting method. The evaluation metric was the success rate of collision-free optimal grasps based on a total of 20 samples, comparing scenarios with and without the implementation of collision avoidance. The comparison results are detailed in Table 2. It is noteworthy that internal grasping presented a significant challenge due to the necessity of locating a feasible grasp within a narrow space, a task that is markedly more complex in the absence of a collision avoidance mechanism. Our findings demonstrate that integrating the collision avoidance function significantly enhanced success rates across all four objects.

4.2. Experiments with a Real Robot

The proposed approach was further validated on a FANUC LR-Mate 200iD (FANUC America Corporation, Rochester Hills, MI, USA), a 6-degree-of-freedom industrial robot, tested with various objects. For detecting human hand movements, an Intel RealSense D435 camera (Intel Corporation, Santa Clara, CA, USA) was utilized in these experiments. Compared with other sensors, such as data gloves or thermal cameras, the advantage of using a 3D camera is that it allows for the same sensor to be used for human hand tracking in human demonstration and object detection in robot execution since the demonstrator and the robot share the same workspace. The well-known Iterative Closest Points (ICP) method was employed for object localization. After detecting the positions of both the human hand pose and the target object, the contact region and approach direction were calculated within the camera coordinate system. Once the optimal grasp was determined in the camera coordinates, the result was subsequently transformed into the robot coordinate system for execution by the robot.
The proposed LfD-based grasp planning was demonstrated on four different objects: a water pipe, water valve, converter, and toy rabbit. The results are shown in Figure 13. The top row of the figure illustrates the human demonstration. The second row shows the hand plane; the identified contact region, marked in red; and the approach direction, marked in yellow. The third row shows the generated optimal grasp, whereas the fourth row shows the robot successfully executing the grasp. Finally, the bottom row illustrates the ability of the proposed method to rotate the object by 90 degrees relative to the ground, simulating a pose-grasp motion. The experimental videos are available in Supplementary Materials.
We also evaluated the proposed method to encompass scenarios involving the internal grasping of four distinct objects. The results, as illustrated in Figure 14, affirm that our method is capable of identifying feasible grasping poses, even within restricted spatial confines. This is primarily attributed to the integrated collision avoidance function, which effectively identifies an escape vector to facilitate the generation of a collision-free grasp.
To quantitatively evaluate the proposed method, we conducted a series of experimental comparisons: Method 1 employed only the contact region; Method 2 employed only the approach direction; and Method 3 utilized both the contact region and approach direction. The evaluation metric was the success rate, indicating the robot’s ability to reach the grasp pose generated from human demonstrations without colliding with the surrounding environment. The results of these experiments are comprehensively detailed in Table 3. Note that only Method 1 and Method 3 were subjected to comparative analysis and testing in the case of internal grasp since it was necessary to specify the contact region for generating an internal grasp. The analysis revealed that Method 1 exhibited the lowest success rates. This outcome can be attributed to the absence of approach direction information, leading to frequent collisions with the ground during grasp attempts. This indicates the critical importance of approach direction in facilitating a functional grasp. Method 2 demonstrated better success rates compared to Method 1. Notably, the integration of the approach direction reduced the frequency of collisions with the surrounding environment. However, the lack of contact area information often resulted in grasping poses tending toward an unstable position, hindering the successful lifting of the object. In contrast, Method 3 exhibited the highest success rates, suggesting that the inclusion of both the contact region and approach direction was effective. It was observed that the incorporation of a good initial contact region and approach direction derived from human demonstrations significantly expedited the convergence of the optimization algorithm to a feasible solution.
These experimental results not only validate the proposed method but also highlight its practical applicability in real-world grasp-planning scenarios. Despite the diverse shapes of the target objects, the proposed LfD-based approach demonstrated its efficiency in synthesizing human-like grasps. A key strength of this approach lies in its ability to adeptly avoid collisions, a critical aspect in robotic grasping scenarios. Furthermore, the method ensures a firm grasp of the object, which is essential for successfully executing post-grasp motions. This adaptability to various object geometries while maintaining a reliable and collision-free grasp underscores the robustness of the proposed method. The effectiveness of replicating human-like grasping strategies enhances the practical applicability of robotic systems in diverse environments.

5. Conclusions

This work introduces a learning from demonstration-based grasp-planning approach that effectively generates stable grasps by integrating grasp skills derived from human demonstrations. We demonstrate that the proposed method can intuitively acquire critical grasp skills, namely the contact region and approach direction, from a single human demonstration using a standard RGB-D camera. This approach contrasts with and offers advantages over more specialized equipment such as data gloves or thermal cameras. In addition, the optimal grasp is solved using a closed-form solution, which incorporates the nuances of human grasp skills. To validate the effectiveness of this approach, a series of experiments is conducted using a FANUC industrial robot. The results from these experiments demonstrate that the proposed method not only achieves stable grasping but also effectively aligns with human intentions for post-grasp motions.
In the future, we plan to extend the algorithm to accommodate multi-finger grippers while considering the kinematic constraints of such grippers. In addition, there is an interest in exploring how general grasp skills can be extracted from human demonstrations across a variety of object shapes. This exploration is expected to leverage deep learning techniques, potentially leading to more versatile and adaptable robot grasping capabilities.

Supplementary Materials

The following supporting information can be downloaded at: https://0-www-mdpi-com.brum.beds.ac.uk/article/10.3390/s24020618/s1, Video S1: Experimental Videos for Learning from Demonstration-Based Grasp Planning.

Author Contributions

Conceptualization, K.W.; methodology, K.W. and Y.F.; software, K.W.; validation, K.W. and Y.F.; formal analysis, K.W.; investigation, K.W.; writing—original draft preparation, K.W.; writing—review and editing, K.W., Y.F. and I.S.; visualization, K.W.; supervision, I.S.; project administration, K.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, H.; Tang, J.; Sun, S.; Lan, X. Robotic Grasping from Classical to Modern: A Survey. arXiv 2022, arXiv:2202.03631. [Google Scholar]
  2. Saito, D.; Sasabuchi, K.; Wake, N.; Takamatsu, J.; Koike, H.; Ikeuchi, K. Task-grasping from human demonstration. arXiv 2022, arXiv:2203.00733. [Google Scholar]
  3. Mandikal, P.; Grauman, K. Learning dexterous grasping with object-centric visual affordances. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 6169–6176. [Google Scholar]
  4. Lin, Y.; Sun, Y. Robot grasp planning based on demonstrated grasp strategies. Int. J. Robot. Res. 2015, 34, 26–42. [Google Scholar] [CrossRef]
  5. Geng, T.; Lee, M.; Hülse, M. Transferring human grasping synergies to a robot. Mechatronics 2011, 21, 272–284. [Google Scholar] [CrossRef]
  6. Wang, K.; Fan, Y.; Sakuma, I. Robot Grasp Planning from Human Demonstration. In Proceedings of the 2023 15th International Conference on Computer and Automation Engineering (ICCAE), Sydney, Australia, 3–5 March 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 421–426. [Google Scholar]
  7. Pinto, L.; Gupta, A. Supersizing self-supervision: Learning to grasp from 50k tries and 700 robot hours. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 3406–3413. [Google Scholar]
  8. Song, S.; Zeng, A.; Lee, J.; Funkhouser, T. Grasping in the wild: Learning 6dof closed-loop grasping from low-cost demonstrations. IEEE Robot. Autom. Lett. 2020, 5, 4978–4985. [Google Scholar]
  9. Deng, Y.; Guo, X.; Wei, Y.; Lu, K.; Fang, B.; Guo, D.; Liu, H.; Sun, F. Deep reinforcement learning for robotic pushing and picking in cluttered environment. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 4–8 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 619–626. [Google Scholar]
  10. Zhao, W.; Queralta, J.P.; Westerlund, T. Sim-to-real transfer in deep reinforcement learning for robotics: A survey. In Proceedings of the 2020 IEEE Symposium Series on Computational Intelligence (SSCI), Canberra, Australia, 1–4 December 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 737–744. [Google Scholar]
  11. Bicchi, A.; Kumar, V. Robotic grasping and contact: A review. In Proceedings of the 2000 ICRA, Millennium Conference, IEEE International Conference on Robotics and Automation, San Francisco, CA, USA, 24–28 April 2000; Symposia Proceedings (Cat. No. 00CH37065). IEEE: Piscataway, NJ, USA, 2000; Volume 1, pp. 348–353. [Google Scholar]
  12. Ciocarlie, M.; Goldfeder, C.; Allen, P. Dexterous grasping via eigengrasps: A low-dimensional approach to a high-complexity problem. In Proceedings of the Robotics: Science and Systems Manipulation Workshop-Sensing and Adapting to the Real World, Atlanta, GA, USA, 30 June 2007. [Google Scholar]
  13. Fan, Y.; Tomizuka, M. Efficient grasp planning and execution with multifingered hands by surface fitting. IEEE Robot. Autom. Lett. 2019, 4, 3995–4002. [Google Scholar] [CrossRef]
  14. Dai, W.; Sun, Y.; Qian, X. Functional analysis of grasping motion. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 3507–3513. [Google Scholar]
  15. Mehrkish, A.; Janabi-Sharifi, F. A comprehensive grasp taxonomy of continuum robots. Robot. Auton. Syst. 2021, 145, 103860. [Google Scholar]
  16. Mehrkish, A.; Janabi-Sharifi, F. Grasp synthesis of continuum robots. Mech. Mach. Theory 2022, 168, 104575. [Google Scholar] [CrossRef]
  17. Feix, T.; Romero, J.; Schmiedmayer, H.B.; Dollar, A.M.; Kragic, D. The grasp taxonomy of human grasp types. IEEE Trans. Hum.-Mach. Syst. 2015, 46, 66–77. [Google Scholar]
  18. Corona, E.; Pumarola, A.; Alenya, G.; Moreno-Noguer, F.; Rogez, G. Ganhand: Predicting human grasp affordances in multi-object scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5031–5041. [Google Scholar]
  19. Kleeberger, K.; Bormann, R.; Kraus, W.; Huber, M.F. A survey on learning-based robotic grasping. Curr. Robot. Rep. 2020, 1, 239–249. [Google Scholar] [CrossRef]
  20. Ozawa, R.; Ueda, N. Supervisory control of a multi-fingered robotic hand system with data glove. In Proceedings of the 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Diego, CA, USA, 29 October–2 November 2007; IEEE: Piscataway, NJ, USA, 2007; pp. 1606–1611. [Google Scholar]
  21. Liu, H.; Xie, X.; Millar, M.; Edmonds, M.; Gao, F.; Zhu, Y.; Santos, V.J.; Rothrock, B.; Zhu, S.C. A glove-based system for studying hand-object manipulation via joint pose and force sensing. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 6617–6624. [Google Scholar]
  22. Lakshmipathy, A.; Bauer, D.; Bauer, C.; Pollard, N.S. Contact transfer: A direct, user-driven method for human to robot transfer of grasps and manipulations. In Proceedings of the 2022 International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 6195–6201. [Google Scholar]
  23. Karunratanakul, K.; Yang, J.; Zhang, Y.; Black, M.J.; Muandet, K.; Tang, S. Grasping field: Learning implicit representations for human grasps. In Proceedings of the 2020 International Conference on 3D Vision (3DV), Fukuoka, Japan, 25–28 November 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 333–344. [Google Scholar]
  24. Wang, P.; Manhardt, F.; Minciullo, L.; Garattoni, L.; Meier, S.; Navab, N.; Busam, B. DemoGrasp: Few-shot learning for robotic grasping with human demonstration. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 5733–5740. [Google Scholar]
  25. Rosales, C.; Ros, L.; Porta, J.M.; Suárez, R. Synthesizing grasp configurations with specified contact regions. Int. J. Robot. Res. 2011, 30, 431–443. [Google Scholar]
  26. Ekvall, S.; Kragic, D. Learning and evaluation of the approach vector for automatic grasp generation and planning. In Proceedings of the 2007 IEEE International Conference on Robotics and Automation, Roma, Italy, 10–14 April 2007; IEEE: Piscataway, NJ, USA, 2007; pp. 4715–4720. [Google Scholar]
  27. Hillenbrand, U.; Roa, M.A. Transferring functional grasps through contact warping and local replanning. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura-Algarve, Portugal, 7–12 October 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 2963–2970. [Google Scholar]
  28. Brahmbhatt, S.; Handa, A.; Hays, J.; Fox, D. Contactgrasp: Functional multi-finger grasp synthesis from contact. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 4–8 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 2386–2393. [Google Scholar]
  29. Wang, K.; Fan, Y.; Sakuma, I. Robot Programming from a Single Demonstration for High Precision Industrial Insertion. Sensors 2023, 23, 2514. [Google Scholar] [PubMed]
  30. Wang, K.; Tang, T. Robot programming by demonstration with a monocular RGB camera. Ind. Robot. Int. J. Robot. Res. Appl. 2023, 50, 234–245. [Google Scholar] [CrossRef]
  31. Cutkosky, M.R. On grasp choice, grasp models, and the design of hands for manufacturing tasks. IEEE Trans. Robot. Autom. 1989, 5, 269–279. [Google Scholar] [CrossRef]
  32. Fan, Y.; Lin, H.C.; Tang, T.; Tomizuka, M. Grasp planning for customized grippers by iterative surface fitting. In Proceedings of the 2018 IEEE 14th International Conference on Automation Science and Engineering (CASE), Munich, Germany, 20–24 August 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 28–34. [Google Scholar]
  33. Van Den Bergen, G. Proximity queries and penetration depth computation on 3d game objects. In Proceedings of the Game Developers Conference, San Jose, CA, USA, 22–24 March 2001; Volume 170, p. 209. [Google Scholar]
  34. Kaskman, R.; Zakharov, S.; Shugurov, I.; Ilic, S. Homebreweddb: Rgb-d dataset for 6d pose estimation of 3d objects. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
  35. Xiang, Y.; Schmidt, T.; Narayanan, V.; Fox, D. Posecnn: A convolutional neural network for 6d object pose estimation in cluttered scenes. arXiv 2017, arXiv:1711.00199. [Google Scholar]
  36. Hodan, T.; Haluza, P.; Obdržálek, Š.; Matas, J.; Lourakis, M.; Zabulis, X. T-LESS: An RGB-D dataset for 6D pose estimation of texture-less objects. In Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision (WACV), Santa Rosa, CA, USA, 24–31 March 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 880–888. [Google Scholar]
Figure 1. The pipeline of the proposed approach consists of two stages: grasp skill recognition and grasp synthesis. First, the contact regions and approach directions are recognized from the human demonstration. Then, grasp optimization, which iteratively fits the surface between the gripper finger surface and the contact region, is employed to synthesize the grasp corresponding to the human demonstration. The generated grasp pose is evaluated to ensure it is collision-free. If a collision is detected, the gripper is moved to a safe pose, and the grasp optimization process is repeated to generate a new pose.
Figure 1. The pipeline of the proposed approach consists of two stages: grasp skill recognition and grasp synthesis. First, the contact regions and approach directions are recognized from the human demonstration. Then, grasp optimization, which iteratively fits the surface between the gripper finger surface and the contact region, is employed to synthesize the grasp corresponding to the human demonstration. The generated grasp pose is evaluated to ensure it is collision-free. If a collision is detected, the gripper is moved to a safe pose, and the grasp optimization process is repeated to generate a new pose.
Sensors 24 00618 g001
Figure 2. The hand-plane definitions in the two- and three-finger Robotiq gripper cases.
Figure 2. The hand-plane definitions in the two- and three-finger Robotiq gripper cases.
Sensors 24 00618 g002
Figure 3. The hand plane is formed by fitting hand joints on the thumb finger and index finger.
Figure 3. The hand plane is formed by fitting hand joints on the thumb finger and index finger.
Sensors 24 00618 g003
Figure 4. The contact region is determined by passing the hand plane through the object’s point cloud (blue). (a) First, select the points (red) within the distance d 1 to the hand plane. (b) Second, select the points (red) from the first step within the distance d 2 to the origin of the hand plane.
Figure 4. The contact region is determined by passing the hand plane through the object’s point cloud (blue). (a) First, select the points (red) within the distance d 1 to the hand plane. (b) Second, select the points (red) from the first step within the distance d 2 to the origin of the hand plane.
Sensors 24 00618 g004
Figure 5. (a) In instances of minor collisions, where the interaction between the gripper and the object is relatively superficial, escaping collisions is typically straightforward. (b) In situations involving deep collisions, the complexity of avoiding collisions escalates significantly.
Figure 5. (a) In instances of minor collisions, where the interaction between the gripper and the object is relatively superficial, escaping collisions is typically straightforward. (b) In situations involving deep collisions, the complexity of avoiding collisions escalates significantly.
Sensors 24 00618 g005
Figure 6. (a) Approximate both the object and the gripper into spherical components. (b) Calculate the Minkowski difference to formulate a union of spheres. (c) Determine the boundary of the union of spheres. (d) Find the shortest escape vector. (e) The result of the collision-free grasp.
Figure 6. (a) Approximate both the object and the gripper into spherical components. (b) Calculate the Minkowski difference to formulate a union of spheres. (c) Determine the boundary of the union of spheres. (d) Find the shortest escape vector. (e) The result of the collision-free grasp.
Sensors 24 00618 g006
Figure 7. Examples of approximation into spherical components.
Figure 7. Examples of approximation into spherical components.
Sensors 24 00618 g007
Figure 8. The red area is the contact region (red) and the yellow arrow is the approach direction.
Figure 8. The red area is the contact region (red) and the yellow arrow is the approach direction.
Sensors 24 00618 g008
Figure 9. The simulation results on different objects. The nine objects were selected from the HomebrewedDB dataset.
Figure 9. The simulation results on different objects. The nine objects were selected from the HomebrewedDB dataset.
Sensors 24 00618 g009
Figure 10. The simulation results on different objects. The nine objects were selected from the YCB-Video dataset.
Figure 10. The simulation results on different objects. The nine objects were selected from the YCB-Video dataset.
Sensors 24 00618 g010
Figure 11. The simulation results on different objects. The nine objects were selected from the T-less dataset.
Figure 11. The simulation results on different objects. The nine objects were selected from the T-less dataset.
Sensors 24 00618 g011
Figure 12. The simulation results of the internal grasping of different objects.
Figure 12. The simulation results of the internal grasping of different objects.
Sensors 24 00618 g012
Figure 13. The results using a real robot. After lifting the objects, the robot rotates the objects 90 degrees to simulate a post-grasp motion.
Figure 13. The results using a real robot. After lifting the objects, the robot rotates the objects 90 degrees to simulate a post-grasp motion.
Sensors 24 00618 g013
Figure 14. The experimental results of the internal grasp case using a real robot.
Figure 14. The experimental results of the internal grasp case using a real robot.
Sensors 24 00618 g014
Table 1. Surface fitting errors and computation times on different objects.
Table 1. Surface fitting errors and computation times on different objects.
DatasetObject NameISFLfD-Based ISF
E total  (mm) t total  (s) E total  (mm) t total  (s)
HomeDBJig13.0051.1915.3260.798
Dryer14.6001.0716.9740.747
Cap9.6091.9905.4590.975
Converter11.3872.5616.4040.519
Motocycle12.0322.4397.2850.661
Dinosaur17.2191.7449.4270.895
Clamp9.4391.2407.3950.614
Cup10.8711.8678.0680.771
Bear12.7581.0437.2150.558
Average12.3241.6837.0610.726
YCB-VBox3.4570.4043.5600.369
Can4.1930.4202.5510.406
Mustard8.7200.8326.1740.559
Spam8.7400.5324.2830.370
Banana3.1450.4343.0310.287
Bowl8.9241.2384.7780.862
Scissor5.8080.6793.6620.277
Magic pen4.8060.5234.0920.317
Clip6.8360.3794.6810.347
Average6.0700.6044.0900.422
T-lessBulb5.1980.5062.6000.327
S-connector8.7720.8723.5880.334
C-plug6.9681.0046.5250.532
R-connector11.6481.2167.8580.524
C-connector10.5931.5626.6230.639
Switch5.1590.9153.4090.359
Box2.4050.4392.5360.412
Plug13.0581.2825.1710.865
Circle box7.7840.5265.4170.329
Average7.9540.9254.8590.480
Total average8.7831.0715.3370.543
Table 2. Success rates (collision-free grasps/total samples), with and without the implementation of collision avoidance.
Table 2. Success rates (collision-free grasps/total samples), with and without the implementation of collision avoidance.
Object NameGearWater ValveWater PipeJig
Without collision avoidance11/2013/2010/207/20
With collision avoidance20/2018/2019/2020/20
Table 3. Success rates (collision-free grasps/total samples).
Table 3. Success rates (collision-free grasps/total samples).
External GraspInternal Grasp
ObjectsWater PipeWater ValveConverterToy RabbitGearWater ValveWater PipeJig
Method 111/206/202/2013/2011/207/2017/2012/20
Method 215/2010/2013/2013/20----
Method 318/2020/2020/2019/2020/2020/2020/2020/20
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, K.; Fan, Y.; Sakuma, I. Robot Grasp Planning: A Learning from Demonstration-Based Approach. Sensors 2024, 24, 618. https://0-doi-org.brum.beds.ac.uk/10.3390/s24020618

AMA Style

Wang K, Fan Y, Sakuma I. Robot Grasp Planning: A Learning from Demonstration-Based Approach. Sensors. 2024; 24(2):618. https://0-doi-org.brum.beds.ac.uk/10.3390/s24020618

Chicago/Turabian Style

Wang, Kaimeng, Yongxiang Fan, and Ichiro Sakuma. 2024. "Robot Grasp Planning: A Learning from Demonstration-Based Approach" Sensors 24, no. 2: 618. https://0-doi-org.brum.beds.ac.uk/10.3390/s24020618

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop