- ARQ (Advanced Robotics at Queen Mary), School of Electronic Engineering and Computer Science, Queen Mary University of London, London, United Kingdom
Grasp stability prediction of unknown objects is crucial to enable autonomous robotic manipulation in an unstructured environment. Even if prior information about the object is available, real-time local exploration might be necessary to mitigate object modelling inaccuracies. This paper presents an approach to predict safe grasps of unknown objects using depth vision and a dexterous robot hand equipped with tactile feedback. Our approach does not assume any prior knowledge about the objects. First, an object pose estimation is obtained from RGB-D sensing; then, the object is explored haptically to maximise a given grasp metric. We compare two probabilistic methods (i.e. standard and unscented Bayesian Optimisation) against random exploration (i.e. uniform grid search). Our experimental results demonstrate that these probabilistic methods can provide confident predictions after a limited number of exploratory observations, and that unscented Bayesian Optimisation can find safer grasps, taking into account the uncertainty in robot sensing and grasp execution.
1 Introduction
Autonomous robotic grasping of arbitrary objects is a challenging problem that is becoming increasingly popular in the research community due to its importance in several applications, such as pick-and-place in manufacturing and logistics, service robots in healthcare and robotic operations in hazardous environments, e.g. nuclear decommissioning (Billard and Kragic, 2019; Graña et al., 2019). Grasping involves several phases: from detecting the object location to choosing the grasp configuration (i.e. how the gripper or robot hand should contact the object) with the final objective of keeping the object stable in the robot grip. Moreover, when we consider dexterous robotic hands with multiple fingers, several contact points on an object must be identified to achieve a robust grasp (Miao et al., 2015; Ozawa and Tahara, 2017). This is particularly challenging when limited or no prior information is available about the object, and therefore it is necessary to rely more heavily on real-time robot perception.
Robot perception for grasping typically includes vision, touch and proprioception; notably, all these modalities provide useful information about different aspects of the grasping problem. Vision is often the dominant modality in the phases that precede the lifting of the grasped object (Du et al., 2019), due to the ability to capture global information about the scene. However, vision is not equally effective at detecting local information about the interaction between the robot hand and the object, including the forces exerted by the hand, the hand configuration, and some physical attributes of the object, such as its stiffness or the friction coefficient of its surface: all these aspects are better perceived by touch. Therefore, the use of tactile sensing has become more and more popular (Luo et al., 2017), not just during the holding of the object (e.g. to react to slips) but also to discover how to grasp the object. In addition, the concept of active perception (Bajcsy et al., 2018), or interactive perception (Bohg et al., 2017), is particularly relevant in this case, because to collect useful tactile information the robot should perform relevant actions (Seminara et al., 2019), e.g. a controlled manual exploration of the object surface. However, one big challenge of relying on active real-time perception is the underlying uncertainty of robotic sensing and action generation (Wang et al., 2020). To cope with this uncertainty, we propose to enrich the visual information with a haptic exploration procedure driven by a probabilistic model, i.e. Bayesian Optimisation. The robot first detects the object location using point cloud data extracted from an RGB-D sensor. Then, an exploration procedure starts in which the robot hand evaluates different grasp configurations selected by Bayesian Optimization, based on a grasp metric computed from tactile sensing (Figure 1A). Finally, after the best grasp configuration is found, the object is picked up. We assume that the object is completely unknown to the system: we do not rely on any model or previous learning, but only on a real-time exploration that is relevant only for the current object and for the current execution of the grasp (i.e. not for any other object or any future execution of the grasp).
FIGURE 1. (A) Overview of the setup and the exploration pipeline. The setup includes a UR5 robot arm, an Allegro robot hand equipped with Optoforce 3D force sensors in the fingertips, a Kinect RGB-D sensor. (B) The Allegro hand visualised with the RVIZ software. (C) The three complex-shaped objects used in the experiment.
We extend the simulation results previously obtained in (Nogueira et al., 2016; Castanheira et al., 2018) by testing the system with a real robot hand, and by performing experiments on three objects with complex shapes. Notably, many additional uncertainties are present in a real-world environment (e.g. insensitivity of sensors, disturbance in the position of the object while exploring) that are not present in a controlled simulated environment. In particular, we show that an unscented version of Bayesian Optimization proves to be even more effective than the classic Bayesian Optimisation to discover robust grasps under uncertainty, with a limited number of exploration steps.
The contributions of the paper are threefold:
1) an approach to predict a safe grasp for an unknown object from a combination of visual and tactile perception.
2) a probabilistic exploration model that considers uncertainties of the real world in order to predict a safe grasp.
3) a series of experiments that demonstrate how the proposed system can find robot grasps that maximise the probability of the object being stable after it has been picked and lifted.
The paper is organised as follows: in Section 2, we describe the state of art for visuo-tactile data fusion and grasping of unknown objects. Section 3 provides an overview of our methodology. In Section 4, we describe the configuration and the experimental protocol. Discussion on the results is presented in Section 5. Finally, in Section 6, we conclude by summarising the performance of our approach and presenting possible directions for improvements and future research.
2 Related Work
Grasping objects of unknown shape is an essential skill for automation in manufacturing industries. Many existing grasping techniques require a 2D or 3D geometrical model, limiting its application in different working environments (Ciocarli and Allen, 2009). 3D reconstruction framework for detection of fruit in real environments is presented by Lin et al. (2020). Vision technology has advanced to detect objects in a natural environment over the years, even in the presence of shadows (Chen et al., 2020). Kolycheva née Nikandrova and Kyrki (2015) introduces a system using RGB-D vision to estimate the shape and pose of the object. The models for grasp stability are learnt over a set of known objects using Gaussian process regression. While 3D vision technology has various applications in the engineering field, acquiring 3D images is an expensive process and mostly simulation-based (Shao et al., 2019).
Merzić et al. (2018) makes use of deep reinforcement learning technique to grasp partially visible/occluded objects. It does not rely on the dataset of the object models but instead uses tactile sensors to achieve grasp stability on unknown objects in a simulation. Zhao et al. (2020) implements probabilistic modelling with a neural network to select a group of grasp points for an unknown object. There is also a work on learning object grasping based on visual cues, and the selection of features are often based on human intuitions (Saxena et al., 2008). However, vision-based accuracy is limited due to its standardization and occlusions. Some details can be overlooked even for known objects, which may cause failure in grasping objects (Kiatos et al., 2020). Our work is different from deep learning or reinforcement learning as there is no training data or an existing dataset to predict stable regions. The method explores an unknown object in real-time and finds a solution that maximizes a given grasp metric.
Tactile sensing is capable of compensating for some of the problems of the vision-only approach. Indeed, being able to perceive touch allows the robot to understand when contact with the object has been made and have a better perception of the occluded areas of the object by making contact with those surfaces. Techniques are proposed to control slippage and grasp stabilization of the objects using tactile sensors only (James and Lepora, 2020; Shaw-Cortez et al., 2020). It is independent of the data of object mass, object centre of mass and forces acting on the object to prevent the object from slipping. Rubert et al. (2019) present seven different kinds of grasp quality metrics to predict how well it performs on the robotic platform and in simulations. Different classifiers are trained on the extensive database, and results are evaluated for each grasp. The human labelled database is used in this work, which requires more accuracy in collecting data using different protocols. To accomplish the autonomous grasping of an unknown object, we aim to predict the grasping stability of the object before lifting the object from the surface. In this paper, we used tactile feedback to predict the stability of the robotic grasp. We present real-time grasp safety prediction by haptic probabilistic modelling exploration with a dexterous robotic hand.
The conventional methods address the stability of the objects during in-hand manipulation. Our method predicts the stability of the grasp before lifting the object off from the surface. Li et al. (2014) maintains the stability of the object grasped in the air by changing the grasp configuration of the robotic hand. The state of the object is disturbed by adding extra weight on the object or manually pulling the object from the grasp. The work of Veiga et al. (2015) focuses on slip detection using tactile sensors during in-hand manipulation. The main difference of our methods to existing approaches in the literature is that we do not use any previous learning/training on any object: all the search is performed in real-time on completely unknown objects, i.e. no prior information and no prior data is used.
3 Methodology
A self-supervised model is used to compute the probability of grasp success using tactile and visual inputs. This allows evaluating the robustness of potential grasps.
3.1 Object Detection
We use 3D point cloud data to calculate the midpoint of the object. We define a specific area in an environment as a workspace in which the robot operates safely. The robot perceives the object placed on the workspace while the remaining point cloud data is filtered out, as shown in Figure 2A.
FIGURE 2. Methodology for the haptic exploration. (A) Point cloud data of the workspace. (B) The bounding box of the object extracted. (C) First step of path planning: towards a location at a fixed distance over the object bounding box. (D) Second step of path planning: lowering down and enclosing the fingers on the object to compute the grasp metric.
We are using Random Sample Consensus (RANSAC), a non-deterministic iterative algorithm for detection of the object (Zuliani et al., 2005). It tries to fit the points from the point cloud into a mathematical model of a dominant plane. RANSAC then identifies the points which do not constitute the dominant plane model. These points that do not fit into the plane model (called outliers) are clustered together to form one object. A minimum threshold is set to avoid the detection of tiny objects and filtering extra noise. We demonstrate our approach only on singulated objects, i.e. not in clutter. The approach could be applied to clutters, but it would require more sophisticated visual perception components to segment each object and identify its boundaries partially.
Dimensions of the object are used to create a 3D bounding box around the object, as shown in Figure 2B. The midpoint of the object is computed as the difference between the maximum and minimum boundary points in an axis parallel to the plane. This point is then used to reference the robot to move close to the object and initiate tactile exploration. Path planning towards the object is executed in two steps to avoid collision with the environment. In the first step (Figure 2C), the arm moves to a safe distance above the object. The second step of path planning is then to move closer to the object (Figure 2D). Moveit! framework (Coleman et al., 2014) is utilised for implementation of motion planning. The process of instructing the robot to align itself closer to the object is described in Algorithm 1. The target pose is saved before the movement of the arm towards the object to avoid end-effector blocking the target during execution.
3.2 Force Metric Calculation
A constant envelop force is useful for the computation of force metric. An extensive review of the different criteria used for computing a grasp metric is described in Roa and Suárez, 2015. Following variables can be taken into account in the evaluation of a grasp metric:
• coordinates of the grasp points on the object.
• directions at which the force is applied at the grasp point.
• magnitude of the force experienced at the grasp point.
• pose of the robotic hand (in our case, Allegro hand).
Tactile exploration consists of closing the robotic hand at multiple points in an object and evaluating the grasp metric. In a closing state, fingers stop moving when the fingertips get in contact with the object. A force vector is created to grasp the object during metric calculation and picking the object. This force is calculated from coordinates of fingertip to virtual frame positioned in the middle of fingers and thumb. The concept of virtual springs is discussed in detail by Solak and Jamone (2019). Equation of grasp force is:
where
The volume of the Force Wrench Space (FWS) by Miller and Allen (1999) is used as a force metric to measure the stability of the grasp during tactile exploration. FWS is defined as the set of all forces applied to the object with all grasp contacts. It is a three-dimensional vector consisting of force components from all the four tactile sensors positioned on the tip of the fingers of the robotic hand. This metric is also independent of the coordinates of reference system. Function
During the closing state, the robotic hand wounds its fingers around the object. The grasp metric is calculated when a connection is established between the hand and the object. The size and coordinates of the object are assumed fixed to limit the size of the exploration space.
3.3 Probabilistic Modelling
We use two probabilistic exploration methods: scented and unscented bayesian optimization, and compare their performance with uniform grid exploration. The uniform grid approach is where all search points in bounded space have an equal probability of being explored.
3.3.1 Bayesian Optimisation
We consider the Bayesian Optimisation (BO) algorithm as one of the probabilistic models to accomplish the task of exploring global optima (Brochu et al., 2010). For n number of iterations, the input dataset of query point is x = {
3.3.2 Unscented Bayesian Optimisation
Unscented Bayesian Optimisation (UBO) is a method to propagate mean and covariance through nonlinear transformation. The basis of the algorithm is better manageability of an approximate probability distribution than approximate arbitrary nonlinear function (Nogueira et al., 2016). To calculate mean and covariance, a set of sigma points are chosen. These sigma points are deterministically chosen points that depict certain information about mean and covariance. The weighted combination of sigma points is then passed through linear function to compute the transformed distribution. The advantage of UBO over classical BO is it’s ability to consider uncertainty in the input space to find an optimal grasp. For dimension d, it requires 2d + 1 sigma points that show its computational cost is negligible compared to others such as Monte Carlo, which requires more samples or Gaussian function.
In UBO, the query is selected based on probability distribution. We choose the best query point considering it as deterministic, but also check its surrounding neighbours. Thus, while considering input noise, we will analyze the resulting posterior distribution through the acquisition function. Assuming that our prior distribution is Gaussian distribution where
where d is dimensional input space, κ parameter tunes magnitude of sigma points and
4 Implementation
The grasp metric of a candidate grasp is evaluated on a real robotic platform. We start from elementary visual perception, which is used by the robot to come closer to the object and to be able to initiate the haptic exploration. Motion planning is initially visualised using the robot operating system (ROS) before execution in the real-world environment. Experiment to pick the object from the surface is designed to the evaluate performance of the exploration algorithm. Objects are manually put in the same approximate location to maintain consistency in the evaluation and show that the grasps found with UBO are more resilient to minor variations in object position.
In a real use-case, the robot hand would approach the object (starting from the visual estimation of the object pose). It would haptically explore the object (without lifting it, only by touching it in the different possible grasp postures/configurations) to maximise a given grasp metric (i.e. based on the measured contact forces), and then it would lift the object by using the best grasp that has been found with the haptic exploration. This is relevant for scenarios in which we want to optimise the safety of the grasp over speed, e.g. nuclear-decommissioning settings, or other scenarios in which we do want to minimise the possibility of the object falling from the grasp, at the cost of requiring more time to find the safest grasp.
4.1 Configuration
For the experimental setup, a camera is required to generate point cloud data of the objects. The generation of the point cloud can be achieved using a stereo camera or RGBD camera. The authors in Vezzani et al., 2017 have used a stereo camera to generate point cloud data, and authors of Rodriguez et al., 2012 are using Kinect. Both have presented that the generated point cloud is satisfactory so that any camera can be selected. We have used a kinect camera for the generation of point cloud data in our experiments.
To achieve our objective of successfully grasping an unknown object, we have set up a UR5 robot in the lab. Allegro hand is mounted at the end of the UR5 arm as an end effector. Kinect is fixed at the top of the base of the robot, facing perpendicular to the workspace. Optoforce OMD 20-SE-40N is a 3-axis force sensors that measure the forces experienced by the fingers of the Allegro hand (at a rate of 1 kHz). The workplace is 72 cm from the kinect frame. Any object within the workplace area (a rectangular area of 31 cm by 40 cm) is processed, and the extra points are filtered out. The orientation of the Allegro hand is fixed parallel to the axis of the workspace plane. The setup is shown in Figure 1A.
4.2 Protocol
To perform the experiments, we apply the following experimental protocol.
1) Object detection: to detect the unknown object in the environment, we use the RANSAC algorithm in point cloud library. This library allows the detection of the desired object and obtains its pose with respect to the camera.
2) Motion planning: once we have detected the pose of the object, the Moveit plans the collision-free movement of the robot to the top of the object.
3) Plan execution: after successful planning, the robot navigates itself to the target pose. This is also the starting pose for haptic exploration.
4) Haptic exploration: robot plans and navigates the robotic hand to search points queried by the exploration model. Search space is confined by limiting the orientation of the Allegro hand parallel to the surface.
5) Gradually gripping the object: when the robotic arm reaches the search point, it starts closing its fingers until contact is detected.
6) Applying grasping force: to ensure the gripper applies enough pressure over the object and not just touches it.
7) Calculation of grasp metric: evaluate grasp score of the candidate grasp.
8) Move the robotic arm to the next pose: open the grip of the robotic hand and move to the next pose directed by the probabilistic model. This process is repeated 50 iterations. Approximately two iterations are completed in a minute.
9) Stability testing: this experiment is performed after the completion of the exploration stage. The robot is manually navigated to the coordinates of the maximum grasp metric score to evaluate its stability.
10) Stability scoring: object is lifted 20 times from the surface and maintained in the air for 10 s.
5 Results
The proposed model is validated by exploring grasp points in the 3D space, but the contact points are searched on two dimensions. Experiments are conducted five times with probabilistic modelling exploration and then compared with the uniformly distributed exploration. BO and UBO models are used for probabilistic modelling exploration. At the end of each experiment, grasp point with the highest metric is used to lift the object from the surface. This process of lifting the object is repeated 20 times to find the stability score of the grasp point.
We used the objects from the dataset1 developed by EU RoMaNS to observe exploration performance. The objects in this dataset are commonly found in nuclear waste and are categorised in different categories such as bottles, cans, pipe joints. We conducted the experiments with three different kinds of complex objects: a c-shaped pipe joint, a mustard plastic bottle and a 3D printed blue object. The diameter of pipe joint is 6 cm from one end and 5.5 cm from another, the height of thread on the ring is 0.2 cm. A complex-shaped 3D printed blue object 11cm × 5.5 cm from dexnet dataset2 is also used to increase the persuasiveness of the data. Images of the objects can be seen in the Figure 1C. Objects were placed on the bubble wrap surface to increase the friction between the object and the plane. This was done because the fingertip force sensors are not very sensitive, and therefore the minimum contact force that can be measured (at first contact) may already produce a consistent displacement of the object (if the friction coefficient of the table surface was too low).
Scatter plots: The Figure 3 represents the points observed by each exploration method in all the experiments. The point represents the location of the middle finger of the robotic arm. A total of 250 search points (5 experiments with 50 iterations each) are plotted for each exploration method. It can be observed that more observations are recorded at the boundaries of the object for probabilistic methods. This is due to the concavity of the tactile sensors and their contact with the edges in the objects. It is expected for probabilistic models to explore the complex part of the object. It can also be seen that BO and UBO exploration converges to a more substantial part of the object. The figure also represents the optimal position with the highest metric score for all experiments for each exploration model. There are a total of 15 points represented, five for each approach. Again, the points are the location of the middle finger of the robotic arm.
FIGURE 3. Scatter Plots of all points explored in uniform, BO and UBO explorations for (A) bottle, (B) pipe joint and (C) blue object. The pose of Allegro hand at the start of experiments is shown in the first column. The last column represents the optimum position in 2D from each experiment.
Optimal position: The position with optimal grasp score is the distance from the world frame along the horizontal plane of the object. The middle point of the bounding box of the bottle is approximately 35 cm from the world frame and 37.5 cm in the pipe joint and blue object. The frames are shown in Figure 1A. Table 1 tabulates the optimal position of the object as observed in each experiment. It also shows the value of grasp metric value in the optimal position. The points are skewed towards one side of the object because of the constraint in the encoders of the thumb, which restricts the movement of the thumb to align with the middle finger (Figure 1B). The results indicate that probabilistic models have an optimum position similar to uniform distributed exploration with minor standard deviation in position and metric score.
TABLE 1. Coordinates of maximum metric observed of explorations for different objects from all experiments in the world frame. Standard deviation of the mean position in x and y axes, metric is also listed.
Convergence: The convergence of each exploration to its maximum grasp metric value reflects confidence in successfully lifting the object from the surface. The left column of Figure 5 presents the performance of BO, UBO and uniform explorations in converging to the final optimum position (x-axis) at each observation. Uniform, BO and UBO are represented by green, blue and orange lines, respectively. A total of five experiments are conducted with 50 observations for three different complex-shaped objects. Plots present convergence in the x-axis only because of the confined range of exploration in the y-axis (
Stability score: There are three possible states of stability when the object is lifted in the air: stable, partial stable and failure. A stable state is when three or four fingers of the robotic hand contact the object, and the object stays in the air for 10 s. Partial stability is when only the thumb and first finger hold the object in the air for 10 s. These two states are shown in Figure 4. Failure state is when the robotic arm fails to lift the object off the surface. In none of our experiments, an object dropped from the air. Table 2 tabulates the performance of each exploration in lifting the object. The results of the experiment to evaluate stability is shown on the right column of Figure 5. Frequency distribution of the five experiments for objects used is tabulated in Tables 3–5. The success rate for each exploration is the percentage of the robot lifting the object from the surface (both stable and partial state) and holds it in the air without a drop-off. In the calculation of stability score, the stable state is given double weight than the partial state. Failure state is excluded from the calculation. The formula is mentioned below:
FIGURE 4. Two states of stability observed while maintaining the object in the air: (A) stable grasp (B) partial stable grasp.
TABLE 2. Grasping state observed for all five experiments in lifting the object from surface. Total of five experiments with 20 iterations each.
FIGURE 5. Left column displays performance of uniform, BO and UBO explorations in converging to final optimum position (x-axis) in all experiments for (A) bottle, (B) pipe joint and (C) blue object. Right column shows stability scores and mean of the highest grasp metric for different exploration methods.
TABLE 5. Frequency distribution of experiments on complex-shaped blue object to evaluate grasp stability.
The results show that probabilistic models can converge to the optimum position with a higher grasp metric score in fewer iterations than uniformly distributed exploration.
The experimental results collected demonstrate:
• the ability of probabilistic methods to provide confidence in predicting a safe grasp in a minimal number of iterations.
• BO and UBO have the advantage of converging sooner than the uniform exploration, even with fewer observations.
• the potential of UBO to find safer grasps: this is evident in the case of the bottle, as the optimum points lie far from the edges.
• the success rate of UBO is the highest in lifting the object from the surface and maintaining it in the air for 10 s (i.e. stable grasp).
6 Conclusion
We presented a pipeline for object detection (using depth-sensing) and exploration (using tactile sensing) with a dexterous robotic hand, aimed at finding grasps that maximise the probability of the object being held robustly in hand after picking and lifting. Our approach is not based on any previous learning or prior information about the object: the system knows nothing about the object before the exploration starts.
The intelligence of the system lies in the real-time decisions about where to explore the object at each exploration step, so that the number of exploratory steps is minimised and the amount of information gathered is maximised. These decisions are based on a probabilistic model (BO). In particular, we show experimentally that an unscented version of the model (UBO) can find the more robust grasps, even in the presence of the natural uncertainty of robotic perception and action execution: we show this by repeating the grasps multiple times, showing that such grasps are robust to the minor inaccuracies/differences between each replication of the grasp. Given the nature of this approach, the most relevant applications are in scenarios in which the cost of dropping the object after grasp is very high, and it is therefore justified to invest some additional time in exploring the object haptically before picking it. For example, handling hazardous materials in a nuclear environment, collecting samples in space or deep sea missions, pick and place of fragile objects in logistics.
In our experiments, we assume to have no prior knowledge about the object; however, such information (if available) could be included in the probabilistic exploration models as a prior, also depending on the specific application. We use depth sensing to limit the search space by identifying a bounding box around the object: a more sophisticated visual perception component could permit defining an even more compact search space, e.g. consisting in a small set of tentatively good grasps.
Another possible improvement of our system is to use better tactile sensors on the robot fingertips that are more sensitive (Jamone et al., 2015; Paulino et al., 2017) and that can provide 3D force measurements on several contact points (Tomo et al., 2018). With such a sensor: we could detect the initial contact with the object earlier (i.e. based on a lower force threshold), therefore minimising undesired motion of the object during exploration; we could obtain a better estimation of the contact forces, that would lead to a more reliable assessment of the force closure metric; we could estimate other object properties (e.g. friction coefficient) that can also be included in the grasp metric, leading to better predictions of the grasp stability.
Data Availability Statement
Publicly available datasets were analyzed in this study. This data can be found here: https://github.com/ARQ-CRISP
Author Contributions
MS was responsible for the development of the pipeline and carrying out the experimental and analytical work. CC designed the exploration algorithms. GS provided technical assistance in controlling the robotic hand. LJ supervised the project and revised the draft. All authors listed approved it for publication.
Funding
Work partially supported by the EPSRC United Kingdom (projects NCNR, EP/R02572X/1, and MAN3, EP/S00453X/1) and by the Alan Turing Institute.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Footnotes
1https://sites.google.com/site/romansbirmingham
2https://berkeley.app.box.com/s/6mnb2bzi5zfa7qpwyn7uq5atb7vbztng
References
Bajcsy, R., Aloimonos, Y., and Tsotsos, J. K. (2018). Revisiting Active Perception. Auton. Robot 42, 177–196. doi:10.1007/s10514-017-9615-3
Billard, A., and Kragic, D. (2019). Trends and Challenges in Robot Manipulation. Science 364. doi:10.1126/science.aat8414
Bohg, J., Hausman, K., Sankaran, B., Brock, O., Kragic, D., Schaal, S., et al. (2017). Interactive Perception: Leveraging Action in Perception and Perception in Action. IEEE Trans. Robot. 33, 1273–1291. doi:10.1109/tro.2017.2721939
Brochu, E., Cora, V. M., and de Freitas, N. (2010). A Tutorial on Bayesian Optimization of Expensive Cost Functions, with Application to Active User Modeling and Hierarchical Reinforcement Learning. CoRR abs/1012.2599.
Castanheira, J., Vicente, P., Martinez-Cantin, R., Jamone, L., and Bernardino, A. (2018). Finding Safe 3d Robot Grasps through Efficient Haptic Exploration with Unscented Bayesian Optimization and Collision Penalty. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE), 1643–1648. doi:10.1109/iros.2018.8594009
Chen, M., Tang, Y., Zou, X., Huang, K., Huang, Z., Zhou, H., et al. (2020). Three-dimensional Perception of Orchard Banana central Stock Enhanced by Adaptive Multi-Vision Technology. Comput. Elect. Agric. 174, 105508. doi:10.1016/j.compag.2020.105508
Ciocarlie, M. T., and Allen, P. K. (2009). Hand Posture Subspaces for Dexterous Robotic Grasping. Int. J. Robotics Res. 28, 851–867. doi:10.1177/0278364909105606
Coleman, D., Sucan, I., Chitta, S., and Correll, N. (2014). Reducing the Barrier to Entry of Complex Robotic Software: a Moveit! Case Study.
Du, G., Wang, K., and Lian, S. (2019). Vision-based Robotic Grasping from Object Localization Pose Estimation Grasp Detection to Motion Planning: A Review. arXiv preprint arXiv:1905.06658.
Graña, M., Alonso, M., and Izaguirre, A. (2019). A Panoramic Survey on Grasping Research Trends and Topics. Cybernetics Syst. 50, 40–57. doi:10.1080/01969722.2018.1558013
James, J. W., and Lepora, N. F. (2021). Slip Detection for Grasp Stabilization with a Multifingered Tactile Robot Hand. IEEE Trans. Robot. 37, 506–519. doi:10.1109/TRO.2020.3031245
Jamone, L., Natale, L., Metta, G., and Sandini, G. (2015). Highly Sensitive Soft Tactile Sensors for an Anthropomorphic Robotic Hand. IEEE Sensors J. 15, 4226–4233. doi:10.1109/JSEN.2015.2417759
Kiatos, M., Malassiotis, S., and Sarantopoulos, I. (2021). A Geometric Approach for Grasping Unknown Objects with Multifingered Hands. IEEE Trans. Robot. 37, 735–746. doi:10.1109/tro.2020.3033696
Kolycheva née Nikandrova, E., and Kyrki, V. (2015). Task-specific Grasping of Simiiar Objects by Probabiiistic Fusion of Vision and Tactiie Measurements. In 2015 IEEE-RAS 15th International Conference on Humanoid Robots (Humanoids). 704–710. doi:10.1109/HUMANOIDS.2015.7363431
Li, M., Bekiroglu, Y., Kragic, D., and Billard, A. (2014). Learning of Grasp Adaptation through Experience and Tactile Sensing. IEEE/RSJ Int. Conf. Intell. Robots Syst., 3339–3346. doi:10.1109/iros.2014.6943027
Lin, G., Tang, Y., Zou, X., Cheng, J., and Xiong, J. (2020). Fruit Detection in Natural Environment Using Partial Shape Matching and Probabilistic Hough Transform. Precision Agric. 21, 160–177. doi:10.1007/s11119-019-09662-w
Luo, S., Bimbo, J., Dahiya, R., and Liu, H. (2017). Robotic Tactile Perception of Object Properties: A Review. Mechatronics 48, 54–67. doi:10.1016/j.mechatronics.2017.11.002
Merzić, H., Bogdanovic, M., Kappler, D., Righetti, L., and Bohg, J. (2018). Leveraging Contact Forces for Learning to Grasp. arXiv , 3615–3621.
Miao, W., Li, G., Jiang, G., Fang, Y., Ju, Z., and Liu, H. (2015). Optimal Grasp Planning of Multi-Fingered Robotic Hands: a Review. Appl. Comput. Math. Int. J. 14, 238–247.
Miller, A. T., and Allen, P. K. (1999). Examples of 3d Grasp Quality Computations. Proc. 1999 IEEE Int. Conf. Robotics Automation (Cat. No.99CH36288C) 2, 1240–1246. doi:10.1109/ROBOT.1999.772531
Zuliani, M. C. K., and Manjunath, B. (2005). The Multiransac Algorithm and its Application to Detect Planar Homographies. In IEEE International Conference on Image Processing doi:10.1109/icip.2005.1530351
Nogueira, J., Martinez-Cantin, R., Bernardino, A., and Jamone, L. (2016). Unscented Bayesian Optimization for Safe Robot Grasping. 1967–1972. doi:10.1109/IROS.2016.7759310
Ozawa, R., and Tahara, K. (2017). Grasp and Dexterous Manipulation of Multi-Fingered Robotic Hands: a Review from a Control View point. Adv. Robotics 31, 1030–1050. doi:10.1080/01691864.2017.1365011
Paulino, T., Ribeiro, P., Neto, M., Cardoso, S., Schmitz, A., Santos-Victor, J., et al. (2017). Low-cost 3-axis Soft Tactile Sensors for the Human-Friendly Robot Vizzy. In 2017 IEEE International Conference on Robotics and Automation (ICRA). 966–971. doi:10.1109/ICRA.2017.7989118
Roa, M. A., and Suárez, R. (2015). Grasp Quality Measures: Review and Performance. Auton. Robot 38, 65–88. doi:10.1007/s10514-014-9402-3
Rodriguez, A., Mason, M. T., and Ferry, S. (2012). From Caging to Grasping. Int. J. Robotics Res. 31, 886–900. doi:10.1177/0278364912442972
Rubert, C., Kappler, D., Bohg, J., and Morales, A. (2019). Predicting Grasp success in the Real World - A Study of Quality Metrics and Human Assessment. Robotics Autonomous Syst. 121, 103274. doi:10.1016/j.robot.2019.103274
Saxena, A., Driemeyer, J., and Ng, A. Y. (2008). Robotic Grasping of Novel Objects Using Vision. Int. J. Robotics Res. 27 (2), 157–173. doi:10.1177/0278364907087172
Seminara, L., Gastaldo, P., Watt, S. J., Valyear, K. F., Zuher, F., and Mastrogiovanni, F. (2019). Active Haptic Perception in Robots: a Review. Front. Neurorobot. 13, 53. doi:10.3389/fnbot.2019.00053
Shao, L., Ferreira, F., Jorda, M., Nambiar, V., Luo, J., Solowjow, E., et al. (2019). Unigrasp: Learning a Unified Model to Grasp with N-Fingered Robotic Hands. arXiv 5, 2286–2293.
Shaw-Cortez, W., Oetomo, D., Manzie, C., and Choong, P. (2020). Technical Note for "Tactile-Based Blind Grasping: A Discrete-Time Object Manipulation Controller for Robotic Hands". IEEE Robot. Autom. Lett. 5, 3475–3476. doi:10.1109/LRA.2020.2977585
Solak, G., and Jamone, L. (2019). Learning by Demonstration and Robust Control of Dexterous In-Hand Robotic Manipulation Skills. In 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). 8246–8251. doi:10.1109/IROS40897.2019.8967567
Tomo, T. P., Schmitz, A., Wong, W. K., Kristanto, H., Somlor, S., Hwang, J., et al. (2018). Covering a Robot Fingertip with Uskin: A Soft Electronic Skin with Distributed 3-axis Force Sensitive Elements for Robot Hands. IEEE Robot. Autom. Lett. 3, 124–131. doi:10.1109/LRA.2017.2734965
Veiga, F., van Hoof, H., Peters, J., and Hermans, T. (2015). Stabilizing Novel Objects by Learning to Predict Tactile Slip. In 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). 5065–5072. doi:10.1109/IROS.2015.7354090
Vezzani, G., Pattacini, U., and Natale, L. (2017). A Grasping Approach Based on Superquadric Models. In 2017 IEEE International Conference on Robotics and Automation (ICRA) (IEEE), 1579–1586. doi:10.1109/icra.2017.7989187
Wang, C., Zhang, X., Zang, X., Liu, Y., Ding, G., Yin, W., et al. (2020). Feature Sensing and Robotic Grasping of Objects with Uncertain Information: A Review. Sensors 20, 3707. doi:10.3390/s20133707
Keywords: grasping, manipulation, dexterous hand, haptics, grasp metric, exploration
Citation: Siddiqui MS, Coppola C, Solak G and Jamone L (2021) Grasp Stability Prediction for a Dexterous Robotic Hand Combining Depth Vision and Haptic Bayesian Exploration. Front. Robot. AI 8:703869. doi: 10.3389/frobt.2021.703869
Received: 30 April 2021; Accepted: 12 July 2021;
Published: 12 August 2021.
Edited by:
Eris Chinellato, Middlesex University, United KingdomReviewed by:
Dibyendu Mukherjee, Duke University, United StatesYunchao Tang, Guangxi University, China
Copyright © 2021 Siddiqui, Coppola, Solak and Jamone. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Muhammad Sami Siddiqui, bS5zLnNpZGRpcXVpQHFtdWwuYWMudWs=