A | B | C | D | E | F | G | H | I | J | K | L | M | N | O | P | Q | R | S | V | W | X | Y | Z | AA | AB | AC | AD | AE | AF | AG | AH | AI | AJ | AK | ||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1 | Latest update: 🤖 = latest additions | July 31, 2025 | ||||||||||||||||||||||||||||||||||
2 | Note: drop-down filters only work when open in Google Spreadsheets | Number of entries: 48 | ||||||||||||||||||||||||||||||||||
3 | Data for many entries sourced from Newbury et al. [2023] | Grasp Datasets | Simulators | |||||||||||||||||||||||||||||||||
4 | Name | Description | Planning Method | Training Data | End-effector Hardware | Object Configuration | Input Data | Output Pose | Corresponding Dataset (see repository linked above) | Simulator (see repository linked above) | Backbone | Metric(s) Used | Camera Position(s) | Language | Link(s) | License | Maintainer(s) | Citation | Year (Initial Release) | |||||||||||||||||
5 | 🤖 DexGrasp Anything | DexGrasp Anything is a method that effectively integrates physical constraints into both the training and sampling phases of a diffusion-based generative model for multi-finger dexterous hand grasp generation, achieving state-of-the-art performance across nearly all open datasets. | Generative | Sim, Real | Multi-finger | Singulated | Point cloud | Dexterous grasp | DexGrasp-Anything | | PointTransformer | Grasp success rate, Maximum Penetration, Diversity | | PyTorch, Python | https://github.com/4DVLab/DexGrasp-Anything https://dexgraspanything.github.io/ | MIT | 4DVLab, ShanghaiTech University | Zhong, Yiming, Qi Jiang, Jingyi Yu, and Yuexin Ma. "Dexgrasp anything: Towards universal robotic dexterous grasping with physics awareness." In Proceedings of the Computer Vision and Pattern Recognition Conference, pp. 22584-22594. 2025. | 2025 | |||||||||||||||||
6 | 🤖 FoundationGrasp | A foundation model-based TOG (Task oriented grasping) framework that leverages the open-ended knowledge from foundation models to learn generalizable TOG skills. Built on top of the public TOG benchmark TaskGrasp dataset, we contribute a multi-modal TOG dataset, Language and Vision Augmented TaskGrasp (LaViATaskGrasp) dataset, to support the training and evaluation of generalizable TOG methods. | Grasp evaluation + Sampler | Sim | Two-finger | Singulated | RGBD image, Point cloud, Natural language, Grasp | Grasp success probability | Language and Vision Augmented TaskGrasp (LaViA-TaskGrasp) | | | Grasp success rate | | Python, PyTorch | https://github.com/mkt1412/GraspGPT_public https://sites.google.com/view/foundationgrasp | | Tang, Chao, Dehao Huang, Wenlong Dong, Ruinian Xu, and Hong Zhang. "Foundationgrasp: Generalizable task-oriented grasping with foundation models." IEEE Transactions on Automation Science and Engineering (2025). | 2025 | ||||||||||||||||||
7 | 🤖 GraspGen | We build upon the recent success on modeling the object-centric grasp generation process as an iterative diffusion process. Our proposed framework - GraspGen - consists of a Diffusion-Transformer architecture that enhances grasp generation, paired with an efficient discriminator to score and filter sampled grasps. We introduce a novel and performant on-generator training recipe for the discriminator. | Generative | Sim | Two-finger, Suction | Singulated | Point cloud, Grasp | Grasp success probability | GraspGen | Issac-Sim | PointTransformer | Grasp success rate | | Python, PyTorch | https://graspgen.github.io/ https://github.com/NVlabs/GraspGen | NVIDIA | NVIDIA | Murali, Adithyavairavan, Balakumar Sundaralingam, Yu-Wei Chao, Wentao Yuan, Jun Yamada, Mark Carlson, Fabio Ramos, Stan Birchfield, Dieter Fox, and Clemens Eppner. "GraspGen: A Diffusion-based Framework for 6-DOF Grasping with On-Generator Training." arXiv preprint arXiv:2507.13097 (2025). | 2025 | |||||||||||||||||
8 | 🤖 GraspGPT | We propose GraspGPT, a large language model (LLM) based TOG framework that leverages the open-end semantic knowledge from an LLM to achieve zero-shot generalization to novel concepts. We conduct experiments on Language Augmented TaskGrasp (LA-TaskGrasp) dataset and demonstrate that GraspGPT outperforms existing TOG methods on different held-out settings when generalizing to novel concepts out of the training set. | Grasp evaluation + Sampler | Sim | Two-finger | Singulated | Point cloud, Natural language, Grasp | Grasp success probability | Language and Vision Augmented TaskGrasp (LaViA-TaskGrasp) | | | Grasp success rate | | Python, PyTorch | https://github.com/mkt1412/GraspGPT_public https://sites.google.com/view/graspgpt/ | | Tang, Chao, Dehao Huang, Wenqi Ge, Weiyu Liu, and Hong Zhang. "Graspgpt: Leveraging semantic knowledge from a large language model for task-oriented grasping." IEEE Robotics and Automation Letters 8, no. 11 (2023): 7551-7558. | 2023 | ||||||||||||||||||
9 | 🤖 GraspSAM | In this paper, we introduce GraspSAM, an innovative extension of the Segment Anything Model (SAM) designed for prompt-driven and category-agnostic grasp detection. Unlike previous methods, which are often limited by small-scale training data, GraspSAM leverages SAM’s large-scale training and prompt-based segmentation capabilities to efficiently support both targetobject and category-agnostic grasping. | Direct regression | Sim | Two-finger | Singulated | RGB image | "2D grasp rectangle (x, y, width, height, angle)" | Jacquard, Grasp-Anything, Grasp-Anything++ | | ViT | Grasp success rate | Overhead | Python, PyTorch | https://github.com/gist-ailab/GraspSAM https://gistailab.github.io/graspsam/ | | Noh, Sangjun, Jongwon Kim, Dongwoo Nam, Seunghyeok Back, Raeyoung Kang, and Kyoobin Lee. "Graspsam: When segment anything model meets grasp detection." arXiv preprint arXiv:2409.12521 (2024). | 2025 | ||||||||||||||||||
10 | 🤖 NeuGraspNet | NeuGraspNet is a novel, fully implicit 6DoF grasp detection method that re-interprets robotic grasping as surface rendering and predicts high-fidelity grasps from any random single viewpoint of a scene. Our method exploits a learned implicit geometric scene representation to perform global and local surface rendering. This enables effective grasp candidate generation (using global features) and grasp quality prediction (using local features from a shared feature space). | Grasp evaluation + Sampler | Sim | Two-finger | Piled, Stacked | Depth image | Grasp success probability | | PyBullet | PointNet | Grasp success rate, Declutter rate | | Python, PyTorch | https://github.com/pearl-robot-lab/neugraspnet https://sites.google.com/view/neugraspnet | | Jauhri, Snehal, Ishikaa Lunawat, and Georgia Chalvatzaki. "Learning Any-View 6DoF Robotic Grasping in Cluttered Scenes via Neural Surface Rendering." In RoboNerF: 1st Workshop On Neural Fields In Robotics at ICRA 2024. | 2024 | ||||||||||||||||||
11 | 🤖 ShapeGrasp | Task-oriented grasping of unfamiliar objects is a necessary skill for robots in dynamic in-home environments. Inspired by the human capability to grasp such objects through intuition about their shape and structure, we present a novel zero-shot task-oriented grasping method leveraging a geometric decomposition of the target object into simple, convex shapes that we represent in a graph structure, including geometric attributes and spatial relationships. Our approach employs minimal essential information—the object's name and the intended task—to facilitate zero-shot task-oriented grasping. | Grasp evaluation + Sampler | Training-less | Three-finger | Singulated | RGBD image | 3D position and angle in table plane | | | | Grasp success rate, Part selection acc. | Overhead | Python | https://github.com/samwli/ShapeGrasp https://shapegrasp.github.io/ | MIT | | Li, Samuel, Sarthak Bhagat, Joseph Campbell, Yaqi Xie, Woojun Kim, Katia Sycara, and Simon Stepputtis. "Shapegrasp: Zero-shot task-oriented grasping with large language models through geometric decomposition." In 2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 10527-10534. IEEE, 2024. | 2024 | |||||||||||||||||
12 | 🤖 UniGraspTransformer | We introduce UniGraspTransformer, a universal Transformer-based network for dexterous robotic grasping that simplifies training while enhancing scalability and performance. Unlike prior methods such as UniDexGrasp++, which require complex, multi-step training pipelines, UniGraspTransformer follows a streamlined process: first, dedicated policy networks are trained for individual objects using reinforcement learning to generate successful grasp trajectories; then, these trajectories are distilled into a single, universal network. | Reinforcement learning, Direct regression | Sim | Multi-finger | Singulated | Point cloud | Dexterous grasp | | Issac-Gym | | Grasp success rate | Multi-view | PyTorch, Python | https://github.com/microsoft/UniGraspTransformer https://dexhand.github.io/UniGraspTransformer/ | MIT | Microsoft | Wang, Wenbo, Fangyun Wei, Lei Zhou, Xi Chen, Lin Luo, Xiaohan Yi, Yizhong Zhang et al. "Unigrasptransformer: Simplified policy distillation for scalable dexterous robotic grasping." In Proceedings of the Computer Vision and Pattern Recognition Conference, pp. 12199-12208. 2025. | 2025 | |||||||||||||||||
13 | 🤖 ZeroGrasp | ZeroGrasp is a framework for simultaneous reconstruction and grasp pose prediction of multiple unseen objects from a single RGB-D image in near real-time. ZeroGrasp achieves the state-of-the-art results across all splits on GraspNet-1B benchmark and demonstrates strong generalization and robustoness in real-world scenarios. | Grasp evaluation + Sampler | Sim | Two-finger | Cluttered | RGBD image | Grasp success probability | ZeroGrasp-11B | Issac-Gym | | Average Precision | Overhead | Python, PyTorch | https://github.com/sh8/ZeroGrasp https://sh8.io/#/zerograsp | CC BY-NC 4.0 | Iwase, Shun, Muhammad Zubair Irshad, Katherine Liu, Vitor Guizilini, Robert Lee, Takuya Ikeda, Ayako Amma et al. "ZeroGrasp: Zero-Shot Shape Reconstruction Enabled Robotic Grasping." In Proceedings of the Computer Vision and Pattern Recognition Conference, pp. 17405-17415. 2025. | 2025 | ||||||||||||||||||
14 | 6-DoF GraspNet | 6-DoF GraspNet formulates the problem of grasp generation as sampling a set of grasps using a variational autoencoder and assess and refine the sampled grasps using a grasp evaluator model. Both Grasp Sampler and Grasp Refinement networks take 3D point clouds observed by a depth camera as input. Our model is trained purely in simulation and works in the real world without any extra steps. | Sampling | Sim | Two-finger | Singulated | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | ACRONYM | PyRender | PointNet++ | Image-based Grasp Quality (IGQ), Grasp success rate, Precision, Robust grasp rate, Planning time | Eye-in-hand | Python, TensorFlow | https://github.com/NVlabs/6dof-graspnet | MIT | NVIDIA | Mousavian, Arsalan, Clemens Eppner, and Dieter Fox. "6-dof graspnet: Variational grasp generation for object manipulation." In Proceedings of the IEEE/CVF international conference on computer vision, pp. 2901-2910. 2019. | 2019 | |||||||||||||||||
15 | AnyGrasp | Utilization of grasp correspondence across observations enables dynamic grasp tracking. Our model, AnyGrasp, can generate accurate, full-DoF, dense and temporally-smooth grasp poses efficiently, and works robustly against large depth sensing noise. | Direct regression | Real, Sim | Two-finger | Piled | Point cloud | "7-DoF grasp pose (x, y, z, r, p, y, gripper width)" | GraspNet-1Billion | PyRender | PointNet++ | Grasp success rate, Completion rate | Eye-in-hand, Overhead | Python, PyTorch | https://github.com/graspnet/anygrasp_sdk https://graspnet.net/anygrasp | CC BY-NC-SA | Machine Vision and Intelligence Group, Shanghai Jiao Tong University | Fang, Hao-Shu, Chenxi Wang, Hongjie Fang, Minghao Gou, Jirong Liu, Hengxu Yan, Wenhai Liu, Yichen Xie, and Cewu Lu. "Anygrasp: Robust and efficient grasp perception in spatial and temporal domains." IEEE Transactions on Robotics 39, no. 5 (2023): 3929-3945. | 2023 | |||||||||||||||||
16 | CAPGrasp | CAPGrasp is an R3 × SO(2)-equivariant 6-Degrees of Freedom (DoF) continuous approach-constrained generative grasp sampler. It includes a novel learning strategy for training CAPGrasp that eliminates the need to curate massive conditionally labeled datasets and a constrained grasp refinement technique that improves grasp poses while respecting the grasp approach directional constraints. | Sampling | Sim, Real | Two-finger | Singulated | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | ACRONYM, Custom | Issac-Gym | VAE | Grasp success rate | Side view | Python | https://github.com/wengzehang/CAPGrasp https://wengzehang.github.io/CAPGrasp/ | | Weng, Zehang, Haofei Lu, Jens Lundell, and Danica Kragic. "CAPGrasp: An $\mathbb {R}^{3}\times\rm {SO (2)-equivariant} $ Continuous Approach-Constrained Generative Grasp Sampler." IEEE Robotics and Automation Letters (2024). | 2024 | ||||||||||||||||||
17 | CaTGrasp | This work proposes a framework to learn task-relevant grasping for industrial objects without the need of time-consuming real-world data collection or manual annotation. To achieve this, the entire framework is trained solely in simulation, including supervised training with synthetic label generation and self-supervised, hand-object interaction. In the context of this framework, CaTGrasp uses a novel, object-centric canonical representation at the category level, which allows establishing dense correspondence across object instances and transferring task-relevant grasps to novel instances. | Sampling | Sim | Two-finger | Piled | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | Custom | PyBullet | PointNet | Grasp success rate | Overhead | Python, PyTorch | https://github.com/wenbowen123/catgrasp https://sites.google.com/view/catgrasp | MIT, Apache 2.0 | Bowen Wen | Wen, Bowen, Wenzhao Lian, Kostas Bekris, and Stefan Schaal. "Catgrasp: Learning category-level task-relevant grasping in clutter from simulation." In 2022 International Conference on Robotics and Automation (ICRA), pp. 6401-6408. IEEE, 2022. | 2022 | |||||||||||||||||
18 | Contact-GraspNet | Contact-GraspNet treats 3D points of the recorded point cloud as potential grasp contacts. By rooting the full 6-DoF grasp pose and width in the observed point cloud, we can reduce the dimensionality of our grasp representation to 4-DoF which greatly facilitates the learning process. Our class-agnostic approach is trained on 17 million simulated grasps and generalizes well to real world sensor data. | Direct regression | Sim | Two-finger | Structured | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | ACRONYM | NVIDIA FleX | PointNet++ | Grasp success rate, Grasp success at first trial, Number of re-grasps | Oblique view | Python, TensorFlow | https://github.com/NVlabs/contact_graspnet | | NVIDIA | Sundermeyer, Martin, Arsalan Mousavian, Rudolph Triebel, and Dieter Fox. "Contact-graspnet: Efficient 6-dof grasp generation in cluttered scenes." In 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 13438-13444. IEEE, 2021. | 2021 | |||||||||||||||||
19 | Deep Dexterous Grasping (DDG) | Deep Dexterous Grasping (DDG) uses deep neural networks to improve the state-of-the-art in dexterous grasping of previously unseen objects using a single view camera. | Sampling | Sim | Multi-finger | Singulated | Depth image | "6-DoF grasp pose (x, y, z, r, p, y)" | Deep Dexterous Grasping (DDG) | MuJoCo | VGG-16, ResNet-50 | Grasp success rate (top ranked) | Oblique view | C++ | https://github.com/rusen/Grasp/ https://rusen.github.io/DDG/ | GPL-3.0 | Umit Rusen Aktas | Aktaş, Ümit Ruşen, Chao Zhao, Marek Kopicki, and Jeremy L. Wyatt. "Deep dexterous grasping of novel objects from a single view." International Journal of Humanoid Robotics 19, no. 02 (2022): 2250011. | 2019 | |||||||||||||||||
20 | Deep Dexterous Grasping in Clutter (DDGC) | Deep Dexterous Grasping in Clutter (DDGC) takes an RGB image of the complete scene and the mask of the object to grasp and predicts a multi-fingered robotic grasp on the target objects. First, the objects' shapes are completed. Next, grasps are generated on the target object by first feeding the RGB image and the mask through an image encoder, and based on the image encoding 6D multi-finger grasp poses and configurations are generated. Finally, the hand is refined to be close to the surface of the target object but not in collision with it by using our parameter-free fully differentiable barrett kinematics layer. | Sampling | Sim | Three-finger | Structured | RGBD image | "6-DoF grasp pose (x, y, z, r, p, y)" | Custom | PyBullet, GraspIt! | | ε-quality grasps (10 top-scoring), Average clearance rate, Average grasp success rate, Grasp sampling time (10 grasps) | Oblique view | Python, PyTorch | https://github.com/aalto-intelligent-robotics/DDGC https://irobotics.aalto.fi/ddgc/ | MIT | Aalto University Intelligent Robotics | Lundell, Jens, Francesco Verdoja, and Ville Kyrki. "Ddgc: Generative deep dexterous grasping in clutter." IEEE Robotics and Automation Letters 6, no. 4 (2021): 6899-6906. | 2021 | |||||||||||||||||
21 | DeepRLManip | Our approach is to formulate the problem as a Markov decision process (MDP) with abstract yet generally applicable state and action representations. Finding a good solution to the MDP requires adding constraints on the allowed actions. We develop a specific set of constraints called hierarchical SE(3) sampling (HSE3S) which causes the robot to learn a sequence of gazes to focus attention on the task-relevant parts of the scene. | Reinforcement learning | Sim | Two-finger | Structured | Depth image | Grasping Policy | Custom | OpenRAVE | DQN with gradient Monte Carlo | Average grasp success rate | Eye-in-hand | Python, MATLAB | https://github.com/mgualti/DeepRLManip | MIT | Marcus Gualtieri | Gualtieri, Marcus, and Robert Platt. "Learning 6-dof grasping and pick-place using attention focus." In Conference on Robot Learning, pp. 477-486. PMLR, 2018. | 2018 | |||||||||||||||||
22 | Dex-Net 1.0 (MV-CNNs) | The Dexterity Network (Dex-Net) is a research project including code, datasets, and algorithms for generating datasets of synthetic point clouds, robot parallel-jaw grasps and metrics of grasp robustness based on physics for thousands of 3D object models to train machine learning-based methods to plan robot grasps. | Sampling, Exemplar methods | Sim | Two-finger | Singulated | 3D object model mesh | 3D position and approach vector | DexNet 1.0 | AUTOLAB | AlexNet | Probability of force closure | Overhead | Python, TensorFlow | https://github.com/BerkeleyAutomation/dex-net | Apache 2.0 | Berkeley Automation Lab | Mahler, Jeffrey, Florian T. Pokorny, Brian Hou, Melrose Roderick, Michael Laskey, Mathieu Aubry, Kai Kohlhoff, Torsten Kröger, James Kuffner, and Ken Goldberg. "Dex-net 1.0: A cloud-based network of 3d objects for robust grasp planning using a multi-armed bandit model with correlated rewards." In 2016 IEEE international conference on robotics and automation (ICRA), pp. 1957-1964. IEEE, 2016. | 2016 | |||||||||||||||||
23 | Dex-Net 2.0 (GQ-CNN) | The gqcnn Python package is for training and analysis of Grasp Quality Convolutional Neural Networks (GQ-CNNs). It is part of the ongoing Dexterity-Network (Dex-Net) project created and maintained by the AUTOLAB at UC Berkeley. | Direct regression | Real | Two-finger | Singulated | Depth image | "2D grasp rectangle (x, y, width, height, angle)", Grasp success probability | DexNet 2.0 | AUTOLAB | None | Image-based Grasp Quality (IGQ), Grasp success rate, Precision, Robust grasp rate, Planning time | Overhead | Python, TensorFlow | https://github.com/BerkeleyAutomation/gqcnn https://berkeleyautomation.github.io/gqcnn/ | | Berkeley Automation Lab | Mahler, Jeffrey, Jacky Liang, Sherdil Niyaz, Michael Laskey, Richard Doan, Xinyu Liu, Juan Aparicio Ojea, and Ken Goldberg. "Dex-net 2.0: Deep learning to plan robust grasps with synthetic point clouds and analytic grasp metrics." arXiv preprint arXiv:1703.09312 (2017). | 2017 | |||||||||||||||||
24 | Dex-Net 2.1 | We model bin picking with a discrete-time Partially Observable Markov Decision Process that specifies states of the heap, point cloud observations, and rewards. We collect synthetic demonstrations of bin picking from an algorithmic supervisor uses full state information to optimize for the most robust collision-free grasp in a forward simulator based on pybullet to model dynamic object-object interactions and robust wrench space analysis from the Dexterity Network (Dex-Net) to model quasi-static contact between the gripper and object. | Reinforcement learning | Sim | Two-finger | Piled | Rollouts | Grasping Policy | DexNet 2.1 | PyBullet | None | Grasp success rate, Average clearance rate, Picks per Hour (PPH), Average Precision | Overhead | | http://bit.ly/3XxKVro | | Berkeley Automation Lab | Mahler, Jeffrey, and Ken Goldberg. "Learning deep policies for robot bin picking by simulating robust grasping sequences." In Conference on robot learning, pp. 515-524. PMLR, 2017. | 2017 | |||||||||||||||||
25 | Dex-Net 3.0 | We propose a compliant suction contact model that computes the quality of the seal between the suction cup and target object and determines whether or not the suction grasp can resist an external wrench (e.g. gravity) on the object. To evaluate a grasp, we measure robustness to perturbations in end-effector and object pose, material properties, and external wrenches. We train a Grasp Quality Convolutional Neural Network (GQ-CNN) on this dataset to classify suction grasp robustness from point clouds. | Direct regression | Sim | Suction | Singulated | Point cloud | Grasping Policy | DexNet 3.0 | N/A | None | Grasp success rate, Planning time, Average Precision | Overhead | | http://bit.ly/3XEPVuo | | Berkeley Automation Lab | Mahler, Jeffrey, Matthew Matl, Xinyu Liu, Albert Li, David Gealy, and Ken Goldberg. "Dex-net 3.0: Computing robust vacuum suction grasp targets in point clouds using a new analytic model and deep learning." In 2018 IEEE International Conference on robotics and automation (ICRA), pp. 5620-5627. IEEE, 2018. | 2018 | |||||||||||||||||
26 | Dex-Net 4.0 | We present Dexterity Network (Dex-Net) 4.0, a substantial extension to previous versions of Dex-Net that learns policies for a given set of grippers by training on synthetic datasets using domain randomization with analytic models of physics and geometry. We train policies for a parallel-jaw and a vacuum-based suction cup gripper on 5 million synthetic depth images, grasps, and rewards generated from heaps of three-dimensional objects. | Reinforcement learning | Sim | Two-finger, Suction | Piled | Depth image | Grasping Policy | DexNet 4.0 | N/A | None | Grasp success rate, Average clearance rate, Picks per Hour (PPH), Average Precision | Overhead | | http://bit.ly/3ViM4BG | | Berkeley Automation Lab | Mahler, Jeffrey, Matthew Matl, Vishal Satish, Michael Danielczuk, Bill DeRose, Stephen McKinley, and Ken Goldberg. "Learning ambidextrous robot grasping policies." Science Robotics 4, no. 26 (2019): eaau4984. | 2019 | |||||||||||||||||
27 | Edge Grasp Network | Edge Grasp Network is a novel method and neural network model that enables better grasp success rates relative to what is available in the literature. The method takes standard point cloud data as input and works well with single-view point clouds observed from arbitrary viewing directions. | Sampling | Sim | Two-finger | Piled | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | Custom | PyBullet | PointNet | Grasp success rate, Declutter rate | Eye-in-hand | Python | https://github.com/HaojHuang/Edge-Grasp-Network | MIT | Haojie Huang | Huang, Haojie, Dian Wang, Xupeng Zhu, Robin Walters, and Robert Platt. "Edge grasp network: A graph-based se (3)-invariant approach to grasp detection." In 2023 IEEE International Conference on Robotics and Automation (ICRA), pp. 3882-3888. IEEE, 2023. | 2023 | |||||||||||||||||
28 | EquiGraspFlow | EquiGraspFlow is a flow-based SE(3)-equivariant 6-DoF grasp pose generative model that can learn complex conditional distributions on the SE(3) manifold while guaranteeing SE(3)-equivariance. Our model achieves the equivariance without relying on data augmentation, by using network architectures that guarantee it by construction. | Sampling | Sim, Real | Two-finger | Structured | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | ACRONYM | Issac-Gym | VN-DGCNN | Grasp success rate, Earth Mover’s Distance (EMD) | Eye-in-hand | Python | https://github.com/bdlim99/EquiGraspFlow https://equigraspflow.github.io/ | MIT | Lim, Byeongdo, Jongmin Kim, Jihwan Kim, Yonghyeon Lee, and Frank C. Park. "Equigraspflow: Se (3)-equivariant 6-dof grasp pose generative flows." In 8th Annual Conference on Robot Learning. 2024. | 2024 | ||||||||||||||||||
29 | GCNGrasp | The GCNGrasp framework uses the semantic knowledge of objects and tasks encoded in a knowledge graph to generalize to new object instances, classes and even new tasks. Our framework shows a significant improvement of around 12% on held-out settings compared to baseline methods which do not use semantics. | Sampling | Sim | Two-finger | Singulated | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | TaskGrasp | Open3D | PointNet++ | | Eye-in-hand | Python, PyTorch | https://github.com/adithyamurali/TaskGrasp https://sites.google.com/view/taskgrasp | MIT | Adithya Murali | Murali, Adithyavairavan, Weiyu Liu, Kenneth Marino, Sonia Chernova, and Abhinav Gupta. "Same object, different grasps: Data and semantic knowledge for task-oriented grasping." In Conference on robot learning, pp. 1540-1557. PMLR, 2021. | 2021 | |||||||||||||||||
30 | geometric-object-grasper | This repository is an implementation of the paper 'A Geometric Approach for Grasping Unknown Objects with Multi-Fingered Hands' in PyBullet. Note that this is a Python implementation and as a result it runs slower than the initial C++ implementation. | Analytical | Sim | Three-finger | Piled | Point cloud | Grasp preshape | Custom | MuJoCo | VAE | Grasp success rate | Oblique view | Python, PyTorch | https://github.com/mkiatos/geometric-object-grasper | | Marios Kiatos | Kiatos, Marios, Sotiris Malassiotis, and Iason Sarantopoulos. "A geometric approach for grasping unknown objects with multifingered hands." IEEE Transactions on Robotics 37, no. 3 (2020): 735-746. | 2021 | |||||||||||||||||
31 | GeomPickPlace | This approach uses (a) object instance segmentation and shape completion to model the objects and (b) a regrasp planner to decide grasps and places displacing the models to their goals. It is critical for the planner to account for uncertainty in the perceived models, as object geometries in unobserved areas are just guesses, so we account for perceptual uncertainty by incorporating it into the regrasp planner's cost function. | Analytical | Sim | Two-finger | Structured | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | Custom | OpenRAVE | PointNet | Place success rate, Grasp success rate, Number of grasp attempts, Number of re-grasps, Packing height of 5 cm | Overhead | Python, TensorFlow | https://github.com/mgualti/GeomPickPlace | MIT | Marcus Gualteri | Gualtieri, Marcus, and Robert Platt. "Robotic pick-and-place with uncertain object instance segmentation and shape completion." IEEE robotics and automation letters 6, no. 2 (2021): 1753-1760. | 2021 | |||||||||||||||||
32 | ggcnn_plus | In this work, we focus on improving an existing shallow network, GG-CNN, and propose a new orthographic pipeline to enable the use of these models independently of the orientation of the camera. | Direct regression | Sim | Two-finger | Singulated | RGBD image | "2D grasp rectangle (x, y, width, height, angle)" | Cornell Grasping Dataset | PyBullet | | | | Python, Keras | https://github.com/m-rios/ggcnn_plus | | Mario Rios Munoz | Rios Munoz, Mario. "Grasping in 6DoF: An Orthographic Approach to Generalized Grasp Affordance Predictions." PhD dissertation, 2021. | 2021 | |||||||||||||||||
33 | Goal-Auxiliary Deep Deterministic Policy Gradient (GA-DDPG) | We propose a new method for learning closed-loop control policies for 6D grasping. Our policy takes a segmented point cloud of an object from an egocentric camera as input, and outputs continuous 6D control actions of the robot gripper for grasping the object. We combine imitation learning and reinforcement learning and introduce a goal-auxiliary actor-critic algorithm for policy learning. | Reinforcement learning | Sim | Two-finger | Singulated | Point cloud | Grasping Policy | ACRONYM | PyBullet | PointNet++ | Grasp success rate, Number of grasp attempts | Eye-in-hand | Python, PyTorch | https://github.com/liruiw/GA-DDPG https://sites.google.com/view/gaddpg | MIT | Lirui (Leroy) Wang | Wang, Lirui, Yu Xiang, Wei Yang, Arsalan Mousavian, and Dieter Fox. "Goal-auxiliary actor-critic for 6d robotic grasping with point clouds." In Conference on Robot Learning, pp. 70-80. PMLR, 2022. | 2022 | |||||||||||||||||
34 | GRAFF (Grasp-Affordances) | GRAFF is a deep RL dexterous robotic grasping policy that uses visual affordance priors learnt from humans for functional grasping. Our proposed model, called GRAFF for Grasp-Affordances, consists of two stages. First, we train a network to predict affordance regions from static images. Second, we train a dynamic grasping policy using the learned affordances. The key upshots of our approach are better grasping, faster learning, and generalization to successfully grasp objects unseen during policy training. | Reinforcement learning | Sim | Multi-finger | Singulated | RGBD image | Grasping Policy | Custom | MuJoCo, Open3D | ResNet-50 | | Eye-in-hand | Python, TensorFlow, PyTorch | https://github.com/priyankamandikal/graff https://vision.cs.utexas.edu/projects/graff-dexterous-affordance-grasp/ | MIT | Priyanka Mandikal | Mandikal, Priyanka, and Kristen Grauman. "Learning dexterous grasping with object-centric visual affordances." In 2021 IEEE international conference on robotics and automation (ICRA), pp. 6169-6176. IEEE, 2021. | 2021 | |||||||||||||||||
35 | Grasp detection via Implicit Geometry and Affordance (GIGA) | GIGA is a network that jointly detects 6 DOF grasp poses and reconstruct the 3D scene. GIGA takes advantage of deep implicit functions, a continuous and memory-efficient representation, to enable differentiable training of both tasks. GIGA takes as input a Truncated Signed Distance Function (TSDF) representation of the scene, and predicts local implicit functions for grasp affordance and 3D occupancy. By querying the affordance implict functions with grasp center candidates, we can get grasp quality, grasp orientation and gripper width at these centers. | Sampling | Sim | Two-finger | Piled | Truncated Signed Distance Function (TSDF) | "6-DoF grasp pose (x, y, z, r, p, y)" | Custom | PyBullet, Open3D | ConvONets | | Eye-in-hand | Python, PyTorch | https://github.com/UT-Austin-RPL/GIGA https://sites.google.com/view/rpl-giga2021 | MIT | UT Robot Perception and Learning Lab | Jiang, Zhenyu, Yifeng Zhu, Maxwell Svetlik, Kuan Fang, and Yuke Zhu. "Synergies between affordance and geometry: 6-dof grasp detection via implicit representations." arXiv preprint arXiv:2104.01542 (2021). | 2021 | |||||||||||||||||
36 | Grasp Pose Detection (GPD) | Grasp Pose Detection (GPD) is a package to detect 6-DOF grasp poses (3-DOF position and 3-DOF orientation) for a 2-finger robot hand (e.g., a parallel jaw gripper) in 3D point clouds. GPD takes a point cloud as input and produces pose estimates of viable grasps as output. The main strengths of GPD are: works for novel objects (no CAD models required for detection), works in dense clutter, and outputs 6-DOF grasp poses (enabling more than just top-down grasps). | Sampling | Sim | Two-finger | Piled | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | Custom | OpenRAVE | 3DNet | Number of objects, Number of grasp attempts, Number of successful grasps, Percentage of successful grasps, Percentage of objects removed | Eye-in-hand | C++, Python, PyTorch | https://github.com/atenpas/gpd | BSD 2-Clause | Andreas Ten Pas | Ten Pas, Andreas, Marcus Gualtieri, Kate Saenko, and Robert Platt. "Grasp pose detection in point clouds." The International Journal of Robotics Research 36, no. 13-14 (2017): 1455-1473. | 2017 | |||||||||||||||||
37 | Grasp Proposal Network (GPNet) | The end-to-end Grasp Proposal Network (GPNet) predicts a diverse set of 6-DOF grasps for an unseen object observed from a single and unknown camera view. GPNet builds on a key design of grasp proposal module that defines anchors of grasp centers at discrete but regular 3D grid corners, which is flexible to support either more precise or more diverse grasp predictions. | Direct regression | Sim | Two-finger | Singulated | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | Grasp Proposal Network (GPNet) | PyBullet | PointNet++ | Grasp success rate, Coverage rate | Agnostic | Python, PyTorch | https://github.com/Gorilla-Lab-SCUT/GPNet | | South China University of Technology | Wu, Chaozheng, Jian Chen, Qiaoyu Cao, Jianchi Zhang, Yunxin Tai, Lin Sun, and Kui Jia. "Grasp proposal networks: An end-to-end solution for visual learning of robotic grasps." Advances in Neural Information Processing Systems 33 (2020): 13174-13184. | 2020 | |||||||||||||||||
38 | GraspXL | We unify the generation of hand-object grasping motions across multiple motion objectives, diverse object shapes and dexterous hand morphologies in a policy learning framework GraspXL. The objectives are composed of the graspable area, heading direction, wrist rotation, and hand position. | Reinforcement learning | Sim | Multi-finger | Singulated | 3D object model mesh | "6-DoF grasp pose (x, y, z, r, p, y)", Dexterous grasp | GraspXL | RaiSim | PPO | Grasp success rate | NA | C++, Python, PyTorch | https://github.com/zdchan/graspxl https://eth-ait.github.io/graspxl/ | MIT, BSD 3-Clause | Hui Zhang | Zhang, Hui, Sammy Christen, Zicong Fan, Otmar Hilliges, and Jie Song. "Graspxl: Generating grasping motions for diverse objects at scale." In European Conference on Computer Vision, pp. 386-403. Cham: Springer Nature Switzerland, 2024. | 2024 | |||||||||||||||||
39 | ICG-Net | The method uses pointcloud data from a single arbitrary viewing direction as an input and generates an instance-centric representation for each partially observed object in the scene. This representation is further used for object reconstruction and grasp detection in cluttered table-top scenes. | Sampling | Sim, Real | Two-finger | Structured | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | VGN | PyBullet | UNet | Grasp success rate, Declutter rate | Oblique view | Python | https://github.com/renezurbruegg/icg_net | MIT | Zurbrügg, René, Yifan Liu, Francis Engelmann, Suryansh Kumar, Marco Hutter, Vaishakh Patil, and Fisher Yu. "ICGNet: a unified approach for instance-centric grasping." In 2024 IEEE International Conference on Robotics and Automation (ICRA), pp. 4140-4146. IEEE, 2024. | 2024 | ||||||||||||||||||
40 | Multi-FinGAN | Multi-FinGAN is a fast generative multi-finger grasp sampling method that synthesizes high quality grasps directly from RGB-D images in about a second. We achieve this by training in an end-to-end fashion a coarse-to-fine model composed of a classification network that distinguishes grasp types according to a specific taxonomy and a refinement network that produces refined grasp poses and joint angles. | Sampling | Sim | Three-finger | Singulated | RGBD image | "6-DoF grasp pose (x, y, z, r, p, y)" | Multi-FinGAN | GraspIt! | ResNet-50 | ε-quality grasps, Grasp success rate, Grasp evaluation time | Oblique view | Python, PyTorch | https://github.com/aalto-intelligent-robotics/Multi-FinGAN | MIT | Aalto University Intelligent Robotics | Lundell, Jens, Enric Corona, Tran Nguyen Le, Francesco Verdoja, Philippe Weinzaepfel, Grégory Rogez, Francesc Moreno-Noguer, and Ville Kyrki. "Multi-fingan: Generative coarse-to-fine sampling of multi-finger grasps." In 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 4495-4501. IEEE, 2021. | 2021 | |||||||||||||||||
41 | MultiModalGrasping | We propose an approach to multi-modal grasp detection that jointly predicts the probabilities that several types of grasps succeed at a given grasp pose. Given a partial point cloud of a scene, the algorithm proposes a set of feasible grasp candidates, then estimates the probabilities that a grasp of each type would succeed at each candidate pose. Predicting grasp success probabilities directly from point clouds makes our approach agnostic to the number and placement of depth sensors at execution time. | Sampling | Sim | Multi-finger | Structured | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | Custom | Drake | PointNet++ | Grasp success rate, Number of grasp attempts, Object removal rate | Agnostic | Python, TensorFlow | https://github.com/mattcorsaro1/MultiModalGrasping | MIT | Matt Corsaro | Corsaro, Matt, Stefanie Tellex, and George Konidaris. "Learning to detect multi-modal grasps for dexterous grasping in dense clutter." In 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 4647-4653. IEEE, 2021. | 2021 | |||||||||||||||||
42 | OrbitGrasp (EquiFormerV2) | Our main contribution is to propose an SE(3)-equivariant model that maps each point in the cloud to a continuous grasp quality function over the 2-sphere S 2 using spherical harmonic basis functions. Compared with reasoning about a finite set of samples, this formulation improves the accuracy and efficiency of our model when a large number of samples would otherwise be needed. In order to accomplish this, we propose a novel variation on EquiFormerV2 that leverages a UNet-style encoder-decoder architecture to enlarge the number of points the model can handle. | Direct regression | Sim | Two-finger | Structured | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | Custom | PyBullet | UNet | Grasp success rate, Grasp success rates averaged over independent episodes | Oblique view | Python | https://github.com/BoceHu/orbitgrasp https://orbitgrasp.github.io/ | MIT | Boce Hu | Hu, B., Zhu, X., Wang, D., Dong, Z., Huang, H., Wang, C., ... & Platt, R. (2024). OrbitGrasp: $ SE (3) $-Equivariant Grasp Learning. arXiv preprint arXiv:2407.03531. | 2024 | |||||||||||||||||
43 | PointNet++ Grasping | PointNet++ Grasping is an end-to-end approach to directly predict the poses, categories and scores (qualities) of all the grasps. It takes the whole sparse point clouds as the input and requires no sampling or search process. Moreover, to generate training data of multi-object scene, we propose a fast multi-object grasp detection algorithm based on Ferrari Canny metrics. | Direct regression | Sim | Two-finger | Piled | Point cloud | "7-DoF grasp pose (x, y, z, r, p, y, approach angle)" | PointNet++ Grasping | PyBullet | PointNet++ | Grasp success rate, Percentage of objects removed | Eye-in-hand | Python | https://github.com/pyni/PointNet2_Grasping_Data_Part/tree/master | | P. Ni, W. Zhang, X. Zhu, and Q. Cao, “PointNet++ Grasping: Learning An End-to-end Spatial Grasp Generation Algorithm from Sparse Point Clouds,” Mar. 21, 2020, arXiv: arXiv:2003.09644. doi: 10.48550/arXiv.2003.09644. | 2021 | ||||||||||||||||||
44 | PointNetGPD | PointNetGPD is an end-to-end grasp evaluation model to address the challenging problem of localizing robot grasp configurations directly from the point cloud. PointNetGPD is light-weighted and can directly process the 3D point cloud that locates within the gripper for grasp evaluation. Taking the raw point cloud as input, our proposed grasp evaluation network can capture the complex geometric structure of the contact area between the gripper and the object even if the point cloud is very sparse. | Sampling | Sim | Two-finger | Piled | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | PointNetGPD | GraspIt! | PointNet | Grasp success rate, Completion rate | Multi-view | Python, PyTorch | https://github.com/lianghongzhuo/PointNetGPD https://lianghongzhuo.github.io/PointNetGPD/ | MIT | Hongzhuo Liang | Liang, Hongzhuo, Xiaojian Ma, Shuang Li, Michael Görner, Song Tang, Bin Fang, Fuchun Sun, and Jianwei Zhang. "Pointnetgpd: Detecting grasp configurations from point sets." In 2019 International Conference on Robotics and Automation (ICRA), pp. 3629-3635. IEEE, 2019. | 2019 | |||||||||||||||||
45 | Probablistic Multi-fingered Grasp Planner | This repo has several learning-based multi-fingered grasp planners implemented. We proposed multiple machine leanring models to predict the probability of grasp success from visual information of the object and grasp configuration. We then formulated grasp planning as inferring the grasp configuration which maximizes the probability of grasp success inside the grasp prediction deep networks. | Active learning | Real | Multi-finger | Singulated | Point cloud | Grasp preshape | Custom | | | Grasp success rate | Oblique view | Python, TensorFlow | https://github.com/QingkaiLu/multi-fingered_grasp_planners https://robot-learning.cs.utah.edu/project/grasp_active_learning | | Qingkai Lu | Lu, Qingkai, Mark Van der Merwe, and Tucker Hermans. "Multi-fingered active grasp learning." In 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 8415-8422. IEEE, 2020. | 2020 | |||||||||||||||||
46 | REgion-based Grasp Network (REGNet) | REGNet is a REgion-based Grasp Network for End-to-end Grasp Detection in Point Clouds. It aims at generating the optimal grasp of novel objects from partial noisy observations. | Direct regression | Sim | Two-finger | Piled | Point cloud | "7-DoF grasp pose (x, y, z, r, p, y, approach angle)" | REgion-based Grasp Network (REGNet) | MuJoCo | PointNet++ | Grasp success rate, Completion rate | Agnostic | Python, PyTorch | https://github.com/zhaobinglei/REGNet_for_3D_Grasping | MIT | Xi'an Jiaotong University | Zhao, B., H. Zhang, X. Lan, H. Wang, Z. Tian, and N. Zheng. "REGNet: REgion-based grasp network for single-shot grasp detection in point clouds. arXiv." arXiv preprint arXiv:2002.12647 (2020). | 2020 | |||||||||||||||||
47 | RGBD-Grasp | RGBD-Grasp is a pipeline that decouples 7-DoF grasp detection into two sub-tasks where RGB and depth information are processed separately. In the first stage, an encoder-decoder like convolutional neural network Angle-View Net(AVN) is proposed to predict the SO(3) orientation of the gripper at every location of the image. Consequently, a Fast Analytic Searching (FAS) module calculates the opening width and the distance of the gripper to the grasp point. | Direct regression | Sim, Real | Two-finger | Structured | RGBD image | "7-DoF grasp pose (x, y, z, r, p, y, gripper width)" | GraspNet-1Billion | | ResNet-50 | Grasp success rate | Eye-in-hand | Python, PyTorch | https://github.com/GouMinghao/rgb_matters | MIT | Shanghai Jiao Tong University | Gou, Minghao, Hao-Shu Fang, Zhanda Zhu, Sheng Xu, Chenxi Wang, and Cewu Lu. "Rgb matters: Learning 7-dof grasp poses on monocular rgbd images." In 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 13459-13466. IEEE, 2021. | 2021 | |||||||||||||||||
48 | Robust Grasp Planning Over Uncertain Shape Completions | We present a method for planning robust grasps over uncertain shape completed objects. For shape completion, a deep neural network is trained to take a partial view of the object as input and outputs the completed shape as a voxel grid. The key part of the network is dropout layers which are enabled not only during training but also at run-time to generate a set of shape samples representing the shape uncertainty through Monte Carlo sampling. | Analytical | Sim | Three-finger | Singulated | voxelized occupancy grid | "6-DoF grasp pose (x, y, z, r, p, y)" | Custom | GraspIt! | | | Oblique view, Side view | Python, PyTorch | https://github.com/aalto-intelligent-robotics/robust_grasp_planning_over_uncertain_shape_completions | MIT | Jens Lundell, Intelligent Robotics Lab, Aalto University | Lundell, Jens, Francesco Verdoja, and Ville Kyrki. "Robust grasp planning over uncertain shape completions." In 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1526-1532. IEEE, 2019. | 2019 | |||||||||||||||||
49 | ROI-GD | ROI-GD is proposed to provide a feasible solution to this problem based on Region of Interest (ROI), which is the region proposal for objects. ROI-GD uses features from ROIs to detect grasps instead of the whole scene. It has two stages: the first stage is to provide ROIs in the input image and the second-stage is the grasp detector based on ROI features. | ROI-based | Real | Two-finger | Piled | RGBD image | "2D grasp rectangle (x, y, width, height, angle)" | Visual Manipulation Relationship Dataset (VMRD) | N/A | ResNet-101 | Prediction success rate, Grasp success rate | Oblique view | Python, PyTorch | https://github.com/ZhangHanbo/Visual-Manipulation-Relationship-Network-Pytorch | | H. Zhang, X. Lan, S. Bai, X. Zhou, Z. Tian, and N. Zheng, “ROI-based Robotic Grasp Detection for Object Overlapping Scenes,” Mar. 14, 2019, arXiv: arXiv:1808.10313. | 2019 | ||||||||||||||||||
50 | Single-Shot SE(3) Grasp Detection (S4G) | S4G is a grasping proposal algorithm to regress SE(3) pose from single camera point cloud (depth only, no RGB information). S4G is trained only on synthetic dataset with YCB Objects, it can generate to real world grasping with unseen objects that has never been used in the training. | Sampling | Sim | Two-finger | Piled | Point cloud | "6-DoF grasp pose (x, y, z, r, p, y)" | Custom | MuJoCo | PointNet | | Oblique view | Python, PyTorch | https://github.com/yzqin/s4g-release https://sites.google.com/view/s4ggrapsing | | Yuzhe Qin | Qin, Yuzhe, Rui Chen, Hao Zhu, Meng Song, Jing Xu, and Hao Su. "S4g: Amodal single-view single-shot se (3) grasp detection in cluttered scenes." In Conference on robot learning, pp. 53-65. PMLR, 2020. | 2020 | |||||||||||||||||
51 | UniGrasp | We propose UniGrasp, an efficient data-driven grasp synthesis method that considers both the object geometry and gripper attributes as inputs. UniGrasp is based on a novel deep neural network architecture that selects sets of contact points from the input point cloud of the object. The proposed model is trained on a large dataset to produce contact points that are in force closure and reachable by the robot hand. By using contact points as output, we can transfer between a diverse set of multifingered robotic hands. | Direct regression | Sim | Multi-finger | Singulated | Point cloud | Grasp Points | UniGrasp | Bullet | PointNet++ | | Eye-in-hand | Python, TensorFlow | https://github.com/stanford-iprl-lab/UniGrasp https://sites.google.com/view/unigrasp | MIT | Stanford Interactive Perception and Robot Learning Lab | Shao, Lin, Fabio Ferreira, Mikael Jorda, Varun Nambiar, Jianlan Luo, Eugen Solowjow, Juan Aparicio Ojea, Oussama Khatib, and Jeannette Bohg. "Unigrasp: Learning a unified model to grasp with multifingered robotic hands." IEEE Robotics and Automation Letters 5, no. 2 (2020): 2286-2293. | 2020 | |||||||||||||||||
52 | Volumetric Grasping Network (VGN) | VGN is a 3D convolutional neural network for real-time 6 DOF grasp pose detection. The network accepts a Truncated Signed Distance Function (TSDF) representation of the scene and outputs a volume of the same spatial resolution, where each cell contains the predicted quality, orientation, and width of a grasp executed at the center of the voxel. The network is trained on a synthetic grasping dataset generated with physics simulation. | Direct regression | Sim | Two-finger | Piled | Truncated Signed Distance Function (TSDF) | "6-DoF grasp pose (x, y, z, r, p, y)" | VGN | PyBullet | | | Eye-in-hand | Python, PyTorch | https://github.com/ethz-asl/vgn | BSD 3-Clause | Autonomous Systems Lab @ ETH Zürich | Breyer, Michel, Jen Jen Chung, Lionel Ott, Roland Siegwart, and Juan Nieto. "Volumetric grasping network: Real-time 6 dof grasp detection in clutter." In Conference on Robot Learning, pp. 1602-1611. PMLR, 2021. | 2021 |