TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/ClickDiff: Click to Induce Semantic Contact Map for Contro...

ClickDiff: Click to Induce Semantic Contact Map for Controllable Grasp Generation with Diffusion Models

Peiming Li, Ziyi Wang, Mengyuan Liu, Hong Liu, Chen Chen

2024-07-28Controllable Grasp GenerationGrasp Generation
PaperPDF

Abstract

Grasp generation aims to create complex hand-object interactions with a specified object. While traditional approaches for hand generation have primarily focused on visibility and diversity under scene constraints, they tend to overlook the fine-grained hand-object interactions such as contacts, resulting in inaccurate and undesired grasps. To address these challenges, we propose a controllable grasp generation task and introduce ClickDiff, a controllable conditional generation model that leverages a fine-grained Semantic Contact Map (SCM). Particularly when synthesizing interactive grasps, the method enables the precise control of grasp synthesis through either user-specified or algorithmically predicted Semantic Contact Map. Specifically, to optimally utilize contact supervision constraints and to accurately model the complex physical structure of hands, we propose a Dual Generation Framework. Within this framework, the Semantic Conditional Module generates reasonable contact maps based on fine-grained contact information, while the Contact Conditional Module utilizes contact maps alongside object point clouds to generate realistic grasps. We evaluate the evaluation criteria applicable to controllable grasp generation. Both unimanual and bimanual generation experiments on GRAB and ARCTIC datasets verify the validity of our proposed method, demonstrating the efficacy and robustness of ClickDiff, even with previously unseen objects. Our code is available at https://github.com/adventurer-w/ClickDiff.

Results

TaskDatasetMetricValueModel
HandGRABAverage MPJPE (mm)40.57ClickDiff
Pose EstimationGRABAverage MPJPE (mm)40.57ClickDiff
Hand Pose EstimationGRABAverage MPJPE (mm)40.57ClickDiff
3DGRABAverage MPJPE (mm)40.57ClickDiff
3D Hand Pose EstimationGRABAverage MPJPE (mm)40.57ClickDiff
Grasp GenerationGRABAverage MPJPE (mm)40.57ClickDiff
1 Image, 2*2 StitchiGRABAverage MPJPE (mm)40.57ClickDiff

Related Papers

GraspGen: A Diffusion-based Framework for 6-DOF Grasping with On-Generator Training2025-07-17A Unified Transformer-Based Framework with Pretraining For Whole Body Grasping Motion Generation2025-07-01Exploiting Radiance Fields for Grasp Generation on Novel Synthetic Views2025-05-16Grasping a Handful: Sequential Multi-Object Dexterous Grasp Generation2025-03-28EvolvingGrasp: Evolutionary Grasp Generation via Efficient Preference Alignment2025-03-18GAGrasp: Geometric Algebra Diffusion for Dexterous Grasping2025-03-06Bring Your Own Grasp Generator: Leveraging Robot Grasp Generation for Prosthetic Grasping2025-03-01Task-Oriented 6-DoF Grasp Pose Detection in Clutters2025-02-24