Papers
Topics
Authors
Recent
2000 character limit reached

DemoFunGrasp: Universal Dexterous Grasping

Updated 23 December 2025
  • DemoFunGrasp is a comprehensive research framework for universal dexterous functional grasping that integrates demonstration-editing with reinforcement learning and multimodal perception.
  • It factorizes grasp synthesis into explicit affordance and style conditioning, enabling physically plausible, robust, and human-inspired grasping across diverse objects.
  • The framework employs state-based and vision-based policies with PPO and vision–language integration, achieving strong sim-to-real transfer and high grasp success rates.

DemoFunGrasp is a research framework for universal dexterous functional grasping that combines single-demonstration editing with reinforcement learning (RL), factorized semantic grasping objectives, and vision–language perception to advance the generalization and usability of dexterous robot hands. DemoFunGrasp is designed to generate physically plausible, functional, and robust grasps across diverse objects, affordance locations, and human-inspired grasping styles, achieving strong sim-to-real transfer and instruction-following capability. This approach synthesizes methodologies from demonstration-based trajectory editing, reward shaping in RL, and multimodal perceptual integration, offering a flexible paradigm for anthropomorphic hand manipulation in both simulation and real-world settings (Mao et al., 15 Dec 2025).

1. Problem Factorization: Affordance and Style Conditioning

Functional grasping is formally defined as achieving both mechanical stability and semantically appropriate hand postures for tool or object use. DemoFunGrasp factorizes this objective into two explicit and complementary components:

  • Affordance (paffordp_{\mathrm{afford}}): Specifies “where to grasp” as a 3D point on the object surface representing a functionally meaningful contact region.
  • Style (lstylel_{\mathrm{style}}): Specifies “how to grasp” via a one-hot label corresponding to predefined human-inspired grasping styles (e.g., palmar pinch, lateral pinch), each with canonical joint configurations qpos\mathbf{q}_{\mathrm{pos}}.

This decomposition reduces the complexity of specifying and generalizing functional grasps. The policy is conditioned on both the object geometry xox_o and the desired tuple (pafford,lstyle)(p_{\mathrm{afford}}, l_{\mathrm{style}}), yielding a grasping action aa that achieves the designated tool-use semantics (Mao et al., 15 Dec 2025).

2. Demonstration-Editing RL Formalism

DemoFunGrasp reformulates grasp synthesis as a one-step MDP restricted to local editing around a single demonstration. The state s=(sr,so,xo,pafford,lstyle)s=(\mathbf{s}_r,\mathbf{s}_o,x_o,p_{\mathrm{afford}},l_{\mathrm{style}}) encodes the robot and object poses, object point cloud, desired affordance, and style. The action a=(ΔT,Δq,k)a=(\Delta T, \Delta \mathbf{q}, k) consists of:

  • ΔTSE(3)\Delta T \in SE(3): A residual wrist-frame transformation applied to the demonstration trajectory.
  • ΔqRn\Delta \mathbf{q} \in \mathbb{R}^n: Residual joint-angle offsets.
  • kRk \in \mathbb{R}: A style pose scaling parameter.

The agent maximizes the single-step RL objective: J(π)=Esρ0Eaπ(s)[r(s,a)],J(\pi) = \mathbb{E}_{s \sim \rho_0} \mathbb{E}_{a \sim \pi(\cdot \mid s)} [r(s, a)], optimized via Proximal Policy Optimization (PPO) (Mao et al., 15 Dec 2025).

Reward structure:

r(s,a)=λaffordrafford+λcloserclose+λqposrqpos+rsuccessr(s, a) = \lambda_{\mathrm{afford}} r_{\mathrm{afford}} + \lambda_{\mathrm{close}} r_{\mathrm{close}} + \lambda_{\mathrm{qpos}} r_{\mathrm{qpos}} + r_{\mathrm{success}}

with:

  • raffordr_{\mathrm{afford}}: Affordance proximity at lift.
  • rcloser_{\mathrm{close}}: Early engagement with the target.
  • rqposr_{\mathrm{qpos}}: Style configuration match.
  • rsuccessr_{\mathrm{success}}: Binary object-lifted bonus.

This framework converts the grasp synthesis problem into a robust, local residual correction task, achieving greater sample efficiency and ease of transfer than full trajectory optimization (Mao et al., 15 Dec 2025).

3. Demonstration Editing and Trajectory Generation

Given a single demonstration D={(ptref,qtref)}t=0TDD = \{(\mathbf{p}_t^{\mathrm{ref}}, \mathbf{q}_t^{\mathrm{ref}})\}_{t=0}^{T^D}, DemoFunGrasp applies editing operations:

  • End-effector editing:

ptΔTptref\mathbf{p}_t \leftarrow \Delta T \cdot \mathbf{p}_t^{\mathrm{ref}}

  • Style-aware hand editing:

qpos=kqpos+Δq,qt=q0ref+f(qtrefq0ref),\mathbf{q}_{\mathrm{pos}}^* = k\,\mathbf{q}_{\mathrm{pos}} + \Delta\mathbf{q},\quad \mathbf{q}_t = \mathbf{q}_0^{\mathrm{ref}} + f(\mathbf{q}_t^{\mathrm{ref}} - \mathbf{q}_0^{\mathrm{ref}}),

where

f=qposq0ref2qTDrefq0ref2f = \frac{\|\mathbf{q}_{\mathrm{pos}}^* - \mathbf{q}_0^{\mathrm{ref}}\|_2}{\|\mathbf{q}_{T^D}^{\mathrm{ref}} - \mathbf{q}_0^{\mathrm{ref}}\|_2}

This editing approach enables the model to synthesize grasp trajectories for novel affordance and style pairs from a single reference, bypassing the need for extensive data or full RL exploration (Mao et al., 15 Dec 2025).

4. Learning Pipeline: Architectures, Perception, and Training

DemoFunGrasp uses both state-based and vision-based policy networks:

  • State-based policy: Simple MLPs or graph networks consume the full state and are trained with PPO in IsaacGym on a set of 175 DexGraspNet and YCB objects, sampling affordances and styles uniformly.
  • Vision-based policy: Trained via imitation learning from 30,000 successful state-based rollouts. The best-performing vision model (DiT + VLM encoder) processes both RGB images (256×256256 \times 256) and a 2D affordance projection caffordc_{\mathrm{afford}}, predicting end-effector and joint commands at each timestep.

Comprehensive domain randomization during simulation—including object/texturing, lighting, camera pose, and initial conditions—is employed to enable zero-shot sim-to-real transfer (Mao et al., 15 Dec 2025).

Integration of a pre-trained vision–LLM (VLM, specifically Embodied-R1) enables autonomous mapping of natural language instructions to 2D affordance points, which are then projected into 3D and fed to the low-level policy. This modular stack facilitates instruction-following grasp execution by decoupling semantic perception from control (Mao et al., 15 Dec 2025).

5. Experimental Results and Comparative Evaluation

Quantitative and qualitative benchmarks demonstrate that DemoFunGrasp achieves high functional accuracy, diversity, and sim-to-real transferability.

Affordance Accuracy (Mean Success Affordance Distance, cm):

Model Train Seen Cat. Unseen Cat.
DemoGrasp 6.29 6.27 6.20
DemoFunGrasp 3.03 3.02 3.21

Style Diversity and Grasp Success Rate (GSR%) in Simulation:

Method GSR% (↑) Style Diversity (↑)
UniDexGrasp 74.3 1.00
DemoFunGrasp 76.26 1.48

Real-world Vision-Based Policy:

  • Human-chosen affordance/style: GSR71%GSR \approx 71\%
  • VLM-predicted affordance: GSR64%GSR \approx 64\%, Intended Affordance Score (IAS) 0.87–0.40, Intended Style Score (ISS) 0.87–1.00

Removing individual reward or editing components in ablation studies degrades performance substantially (e.g., style disturbance ablation reduces GSR from 77.04% to 58.67%) (Mao et al., 15 Dec 2025).

DemoFunGrasp’s factorization and demonstration-editing strategy generalizes several earlier lines of research:

  • Continuous grasping functions from demonstration: Prior work on Conditional VAE-based continuous trajectory synthesis for dexterous hands (Ye et al., 2022) focuses on efficiency and generalization for smooth trajectory generation from human demonstrations but lacks explicit affordance/style conditioning and single-edit RL framing.
  • Contact-based transfer and optimization: Contact field, anchor-point, and category-level correspondence methods leverage physical and geometric priors for robust tool-use grasps (Wei et al., 2023), offering strong functionality and transfer across hands; however, these approaches do not employ residual editing or language/vision conditioning.
  • Parametric mixture models for generative grasping: Approaches that model grasp synthesis with GMMs for contact points and poses (Arruda et al., 2019) achieve computational efficiency and modularity, but are limited in functional goal representation and closed-loop execution.

A plausible implication is that DemoFunGrasp unifies trajectory-level editing, semantic factorization, and multimodal policy architectures in a manner that subsumes the strengths of these earlier approaches, while also providing a direct route to autonomous, language-guided grasping (Mao et al., 15 Dec 2025).

7. Limitations and Future Directions

Current performance achieves robust functional grasping and instruction-following at centimeter-scale positional accuracy. Limitations include:

  • No sub-centimeter manipulation precision (e.g., button pressing or threading).
  • Lack of in-hand re-grasping or multi-step correction policies.
  • Possible degradation for objects with highly variable or thin affordance surfaces, particularly with reliance on a single demonstration.
  • No feedback from tactile or force sensors during execution.

Planned future enhancements include the incorporation of tactile sensing, force-closure feedback for precise functional grasping, multi-step editing policies for real-time adjustment, end-to-end joint training of VLM and control policy, and extension to multi-object, sequential, or bimanual tasks (Mao et al., 15 Dec 2025).


References

  • "Universal Dexterous Functional Grasping via Demonstration-Editing Reinforcement Learning" (Mao et al., 15 Dec 2025)
  • "Learning Continuous Grasping Function with a Dexterous Hand from Human Demonstrations" (Ye et al., 2022)
  • "Generalized Anthropomorphic Functional Grasping with Minimal Demonstrations" (Wei et al., 2023)
  • "Generative grasp synthesis from demonstration using parametric mixtures" (Arruda et al., 2019)

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to DemoFunGrasp.