Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 24 tok/s Pro
GPT-5 High 23 tok/s Pro
GPT-4o 77 tok/s Pro
Kimi K2 159 tok/s Pro
GPT OSS 120B 431 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Learning to Grasp Anything by Playing with Random Toys (2510.12866v1)

Published 14 Oct 2025 in cs.RO and cs.CV

Abstract: Robotic manipulation policies often struggle to generalize to novel objects, limiting their real-world utility. In contrast, cognitive science suggests that children develop generalizable dexterous manipulation skills by mastering a small set of simple toys and then applying that knowledge to more complex items. Inspired by this, we study if similar generalization capabilities can also be achieved by robots. Our results indicate robots can learn generalizable grasping using randomly assembled objects that are composed from just four shape primitives: spheres, cuboids, cylinders, and rings. We show that training on these "toys" enables robust generalization to real-world objects, yielding strong zero-shot performance. Crucially, we find the key to this generalization is an object-centric visual representation induced by our proposed detection pooling mechanism. Evaluated in both simulation and on physical robots, our model achieves a 67% real-world grasping success rate on the YCB dataset, outperforming state-of-the-art approaches that rely on substantially more in-domain data. We further study how zero-shot generalization performance scales by varying the number and diversity of training toys and the demonstrations per toy. We believe this work offers a promising path to scalable and generalizable learning in robotic manipulation. Demonstration videos, code, checkpoints and our dataset are available on our project page: https://lego-grasp.github.io/ .

Summary

  • The paper introduces a toy-play training strategy that uses a limited set of geometric primitives to achieve zero-shot grasp generalization.
  • It employs the LEGO architecture with an object-centric DetPool mechanism, leading to a 67% success rate across 64 YCB objects.
  • The approach proves data-efficient and scalable, outperforming larger models while transferring across different robotic embodiments.

Generalizable Robotic Grasping via Compositional Toy Play

Introduction and Motivation

The paper "Learning to Grasp Anything by Playing with Random Toys" (2510.12866) addresses the persistent challenge of generalization in robotic manipulation. While recent advances in robotic policies have enabled impressive performance on in-domain tasks, these systems typically exhibit poor transfer to novel, out-of-distribution (OOD) objects. Drawing inspiration from developmental psychology, which suggests that infants acquire generalizable manipulation skills by interacting with a small set of simple toys, the authors propose a compositional approach: training robots exclusively on objects constructed from a limited set of geometric primitives, then evaluating zero-shot transfer to real-world objects. Figure 1

Figure 1: The grasping policy is trained solely on random compositions of four basic primitives and achieves 67% zero-shot success on 64 YCB objects.

CÉZANNE Toy Dataset: Compositional Object Generation

The core of the methodology is the construction of a synthetic dataset of "Cézanne toys," each generated by randomly assembling 1–5 instances of four shape primitives: spheres, cuboids, cylinders, and rings. The primitives' dimensions and colors are randomized, and their spatial arrangement ensures physical connectivity and compositional diversity. Figure 2

Figure 2: Example Cézanne toys, each a random composition of 1–5 primitives with randomized dimensions and colors.

This compositional dataset is both OOD with respect to real-world objects and structurally rich, providing a principled testbed for evaluating generalization. The dataset is instantiated both in simulation (ManiSkill) and physically via 3D printing, enabling data collection across both domains.

LEGO: Object-Centric Policy Architecture

The proposed policy architecture, LEGO (LEarning to Grasp from tOys), is designed to maximize generalization by enforcing object-centricity in the visual representation. The architecture consists of:

  • Vision Encoder: A pretrained ViT-L (MVP) encodes multi-view RGB observations.
  • Detection Pooling (DetPool): A novel pooling mechanism that uses segmentation masks (from SAM 2 or ground truth) to restrict attention within the ViT to object patches, followed by mean pooling over object tokens. This ensures the extracted visual features are object-centric and invariant to background distractors.
  • Policy Transformer: A ViT-Base transformer backbone receives concatenated visual and proprioceptive embeddings over a temporal window and autoregressively predicts future action sequences. Figure 3

    Figure 3: LEGO architecture with DetPool. (a) ViT with DetPool extracts object-centric features; (b) DetPool restricts attention to object patches and pools only object tokens.

The policy is trained via behavior cloning with an ℓ1\ell_1 loss on action sequences, using only demonstration data collected on Cézanne toys.

Experimental Evaluation

Zero-Shot Generalization

The primary evaluation is zero-shot grasping on the YCB object set, both in simulation and on real hardware (Franka Emika Panda with Robotiq gripper, and Unitree H1-2 humanoid with Inspire dexterous hands). The policy is trained exclusively on Cézanne toys and never exposed to real objects during training.

  • Franka/YCB (real): 67% mean success rate on 64 YCB objects, using only 1,500 demonstrations and a model with 86M parameters.
  • Comparison: Outperforms OpenVLA-OFT (7B parameters, large-scale pretraining) and Ï€0\pi_0-FAST (3B parameters, in-domain pretraining) when both are fine-tuned on the same toy data. Only a fully fine-tuned Ï€0\pi_0-FAST with additional in-domain data marginally surpasses LEGO.
  • H1-2 Humanoid: 50.8% mean success rate on 13 everyday objects, demonstrating embodiment-agnostic generalization. Figure 4

    Figure 4: Real-world evaluation settings: DROID Franka with YCB objects (left), H1-2 humanoid with dexterous hands and 13 objects (right).

Scaling Laws and Ablations

The authors conduct extensive scaling studies:

  • Data Scaling: Zero-shot performance increases with both the number of unique toys and the number of demonstrations, but the latter is more critical. With sufficient demonstrations, as few as 25 unique toys suffice for robust transfer.
  • Model Size: Performance saturates at 86M parameters (ViT-Base), with no significant gains from larger backbones.
  • Primitive Importance: Spheres and cuboids are most critical for generalization; omitting rings or cylinders has less impact.
  • Toy Complexity: Toys with two primitives contribute most to generalization, likely due to the distribution of part counts in real objects.
  • Color Diversity: Training on multi-color toys yields a 1–4% improvement in generalization over single-color toys. Figure 5

    Figure 5: Left: Zero-shot success scales with both number of demos and unique toys; 25 toys suffice with enough demos. Right: Policy transformer performance saturates at 86M parameters.

Hardware and Data Collection

The approach is validated on two distinct robotic platforms, with detailed hardware configurations and teleoperation-based demonstration collection. The policy is robust to embodiment changes, requiring only proprioceptive adaptation. Figure 6

Figure 6: Franka Emika Panda with Robotiq Gripper.

Figure 7

Figure 7: H1-2 Humanoid with Inspire Dexterous Hands.

Analysis and Implications

Object-Centricity as the Key to Generalization

Ablations demonstrate that DetPool is essential: replacing it with standard mean, attention, or CLS pooling degrades zero-shot performance by 22–48%. The object-centric representation induced by DetPool enables the policy to ignore background and context, focusing exclusively on the target object. This is consistent with findings in cognitive science and object-centric vision literature, and contrasts with prior approaches that rely on large-scale pretraining or heavy data augmentation.

Data Efficiency and Model Simplicity

The results challenge the prevailing assumption that large-scale, in-domain data and massive models are prerequisites for generalization in robotic manipulation. The LEGO policy, with only 86M parameters and a few hours of demonstration data, matches or exceeds the performance of models orders of magnitude larger and more data-hungry.

Practical and Theoretical Implications

  • Practical: The approach enables scalable, data-efficient training of generalist grasping policies, reducing the need for costly real-world data collection and annotation. The compositional toy dataset is easy to generate and physically instantiate.
  • Theoretical: The findings support the hypothesis that compositionality and object-centricity are sufficient for generalization in manipulation, echoing theories of human cognitive development and compositional representation learning.
  • Limitations: The method's generalization may degrade for objects with physical properties not represented in the toy set (e.g., deformable, articulated, or highly textured objects). The current focus is on single-step grasping; extension to long-horizon, multi-step tasks remains open.

Future Directions

Potential avenues for future research include:

  • Extending the compositional approach to more complex manipulation tasks (e.g., tool use, assembly, deformable object handling).
  • Incorporating additional primitive types or physical properties (e.g., mass, compliance) to further enhance generalization.
  • Exploring more efficient architectures for deployment on resource-constrained hardware.
  • Investigating the integration of language or multimodal conditioning for task specification.

Conclusion

This work demonstrates that robust, general-purpose robotic grasping can be achieved by training on a small set of compositional, primitive-based toys, provided that the policy architecture enforces object-centricity via detection pooling. The approach achieves strong zero-shot transfer to real-world objects and diverse robotic embodiments, outperforming state-of-the-art models that rely on large-scale pretraining. These results suggest a scalable and theoretically grounded path toward generalizable robotic manipulation, with significant implications for both research and practical deployment.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Github Logo Streamline Icon: https://streamlinehq.com
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 14 tweets and received 64 likes.

Upgrade to Pro to view all of the tweets about this paper: