Stacking Task Training Data

Stacking datasets for robotic manipulation research and industrial automation — block towers, plate stacking, pallet layering, and multi-object stability reasoning with contact physics annotations and balance prediction labels.

Data Requirements

Modality

RGB-D + proprioception + object poses + optional force/torque

Volume Range

100-500 demonstrations per configuration

Temporal Resolution

30 Hz video, per-placement object pose snapshots, stability annotations

Key Annotations
Per-placement 6-DoF object pose relative to support surfacePost-placement stability label and settling timeTower height and lean angle measurementsObject identity, dimensions, and weightPlacement force profile (if force-instrumented)Success/failure with collapse-frame annotation
Compatible Models
PerActACT / ALOHADiffusion PolicyRT-2SAC (Soft Actor-Critic)SayCan
Environment Types
Research tabletop workspaceBlock stacking benchmark stationIndustrial palletizing cellWarehouse pallet loading areaKitchen plate stackingConstruction material staging

How Claru Supports This Task

Claru provides stacking data collection from research block-tower benchmarks to industrial palletizing at full production scale. Our workstations feature calibrated multi-view cameras with sub-millimeter object tracking via fiducial markers or structured-light scanning, optional 6-axis force/torque sensing for placement force capture, and side-mounted cameras for tower profile monitoring. Operators follow stacking protocols emphasizing placement precision and stability, with quality scoring on placement accuracy, settling time, and lean angle. We collect across configurable difficulty levels from 2-object baselines through 8-object towers and mixed-geometry configurations, including intentional recovery behaviors. Deliverables include per-placement stability annotations, 6-DoF object poses, force profiles, and tower profile measurements formatted for PerAct, ACT, Diffusion Policy, or RL pipelines. For industrial palletizing, we provide layer-by-layer packing patterns on standard pallets with real shipping boxes. Daily throughput of 300-800 stacking demonstrations per station enables rapid scaling.

What Is Robotic Stacking and Why Does Data Matter?

Stacking — placing objects on top of each other to form stable structures — is a foundational manipulation skill that tests a robot's understanding of gravity, friction, contact geometry, and structural stability. From building a 6-block tower to layering mixed-size boxes on a pallet, stacking requires the robot to predict whether a placement will result in a stable configuration or a collapse. This is fundamentally a physics reasoning problem: the center of mass must remain within the support polygon, contact surfaces must provide sufficient friction, and the placement must be accurate enough that cumulative errors do not cause the tower to topple as it grows taller.

Stacking benchmarks have become a standard evaluation in robot learning research precisely because they isolate key manipulation competencies. The Block Stacking benchmark in RLBench (James et al., 2020) requires a robot to stack colored blocks in a specified order, testing both grasp planning and placement precision. Performance on this task correlates strongly with general manipulation capability — policies that achieve 90%+ stacking success typically also excel on other contact-rich tasks. The reason is that stacking demands accurate pose estimation, gentle placement with controlled descent speed, and real-time stability assessment — skills that transfer broadly.

Industrial stacking applications extend far beyond toy blocks. Palletizing — arranging boxes, cases, and bags onto pallets for shipping — is a $3.2 billion market segment within warehouse automation. Mixed-case palletizing, where boxes of different sizes must be arranged into stable, space-efficient layers, remains largely manual because heuristic algorithms struggle with the combinatorial complexity of heterogeneous item arrangements. The Robotics Industry Association reports that palletizing robots handle only 15-20% of warehouse palletizing operations, with the remainder performed by human workers at an average cost of $18-25 per hour. Learning-based stacking policies that reason about stability across diverse object geometries could unlock the remaining 80% of this market.

The core data challenge is that stability is a binary outcome that depends on continuous physical parameters. A 1-mm shift in placement position can mean the difference between a stable 8-block tower and a collapse at block 5. Force profiles during placement — the contact force, settling time, and post-placement vibration — contain critical information about whether the structure is marginally stable or robustly stable. Demonstrations must capture not just the final placement pose but the approach trajectory, descent speed, release timing, and post-placement dynamics to teach policies the full physics of stable stacking.

Stacking Data by the Numbers

$3.2B
Palletizing robot market size (2024)
90%+
Stacking success rate for strong manipulation policies
1 mm
Placement accuracy threshold for tall towers
80%
Palletizing still done manually in warehouses
100-500
Demos per object set for block stacking BC
6-axis
Force/torque DOF for placement stability

Data Requirements by Stacking Approach

Stacking policies range from simple behavioral cloning to physics-informed methods that reason explicitly about stability. Data needs vary accordingly.

ApproachData VolumeKey ModalitiesStability SignalStrengths
Behavioral Cloning100-500 demos per tower heightRGB-D + proprioceptionImplicit in demo successSimple; works for fixed object sets
Diffusion Policy50-200 demos per configurationMulti-view RGB + proprioceptionMultimodal placement distributionHandles placement ambiguity; multiple stable positions
Physics-informed RL500K+ sim episodes + 100 real demosSim physics state + real force/torquePhysics simulator + stability rewardGeneralizes to novel objects; explicit stability reasoning
Sim-to-Real with IsaacGym1M+ sim episodes + 200 real calibrationSim contact + real RGB-D for domain gapGround-truth physics in simulationScalable; diverse object geometry coverage
Foundation model stacking (RT-2, Octo)5K-20K demos + language instructionsRGB + language + proprioceptionLearned from large-scale pretrainingLanguage-conditioned; zero-shot to new objects

State of the Art in Learned Stacking

RLBench (James et al., 2020) established block stacking as a standard benchmark, with the task requiring a robot arm to stack colored cubes in a specified order. Early behavioral cloning methods achieved 30-50% success on 3-block towers, but modern architectures have dramatically improved. PerAct (Shridhar et al., 2023) uses a 3D voxel-based perceiver to predict next-best actions and achieves 82% success on the RLBench stack_blocks task with 100 demonstrations. The key insight is that 3D spatial reasoning, rather than 2D image features, enables accurate placement prediction for tasks where vertical precision matters.

ACT (Action Chunking with Transformers, Zhao et al., 2023) demonstrated that predicting action sequences rather than single actions dramatically improves stacking performance. On a real-robot block stacking task with the ALOHA bimanual platform, ACT achieved 96% success on 2-block stacking and 78% on 4-block towers using only 50 teleoperated demonstrations. The action chunking approach smooths the transition between approach and placement phases, avoiding the jerky motions that cause instability during block release — a critical failure mode in naive frame-by-frame policies.

For industrial palletizing, DeepMind's work on robotic stacking (Lee et al., 2021) demonstrated multi-object stacking policies learned entirely from simulation. Training in MuJoCo with domain randomization over object sizes, masses, and friction coefficients, their SAC-based policy achieved 85% success on stacking 5 diverse objects into stable configurations in the real world. The 15% failure rate was dominated by thin, flat objects where the contact surface was too small for stable support — a case where real demonstration data showing human strategies for handling thin objects would directly address the failure mode.

The most recent advance combines vision-language models with physical reasoning for instruction-conditioned stacking. SayCan (Ahn et al., 2022) and RT-2 (Brohan et al., 2023) demonstrate that foundation models can decompose high-level stacking instructions ('build a pyramid with red blocks on the bottom') into primitive stacking actions. RT-2 achieves 73% success on novel stacking configurations described in natural language, compared to 23% for the non-VLM baseline. However, these models still struggle with tall structures (5+ objects) where cumulative placement error exceeds stability margins, highlighting the need for precise placement demonstrations.

Collection Methodology for Stacking Data

Stacking data collection requires precise tracking of both the robot end-effector and the objects being stacked. The minimum sensor setup includes a calibrated overhead RGB-D camera for tracking object poses on the workspace, a wrist-mounted camera for close-range placement guidance, and optionally a 6-axis force/torque sensor at the wrist to capture placement forces and detect instability. Object pose tracking accuracy should be 1 mm or better — achievable with AprilTag fiducial markers on the objects or high-resolution structured-light sensors. For tall tower experiments, a side-mounted camera is essential to observe the tower profile and detect lean or tilt before collapse.

Demonstrations should cover the full spectrum of stacking complexity: 2-object stacks (baseline placement precision), 3-5 object towers (cumulative error management), mixed-size stacking (support polygon reasoning), and unstable object stacking (non-convex bases, round objects). For each configuration, collect both successful demonstrations and near-failure demonstrations where the tower is marginally stable — these edge cases teach the policy the boundary between stable and unstable placements. Record 50-200 demonstrations per configuration, with 20% of demonstrations intentionally including recovery behaviors (nudging a misaligned object, re-grasping after a near-drop).

Annotations for stacking data must capture per-placement stability information: pre-placement tower height and center-of-mass estimate, placement pose (6-DoF relative to the support surface), post-placement settling time (time from release to zero velocity), tower lean angle after placement, and binary stability label (does the tower survive a 5-second observation period without collapse). For force-instrumented setups, annotate the impact force at contact, the damping profile during descent, and the steady-state load distribution. These stability annotations enable reward shaping for RL fine-tuning and quality filtering for behavioral cloning.

For industrial palletizing data, the collection setup scales to full-size boxes (up to 600 mm x 400 mm x 400 mm) on standard 1200 mm x 1000 mm pallets. Overhead structured-light scanning after each box placement captures the evolving pallet surface profile. The key challenge is layer transitions — starting a new layer on top of a completed layer requires precise placement to distribute weight evenly and maintain stack stability. Collect demonstrations that cover single-SKU palletizing (uniform boxes), mixed-SKU palletizing (2-5 box sizes), and interlocking patterns (alternating box orientations for cross-layer stability).

Key Datasets and Benchmarks for Robotic Stacking

Stacking appears in many manipulation benchmarks, but few datasets focus specifically on stacking with stability annotations.

Dataset / BenchmarkYearScaleObject TypesKey FeaturesLimitations
RLBench stack_blocks (James et al.)2020100-200 demos + unlimited sim generationColored cubes (fixed size)Standardized benchmark; multiple color ordersOnly uniform cubes; no stability metrics
ALOHA stacking (Zhao et al.)202350 demonstrations per task variantCubes, cylinders, cupsBimanual; real robot; action chunkingSmall scale; limited object diversity
ManiSkill2 stacking2023100K+ sim episodesYCB objects + procedural shapesDiverse objects; sim physics; large scaleSim-only; no real stability data
DeepMind stacking (Lee et al.)2021Sim training + real evaluation5 diverse household objectsMulti-object; real-world transferNot publicly released; limited object count
LIBERO stacking subtasks202350 demos per subtaskTabletop objects in LIBERO suiteLifelong learning benchmark; multi-taskStacking is one of many tasks; limited focus

How Claru Supports Stacking Data Needs

Claru provides stacking data collection covering the full range from research block-stacking benchmarks to industrial palletizing demonstrations. Our workstations are equipped with calibrated multi-view cameras (overhead + side + wrist-mounted at 30 Hz), optional 6-axis force/torque sensing at the wrist for placement force capture, and precision object tracking via fiducial markers or structured-light scanning with sub-millimeter accuracy. For palletizing data, we operate full-scale stations with standard pallets, real shipping boxes, and overhead scanning for pallet surface profiling.

Our operators are trained on stacking protocols that emphasize placement precision and stability awareness. Each demonstration is quality-scored on placement accuracy (deviation from target pose), tower stability (post-placement settling time and lean angle), and completion rate (successful tower height versus target). We collect demonstrations across configurable difficulty levels: baseline 2-3 object stacks, challenging 5-8 object towers, mixed-geometry stacking, and recovery behaviors from near-failure states.

Claru delivers stacking datasets with per-placement stability annotations, 6-DoF object poses, tower profile measurements, and optional force data — formatted for direct ingestion by PerAct, ACT/ALOHA, Diffusion Policy, or RL fine-tuning pipelines. For palletizing clients, we provide layer-by-layer packing patterns, box placement sequences, and pallet stability metrics. Our collection throughput of 300-800 stacking demonstrations per day per station enables rapid dataset scaling for both research and production applications.

References

  1. [1]James et al.. RLBench: The Robot Learning Benchmark & Learning Environment.” RA-L 2020, 2020. Link
  2. [2]Zhao et al.. Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware.” RSS 2023, 2023. Link
  3. [3]Shridhar et al.. Perceiver-Actor: A Multi-Task Transformer for Robotic Manipulation.” CoRL 2023, 2023. Link
  4. [4]Lee et al.. Beyond Pick-and-Place: Tackling Robotic Stacking of Diverse Shapes.” CoRL 2021, 2021. Link
  5. [5]Brohan et al.. RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control.” CoRL 2023, 2023. Link

Frequently Asked Questions

For uniform cube stacking, 50-100 demonstrations per target height (2-block, 3-block, etc.) suffice with ACT or Diffusion Policy. For mixed-object stacking with diverse geometries, 200-500 demonstrations per object set are recommended. PerAct achieves 82% success on RLBench stack_blocks with 100 demos. Start with 50 demonstrations of your target configuration to validate the pipeline, then scale to 200+ for production reliability above 90%.

Force data significantly improves placement quality but is not strictly required for simple stacking. Vision-only policies achieve 70-85% success on 3-block towers. Force/torque data becomes critical for tall towers (5+ objects) where gentle placement with controlled descent is essential to avoid toppling, and for industrial palletizing where box weight varies and contact forces indicate whether the placement is properly seated. If you have force sensors, collect the data — it improves success rates by 10-15 percentage points on challenging configurations.

Transfer depends on the policy architecture. Position-regression methods (standard BC) show poor transfer — a policy trained on cubes achieves only 30-40% success on cylinders. 3D-aware architectures like PerAct transfer better (60-70% success on novel shapes) because they reason about support surfaces in voxel space. The best transfer comes from physics-informed methods trained on diverse object sets in simulation with 10-20% real demonstrations for domain adaptation. Include at least 5 distinct object geometries in your training data for reasonable generalization.

Tower collapses during demonstration are valuable negative examples — include them in the dataset with a failure label and collapse-frame annotation. Aim for 70-85% success rate in raw data collection. If success exceeds 95%, the operator is being too conservative (short towers, centered placements only) and the dataset lacks the challenging edge cases needed for robust training. After a collapse, have the operator rebuild from scratch rather than resuming — this captures the full stacking sequence from the beginning.

Each placement adds positional error. For a 6-block tower with 50 mm cubes, a 2 mm per-placement error accumulates to 12 mm total offset at the top — enough to shift the center of mass outside the support polygon and cause collapse. Target 1 mm or better placement accuracy for towers above 4 objects. Track accuracy using fiducial markers or structured-light scanning after each placement. For palletizing, tolerance is more relaxed (5-10 mm) because box sizes are larger and layer interlocking provides lateral stability.

Get a Custom Quote for Stacking Task Data

Tell us about your stacking requirements — object types, target heights, and stability criteria — and we will design a data collection plan for your specific application.