Table Setting Training Data

Table setting and arrangement datasets — precise placement of plates, utensils, glasses, and napkins with spatial reasoning annotations, cultural arrangement variations, and placement accuracy measurements.

Data Requirements

Modality

RGB overhead (4K) + side cameras + depth + wrist camera + proprioception

Volume Range

1K-5K complete setting demonstrations across styles

Temporal Resolution

30 Hz video, 50 Hz proprioception

Key Annotations
Sub-centimeter item placement coordinatesRelational predicates (left-of, above, aligned-with)Arrangement rule compliance per cultural stylePlacement quality score (centered, aligned, stable)Cultural setting style labelSequential ordering constraints
Compatible Models
Diffusion PolicySayCanTidyBotInner MonologueOpenVLAACT
Environment Types
Dining roomRestaurantBanquet hallTest tabletopKitchen counter

How Claru Supports This Task

Claru collects table setting data using instrumented tabletop workstations with overhead 4K cameras, precision-calibrated coordinate systems, and inventories of 20+ real tableware sets spanning Western formal, Western casual, Asian, and custom conventions. Each demonstration is annotated with sub-centimeter placement coordinates, relational spatial predicates between items, cultural compliance labels, and placement quality scores. Our operators demonstrate complete place settings following specified cultural rules with natural variation in approach trajectories, and every dataset includes train/validation/test splits stratified by tableware set and cultural style. We deliver in RLDS, HDF5, or Zarr formats compatible with Diffusion Policy, ACT, SayCan, and foundation model architectures, with standard throughput of 100-200 annotated settings per day per workstation.

What Is Robotic Table Setting and Why Does Data Matter?

Table setting is a canonical benchmark for long-horizon spatial reasoning in service robotics. Unlike simple pick-and-place tasks where objects have a single target location, table setting requires understanding relational constraints: the fork goes to the left of the plate, the knife to the right with the blade facing inward, the water glass above the knife. These spatial relationships are culturally determined, context-dependent (formal vs. casual), and must be maintained across varying table shapes, sizes, and existing place settings. A robot that can reliably set a table demonstrates the spatial reasoning capabilities needed for dozens of downstream service tasks.

The challenge compounds because table setting is inherently multi-step and multi-object. A typical formal place setting involves 8-12 items that must be placed in a specific sequence to avoid collisions and occlusion — you cannot place the bread plate after the main plate if the bread plate position overlaps the approach trajectory. SayCan (Ahn et al., 2022) demonstrated that combining LLM planning with learned manipulation primitives enables table-setting-like sequential tasks, but the manipulation policies themselves require demonstration data that captures the precise spatial relationships and gentle placement dynamics for fragile tableware.

Current public datasets for table setting are extremely limited. The BEHAVIOR-1K benchmark (Li et al., 2023) includes table setting as one of 1,000 household activities but provides only simulated demonstrations. The RT-2 evaluation suite tests table setting with a few dozen configurations. Real-world table setting data that captures the diversity of tableware (ceramic plates, glass stemware, metal utensils of varying sizes), table surfaces (wood, glass, cloth-covered), and cultural conventions (Western formal, Japanese kaiseki, Chinese banquet) simply does not exist at scale in any public dataset.

Building production-grade table setting policies requires 1,000-5,000 demonstrations across at least 20 distinct tableware sets and 5+ table configurations. Each demonstration must include precise placement coordinates (sub-centimeter accuracy), the spatial relationships between items (relative positions encoded as relational predicates), and arrangement rule compliance labels (correct/incorrect placement per cultural standard). Without this structured annotation, policies learn to place objects at approximate positions without understanding the relational rules that define a correctly set table.

Table Setting Data by the Numbers

8-12
Items in a formal Western place setting
<1 cm
Placement accuracy target for production
1K-5K
Demonstrations needed for robust policy
20+
Distinct tableware sets for generalization
5+
Cultural setting styles in a diverse dataset
90%+
Arrangement rule compliance target

Data Requirements by Table Setting Approach

Different approaches to learned table setting have distinct data requirements depending on how spatial relationships are represented.

ApproachData VolumePrimary ModalityKey AnnotationsBest For
LLM planner + manipulation primitives (SayCan)100-500 demos per primitive + language rulesRGB + language instructionsPrimitive success labels + language scene descriptionsFlexible table configurations with verbal instructions
End-to-end behavioral cloning (Diffusion Policy)1K-5K full setting demonstrationsMulti-view RGB + proprioceptionFull trajectories + placement coordinatesHigh-precision placement with consistent style
Spatial relationship prediction + planning5K-20K annotated table images + 500 demosOverhead RGB + depthRelational predicates (left-of, above, aligned-with)Novel tableware generalization via spatial rules
Foundation model fine-tuning (OpenVLA/RT-2)2K-10K language-annotated setting demosRGB + language instruction per stepStep-by-step language + cultural style labelMulti-style settings with language-specified conventions

State of the Art in Learned Table Setting

SayCan (Ahn et al., 2022) demonstrated that grounding LLM planning in learned manipulation affordances enables multi-step tasks like table setting. The system decomposes natural language instructions into sequences of manipulation primitives, with each primitive scored by a learned value function trained on real robot data. For table setting, SayCan successfully planned and executed 8-item place settings by chaining pick-and-place primitives, achieving 74% end-to-end success rate on the full setting sequence. The primary failure mode was accumulating placement error across sequential items rather than individual primitive failure.

TidyBot (Wu et al., 2023) addressed the personalization challenge in household arrangement tasks. By combining a small set of user-specific examples (5-10 demonstrations) with LLM-based category reasoning, TidyBot achieved 85% placement accuracy on novel items following the user's personal organization preferences. For table setting, this approach enables robots to learn household-specific conventions (which side the napkin goes on, whether wine glasses are used) from minimal examples rather than requiring exhaustive demonstrations of every configuration.

Inner Monologue (Huang et al., 2023) showed that providing LLMs with continuous environmental feedback during multi-step tasks dramatically improves success rates. For table setting, Inner Monologue uses object detection to verify each placement before proceeding to the next item, catching and correcting errors in real-time. This closed-loop approach achieved 91% success on 6-item casual settings versus 67% for open-loop execution, demonstrating the importance of verification feedback in long-horizon sequential tasks.

For the manipulation component, Diffusion Policy (Chi et al., 2023) has emerged as the preferred architecture for precise placement tasks. Diffusion Policy's ability to represent multi-modal action distributions is particularly valuable for table setting, where multiple valid placement trajectories exist (approach from left vs. right) but the final position must be precise. Training on 200-500 demonstrations per placement type (plate, fork, knife, glass), Diffusion Policy achieves sub-centimeter placement accuracy with 94% success rate on individual items.

Collection Methodology for Table Setting Data

Table setting data collection requires instrumented workstations with overhead and angled RGB-D cameras, a calibrated tabletop workspace, and a diverse inventory of real tableware. The primary sensor setup uses a high-resolution overhead camera (4K RGB at 30 Hz) for precise placement verification, supplemented by two angled side cameras for approach trajectory capture and a wrist-mounted camera for close-up grasp views. A depth sensor provides the 3D surface geometry needed for height-sensitive placement (stacking plates, positioning stemware).

Tableware diversity is critical for generalization. A production dataset should include at least 20 distinct place setting sets covering: standard ceramic dinnerware (various sizes and shapes), fine china (thin, fragile, with varying rim profiles), glassware (water tumblers, wine stems, champagne flutes), metal utensils (different handle profiles and weights), and cloth napkins (folded in multiple styles). Each setting should be demonstrated on at least 3 different table surfaces (bare wood, tablecloth, glass) to ensure the policy handles varying friction and reflectivity.

Annotation must capture both the absolute placement position and the relational structure of the setting. For each item placed, record: (1) final position in table coordinates (x, y, orientation, to millimeter precision), (2) relational predicates relative to neighboring items (left-of plate, above knife, aligned-with plate-center), (3) cultural compliance label (correct/incorrect per the specified setting style), and (4) placement quality score (centered, aligned, stable). These structured annotations enable training both end-to-end policies and hybrid systems that separate spatial reasoning from manipulation execution.

Claru collects table setting data using instrumented tabletop workstations with overhead 4K cameras, precision-calibrated table coordinate systems, and inventories of 20+ tableware sets spanning casual, formal, and culturally diverse conventions. Our operators demonstrate complete place settings following specified cultural rules, with each item placement annotated with sub-centimeter coordinates, relational predicates, compliance labels, and quality scores. We deliver datasets formatted for Diffusion Policy, ACT, and VLA architectures with train/validation/test splits stratified by tableware set and cultural style.

Key Datasets for Table Setting and Arrangement

Public datasets for table setting are sparse. Most provide simulated or limited real-world coverage.

DatasetYearScaleReal/SimSetting StylesSpatial Annotations
BEHAVIOR-1K (Li et al.)20231,000 activity definitions incl. table settingSimulation (NVIDIA Omniverse)Western casual onlyObject poses in sim
RT-2 evaluation (Brohan et al.)2023~50 table setting configurationsReal (Everyday Robots)Simple 3-4 item settingsSuccess/failure only
SayCan kitchen (Ahn et al.)2022101 tasks incl. table arrangementRealInformal kitchen counterTask completion labels
TidyBot household (Wu et al.)202330 household environmentsRealPersonalized placementUser preference labels

References

  1. [1]Ahn et al.. Do As I Can, Not As I Say: Grounding Language in Robotic Affordances.” arXiv 2204.01691, 2022. Link
  2. [2]Wu et al.. TidyBot: Personalized Robot Assistance with Large Language Models.” IROS 2023, 2023. Link
  3. [3]Chi et al.. Diffusion Policy: Visuomotor Policy Learning via Action Diffusion.” RSS 2023, 2023. Link
  4. [4]Huang et al.. Inner Monologue: Embodied Reasoning through Planning with Language Models.” CoRL 2023, 2023. Link
  5. [5]Li et al.. BEHAVIOR-1K: A Human-Centered Benchmark for Embodied AI.” CoRL 2023, 2023. Link

Frequently Asked Questions

For a single cultural style (e.g., Western casual 4-item setting), 200-500 demonstrations with Diffusion Policy achieve 94% single-item placement success. For a complete formal setting (8-12 items), 1,000-2,000 demonstrations capture the sequential dependencies and collision avoidance needed for end-to-end execution. For cross-style generalization (Western formal + casual + Asian conventions), 3,000-5,000 demonstrations across styles are recommended, with at least 500 per cultural convention. Each demonstration should use a different tableware set to prevent overfitting to specific item geometries.

The minimum viable setup is an overhead RGB camera at 4K resolution for sub-centimeter placement verification, plus a wrist-mounted camera for grasp-phase observations. Adding a depth sensor (structured light or stereo) enables 3D workspace understanding critical for height-sensitive placement like stacking plates or positioning stemware. Proprioceptive data at 50+ Hz captures the gentle placement dynamics needed to avoid chipping fragile tableware. Force/torque sensing at the wrist is valuable for detecting surface contact during placement, reducing the risk of excessive impact force on delicate items like wine glasses.

Cultural setting styles should be treated as distinct task categories with shared low-level manipulation skills. Western formal, Western casual, Japanese kaiseki, Chinese banquet, and Indian thali settings each have specific spatial rules governing item placement. Annotate each demonstration with the cultural style label, and within each style, annotate per-item compliance with the style rules. A hybrid architecture works best: train a shared manipulation policy on all styles for the low-level pick-and-place skills, then train style-specific spatial planners on the relational rules for each convention. This architecture requires only 200-300 demonstrations per new cultural style once the base manipulation policy is trained.

Simulation is useful for pre-training spatial reasoning (learning that forks go left of plates) but cannot replace real data for the manipulation component. The sim-to-real gap is particularly severe for tableware because of material diversity (ceramic friction, glass reflectivity, metal weight distribution), fragility constraints (force limits that vary by material), and visual appearance (the sheen of porcelain, transparency of glass, and patina of silver cannot be rendered accurately enough for visual policy transfer). A practical approach uses simulation for spatial layout planning (80% sim) combined with real demonstrations for manipulation execution (100% real). The spatial planner transfers well; the manipulation policy does not.

Sub-centimeter absolute accuracy and sub-2-degree rotational accuracy are the minimum for production deployment. Human table setters typically achieve 3-5mm placement consistency, and diners notice misalignment beyond 1cm. For utensil placement, rotational accuracy matters more than positional: a fork rotated 5 degrees looks wrong even if positioned correctly. For glass placement, height accuracy matters — a wine glass placed too close to the plate edge risks being knocked over. Annotation should capture position error, rotation error, and stability margin (distance from nearest collision) for each placement, enabling curriculum training from coarse to fine placement accuracy.

Get a Custom Quote for Table Setting Data

Share your target setting styles and tableware types, and we will design a data collection plan with the precise spatial annotations your policy needs.