Training Data for PaLM-E

PaLM-E is Google DeepMind's 562-billion-parameter embodied multimodal language model that grounds language reasoning in real-world sensor data. This guide covers its multi-modal input requirements, robot interaction data formats, training volumes, and how Claru delivers data for PaLM-E-style embodied reasoning pipelines.

OrganizationGoogle DeepMind / TU Berlin
Year2023

Input/Output Specification

Observation

Multi-modal: 224x224 or 512x512 RGB images (ViT-22B encoder) + optional scene state vectors and object-centric representations

Action

High-level natural language plans (step-by-step action sequences), not low-level motor commands; downstream controllers (RT-1, SayCan primitives) execute the plans

Language

Natural language task instructions and reasoning queries, processed by PaLM 540B language backbone

Control Freq

Plan-level (~1-2 Hz for plan generation); downstream motor controllers operate at 3-10 Hz

How Claru Data Integrates with PaLM-E

Claru provides the embodied interaction data that PaLM-E-style models require: RGB observation sequences paired with natural language task descriptions and step-by-step execution plans. Our annotators produce planning-level descriptions ('pick up the blue mug, move to the sink, place mug in sink') rather than motor-level labels, matching PaLM-E's training format. We deliver data as interleaved image-text episodes in JSON format compatible with open-source VLM fine-tuning pipelines. Our collection covers kitchen manipulation, tabletop rearrangement, mobile manipulation, and multi-object sequential tasks -- the task domains PaLM-E was designed to handle.

What Is PaLM-E?

PaLM-E (Pathways Language Model - Embodied) is an embodied multimodal large language model introduced by Danny Driess, Fei Xia, Mehdi S. M. Sajjadi, and collaborators at Google DeepMind and TU Berlin in March 2023 (arXiv:2303.03378). At 562 billion parameters, PaLM-E is the largest vision-language model ever applied to robotics. Its core innovation is embedding continuous sensor data -- RGB images, scene state vectors, and object-centric representations -- directly into the token sequence of a large language model (PaLM 540B), enabling the LLM to reason about the physical world and generate executable robot plans.

The key insight behind PaLM-E is that robot planning requires grounded reasoning: understanding what objects are in a scene, what actions are physically possible, and how those actions change the state of the world. Traditional robot planners rely on hand-crafted symbolic representations of the world, which break down when the environment is complex or novel. PaLM-E replaces this symbolic pipeline by training an LLM to reason directly from raw sensor data, generating step-by-step plans in natural language that a low-level controller then executes.

PaLM-E demonstrated several breakthrough capabilities. It could perform multi-step mobile manipulation planning from a single RGB image ('bring me the chips from the kitchen counter'), reason about affordances ('can I push the heavy box?' vs. 'can I push the paper cup?'), handle failures by re-planning when an action does not produce the expected outcome, and transfer knowledge across robot embodiments. Crucially, Driess et al. showed that PaLM-E exhibited positive transfer: training on a mixture of internet-scale vision-language data and robot-specific interaction data produced a model that was better at both tasks than models trained on either data source alone.

PaLM-E operates at the planning level, not the motor control level. It generates high-level action plans ('1. Navigate to counter. 2. Pick up bag of chips. 3. Navigate to user. 4. Hand over chips.') that are then executed by lower-level policies (e.g., RT-1 or SayCan primitives). This architectural separation means PaLM-E's data requirements are fundamentally different from end-to-end VLA models like OpenVLA or Octo: instead of continuous motor commands, PaLM-E consumes and produces natural language plans grounded in visual observations.

Key Metrics from the Paper

562B
Total parameters
540B
PaLM language backbone
22B
ViT visual encoder (ViT-22B)
3
Robot embodiments tested
6
Task domains demonstrated
1st
LLM to show positive transfer across embodied+VL tasks

PaLM-E Input/Output Specification

ParameterSpecification
Visual InputRGB images (224x224 or 512x512) encoded by ViT-22B or ViT-4B into continuous embedding tokens
State Input (optional)Scene state vectors (object positions, bounding boxes) or object-centric representations (OSRT)
Language InputNatural language task instructions or questions about the scene
OutputNatural language plans (step-by-step action sequences) or answers to embodied reasoning questions
Action LevelHigh-level planning (not low-level motor commands); plans executed by downstream controllers like RT-1 or SayCan
Sensor EmbeddingContinuous sensor tokens interleaved with text tokens in the PaLM input sequence
Control FrequencyPlan-level (~1-2 Hz for plan generation; downstream controllers operate at 3-10 Hz)

Architecture and Key Innovations

PaLM-E's architecture is a standard decoder-only transformer (PaLM 540B) with one critical modification: continuous sensor inputs are projected into the language model's token embedding space and interleaved with text tokens in the input sequence. For visual inputs, ViT-22B encodes an RGB image into a sequence of patch embeddings (e.g., 256 tokens for a 224x224 image with 14x14 patches). These visual tokens are then linearly projected into the same dimensionality as PaLM's text token embeddings and inserted at specific positions in the input sequence marked by special placeholder tokens.

The paper explored three types of sensor encodings: (1) ViT-based image encoding, where raw RGB images are processed through a Vision Transformer to produce patch-level embeddings; (2) object-centric representations using OSRT (Object Scene Representation Transformer), which encodes the scene as a set of object-level feature vectors; and (3) state vector encoding, where low-dimensional scene state (object positions, orientations, bounding boxes) is linearly projected into the token space. The ViT-based encoding performed best overall, but combining ViT features with state vectors showed the strongest results on tasks requiring precise spatial reasoning.

A major innovation in PaLM-E is the demonstration of positive transfer between internet-scale VLM training and robot-specific training. When trained jointly on visual question answering (VQA), image captioning, and robot planning tasks, PaLM-E performed better on all tasks compared to models trained on any single data mixture. This transfer was especially pronounced at scale: the 562B model showed stronger positive transfer than the 12B and 84B variants. This finding suggests that the knowledge encoded in web-scale language and vision data (commonsense reasoning, object recognition, spatial understanding) directly benefits embodied reasoning.

PaLM-E also introduced geometric input encodings that inject 3D structural information into the language model. Neural radiance fields (NeRFs) and point clouds can be encoded as token sequences and inserted alongside image and text tokens. This allows PaLM-E to reason about 3D spatial relationships, occlusions, and viewpoint-dependent visibility -- capabilities that are critical for real-world robot planning but absent from most 2D vision-language models.

PaLM-E vs. Related Embodied Reasoning Models

FeaturePaLM-E (2023)SayCan (2022)RT-2 (2023)Code as Policies (2023)
ArchitectureLLM + multi-modal sensor embeddingLLM + value function scoringVLM + direct action predictionLLM generating executable Python code
Action LevelHigh-level NL plansSelects from primitive skill libraryLow-level motor commandsCode-based policy composition
Sensor GroundingDirect (images embedded as tokens)Indirect (via affordance scoring)Direct (image-to-action)None (text-only reasoning)
Failure RecoveryRe-plans from new observationRe-scores affordancesNo explicit re-planningConditional logic in generated code
Parameters562B540B (PaLM) + learned affordances55B540B (PaLM, no fine-tuning)
Open SourceNo (paper only)NoNoCode released, model closed

Training Data Requirements

PaLM-E's training data is a mixture of internet-scale vision-language data and robot-specific interaction data. The internet-scale component is standard VLM training data: image-text pairs from web crawls, visual question answering datasets (VQAv2, OK-VQA), image captioning datasets (COCO Captions), and other multimodal benchmarks. This data teaches the model visual recognition, commonsense reasoning, and language fluency. The paper reports training on the same data mixture as PaLI (Chen et al., 2022), which includes billions of image-text pairs.

The robot-specific component consists of embodied interaction data from three platforms: (1) a mobile manipulation robot (similar to the Everyday Robots platform) performing kitchen tasks like fetching objects, opening drawers, and wiping surfaces; (2) a tabletop robot arm performing language-conditioned pick-and-place and stacking tasks; and (3) a simulated tabletop environment (Language-Table) with pushing and arrangement tasks. Each robot interaction episode contains: an initial RGB observation, a natural language task description, a sequence of intermediate observations, and the step-by-step plan (in natural language) that was executed to accomplish the task.

The critical data format for robot interaction episodes is the interleaved multi-modal sequence. A training example looks like: [image_1] 'Bring me the chips from the counter.' [image_2] 'Step 1: Navigate to counter.' [image_3] 'Step 2: Pick up bag of chips.' [image_4] 'Step 3: Navigate to user.' [image_5] 'Step 4: Hand over chips.' -- where each [image_N] is an RGB frame from the robot's camera encoded as visual tokens. The model learns to generate the step-by-step text plan conditioned on the interleaved visual observations.

For teams building PaLM-E-style embodied reasoning systems, the key data requirement is paired visual observation + language plan data. Each interaction episode needs: (1) RGB images captured at key moments during task execution (typically 3-10 frames per episode showing the start state, intermediate states, and goal state); (2) a natural language task description; and (3) a step-by-step natural language plan describing the actions taken. Annotations should describe what the robot does in each step at a semantic level ('pick up the blue cup'), not at a motor level ('move end-effector to [0.3, 0.2, 0.1]'). This planning-level annotation is what distinguishes PaLM-E data from the motor-level data required by VLA models like OpenVLA or Octo.

How Claru Data Integrates with PaLM-E

Claru provides the embodied interaction data that PaLM-E-style models require: RGB observation sequences paired with natural language task descriptions and step-by-step execution plans. Our annotation pipeline produces planning-level descriptions (not motor-level action labels) -- human annotators watch robot task execution videos and write the multi-step plans that a language model should learn to generate. Each episode includes 3-10 keyframe observations with paired plan steps, matching the interleaved multi-modal format PaLM-E uses.

For teams building PaLM-E-inspired systems at smaller scale (using open-source LLM backbones like Llama 3 or Gemma 2 instead of the closed PaLM 540B), Claru delivers data in a flexible format: JSON episodes with base64-encoded or file-referenced images, task descriptions, and step-by-step plans that can be tokenized and interleaved by any VLM training pipeline. We also provide scene-level metadata -- object identities, approximate positions, and affordance labels -- that can be used as supplementary state vector inputs alongside RGB observations.

Claru's collection network spans the task domains PaLM-E was designed to handle: kitchen manipulation, tabletop rearrangement, mobile manipulation in home and office environments, and multi-object interaction tasks requiring sequential planning. We capture both RGB keyframes and full video for each episode, enabling teams to choose their preferred observation sampling strategy. All language annotations are written by trained annotators following a planning-level description protocol that produces the grounded, actionable text PaLM-E's training objective expects.

Key References

  1. [1]Driess, D., Xia, F., Sajjadi, M. S. M., Lynch, C., Chowdhery, A., et al.. PaLM-E: An Embodied Multimodal Language Model.” ICML 2023 / arXiv 2303.03378, 2023. Link
  2. [2]Chowdhery, A., Narang, S., Devlin, J., et al.. PaLM: Scaling Language Modeling with Pathways.” arXiv 2204.02311, 2022. Link
  3. [3]Ahn, M., Brohan, A., Brown, N., et al.. Do As I Can, Not As I Say: Grounding Language in Robotic Affordances.” arXiv 2204.01691, 2022. Link
  4. [4]Brohan, A., Brown, N., Carbajal, J., et al.. RT-1: Robotics Transformer for Real-World Control at Scale.” RSS 2023 / arXiv 2212.06817, 2023. Link
  5. [5]Brohan, A., Brown, N., Carbajal, J., et al.. RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control.” CoRL 2023 / arXiv 2307.15818, 2023. Link
  6. [6]Liang, J., Huang, W., Xia, F., et al.. Code as Policies: Language Model Programs for Embodied Control.” ICRA 2023 / arXiv 2209.07753, 2023. Link

Frequently Asked Questions

VLA models like OpenVLA and Octo are end-to-end: they take an image and language instruction and directly output low-level motor commands (e.g., 7-DoF end-effector deltas at 5-10 Hz). PaLM-E operates at the planning level: it takes an image and a task description and generates a step-by-step natural language plan ('1. Pick up cup. 2. Move to sink. 3. Place cup in sink.'). This plan is then executed by a separate low-level policy (like RT-1) that handles the actual motor control. The advantage is that PaLM-E can reason about multi-step tasks requiring sequential logic, while VLA models are typically limited to single-step actions.

Yes. PaLM-E's architecture -- embedding visual tokens into an LLM's input sequence -- has been replicated with open-source components. You can use Llama 3 or Gemma 2 as the language backbone and a ViT (e.g., DINOv2 or SigLIP) as the visual encoder. The key ingredient is the robot interaction training data: interleaved image-text sequences where the model learns to generate grounded plans. Claru provides this data in formats compatible with open-source VLM training pipelines. The quality gap between open replicas and the original PaLM-E comes primarily from scale (562B vs. 7-70B parameters) and the volume of internet-scale pre-training data.

PaLM-E needs embodied interaction episodes consisting of: (1) RGB images captured at key moments during task execution (3-10 frames per episode), (2) a natural language task description, and (3) a step-by-step natural language plan describing the actions taken. The plans should be at the semantic level ('pick up the red cup') not the motor level ('move to [x,y,z]'). This is fundamentally different from VLA training data, which requires continuous action labels at every timestep. PaLM-E's data is cheaper to annotate because the labels are natural language descriptions rather than precise motor commands.

The paper does not disclose exact counts, but the robot data component is relatively small compared to the internet-scale VLM data. PaLM-E used interaction data from three platforms: a mobile manipulator (thousands of episodes across kitchen tasks), a tabletop robot arm (hundreds of episodes of pick-and-place and stacking), and a simulated environment (Language-Table, with thousands of episodes). The critical finding was positive transfer: internet-scale VLM training dramatically improved robot planning performance, meaning the model needed less robot-specific data than expected.

No. PaLM-E is a Google DeepMind research project with no public model weights or training code. The paper describes the architecture and training procedure in detail, and several groups have built open-source replicas using Llama-based backbones. For teams that want PaLM-E-style embodied reasoning, the practical path is to use an open-source VLM (e.g., Llava, Prismatic, or Idefics2) and fine-tune it on interleaved image-plan data. Claru provides the robot interaction data needed for this fine-tuning step.

Get Embodied Reasoning Data for PaLM-E-Style Models

Tell us about your embodied reasoning project -- target tasks, robot platform, and LLM backbone -- and we will deliver interleaved image-plan datasets with the observation format, planning-level annotations, and scene metadata your training pipeline requires.