Force-Torque Sensing: Measuring Contact Forces for Robot Manipulation

Force-torque (F/T) sensing measures the six-dimensional forces and moments (three linear forces, three rotational torques) at a robot's joints or end-effector. This data modality is essential for contact-rich manipulation tasks where vision alone cannot determine whether a grasp is secure, an insertion is aligned, or a surface is being contacted with appropriate pressure.

What Is Force-Torque Sensing?

Force-torque (F/T) sensing is the measurement of the six-dimensional generalized force vector at a point on a robot's kinematic chain. The six components consist of three linear forces (Fx, Fy, Fz) measuring push/pull along each Cartesian axis, and three torques (Tx, Ty, Tz) measuring rotational moments about each axis. This measurement provides a complete description of the mechanical interaction between the robot and its environment at the sensing point.

F/T sensing is fundamental to contact-rich manipulation — any task where the robot must control or respond to physical forces rather than simply position its end-effector in free space. Insertion tasks require detecting and responding to contact forces to achieve alignment. Assembly tasks require monitoring forces to verify component mating. Surface treatment tasks (polishing, cleaning, painting) require maintaining consistent contact pressure. In all these cases, visual observation alone cannot provide the necessary feedback because forces are not directly observable from images.

The two primary sensing approaches are dedicated F/T sensors (strain-gauge-based devices mounted at the wrist, manufactured by companies like ATI Industrial Automation and OnRobot) and joint torque sensors (integrated into each robot joint, as in the Franka Emika Panda and KUKA LBR iiwa). Dedicated sensors offer higher accuracy and bandwidth but add cost and mechanical complexity. Joint torque sensors are less accurate for end-effector force estimation due to model uncertainties in the robot's dynamics but require no additional hardware.

Historical Context

Force-torque sensing in robotics predates the modern era of learned policies by decades. The first 6-axis F/T sensors for robotics were developed in the 1970s and 1980s, primarily for force-controlled assembly in manufacturing. Raibert and Craig (1981) formalized hybrid position/force control, which used F/T measurements to switch between position control (in free space) and force control (during contact). This framework dominated industrial robotics for two decades.

The integration of F/T data into learned manipulation policies began with the resurgence of deep learning in robotics. Lee et al. (2019) demonstrated that self-supervised learning from vision and touch (including F/T measurements) produced better manipulation policies than either modality alone. Subsequent work by multiple groups confirmed that multi-modal policies incorporating F/T data outperform vision-only policies on contact-rich tasks by significant margins.

The emergence of collaborative robots with built-in joint torque sensing (Franka Panda in 2017, KUKA LBR iiwa in 2013) democratized access to force-aware manipulation. Previously, adding F/T sensing required purchasing expensive dedicated sensors. Now, many research labs have force-sensing capability built into their standard research platforms, leading to a rapid increase in datasets and policies that include F/T observations.

Practical Implications

For teams building manipulation datasets that include F/T sensing, the most critical practical concern is calibration. An uncalibrated F/T sensor — or one where the gravity compensation parameters are incorrect — introduces systematic biases that corrupt the learned policy. Best practice is to perform a calibration routine at the start of each data collection session: record sensor readings in multiple static orientations to estimate the tool mass, center of gravity, and sensor bias.

Data storage for F/T measurements is straightforward but requires attention to synchronization. At 500 Hz with 6 channels of 32-bit float data, a 30-second demonstration generates approximately 360 KB of F/T data — negligible compared to camera images. The challenge is precise temporal alignment with visual frames (30 Hz) and action labels (10-100 Hz). Hardware synchronization via shared triggers or timestamps is strongly recommended.

Claru's data collection infrastructure includes calibrated F/T sensing across our fleet of robot platforms. Each dataset includes raw and gravity-compensated F/T streams, calibration metadata (tool mass, center of gravity, sensor offset), and hardware-synchronized timestamps. This eliminates the sensor calibration and synchronization burden from downstream teams training contact-rich manipulation policies.

Common Misconceptions

MYTH

Vision-based policies can infer contact forces from deformation observations.

FACT

While large deformations (bending, squishing) are visible, the contact forces relevant to manipulation (0.1-50 N) typically produce sub-millimeter deformations on rigid objects that are invisible to standard cameras. Direct F/T measurement provides ground-truth force information that no vision system can reliably recover for rigid-body contact.

MYTH

Higher F/T sensor sample rates always produce better learned policies.

FACT

For learning-based manipulation, the bottleneck is typically the camera frame rate (30 Hz), not the F/T sample rate. Logging F/T at 10 kHz when the policy only processes observations at 30 Hz wastes storage without improving policy quality. A 100-500 Hz F/T log with proper downsampling to the policy's inference rate is optimal for most applications.

MYTH

All robot arms have built-in force-torque sensing.

FACT

Many popular robot platforms — Universal Robots UR5/UR10, Kinova Gen3, xArm — do not have direct joint torque sensors. They can estimate external forces from motor current, but this is significantly less accurate than dedicated torque sensors. Only a few platforms (Franka Panda, KUKA iiwa, Sawyer) include direct joint torque sensing.

Key Papers

  1. [1]Lee et al.. Making Sense of Vision and Touch: Self-Supervised Learning of Multimodal Representations for Contact-Rich Tasks.” ICRA 2019, 2019. Link
  2. [2]Luo et al.. Multisensory Robot Manipulation — Learning from Diverse Sensory Modalities.” arXiv 2024, 2024. Link
  3. [3]Haddadin et al.. Robot Collisions: A Survey on Detection, Isolation, and Identification.” IEEE T-RO 2017, 2017. Link

How Claru Supports This

Claru's multi-modal data collection pipelines include calibrated force-torque sensing for contact-rich manipulation tasks. Each demonstration includes synchronized 6-axis F/T measurements at 100+ Hz alongside calibrated camera streams and robot proprioception, ready for direct consumption by multi-modal policy training.

What Is Force-Torque Sensing?

Force-torque sensing refers to the measurement of the six components of generalized force at a point on a robot: three linear forces (Fx, Fy, Fz) and three rotational torques (Tx, Ty, Tz). These sensors are typically mounted between the robot's wrist and its end-effector (wrist-mounted F/T sensors) or integrated into the robot's joints (joint torque sensors). The combined 6-axis measurement provides a complete picture of how forces are distributed at the contact interface.

In robot learning, F/T data serves as a critical observation modality for tasks that involve physical contact. When a robot inserts a peg into a hole, the lateral forces indicate misalignment direction and magnitude. When a robot grasps a fragile object, the normal force indicates grip security without visual occlusion. When a robot wipes a surface, the normal force determines whether adequate pressure is being applied. These physical quantities cannot be reliably inferred from vision alone — they must be directly measured.

Modern collaborative robots like the Franka Emika Panda include joint torque sensors at every joint, providing estimated external forces at the end-effector through the robot's dynamics model. Dedicated wrist F/T sensors (ATI, OnRobot, Robotiq) offer higher accuracy and faster update rates (up to 10 kHz) than joint-torque-derived estimates. For learning-based manipulation, F/T data is typically logged at 100-1000 Hz alongside camera images (30 Hz) and robot proprioception (100-1000 Hz).

Force-Torque Sensing at a Glance

6-axis
Fx, Fy, Fz, Tx, Ty, Tz
10 kHz
Max sampling rate (dedicated)
0.01 N
Resolution of top sensors
ATI
Leading F/T sensor manufacturer
Franka
Built-in joint torque sensing
100 Hz+
Typical logging rate for ML

F/T Data in Robot Learning Pipelines

Incorporating F/T data into learned manipulation policies requires careful data pipeline design. The raw sensor signal is often noisy and includes gravitational and inertial components that must be compensated. Gravity compensation removes the weight of the end-effector and any attached tool from the measurement, isolating the contact forces. Inertial compensation accounts for dynamic effects during fast motions. Both compensations require accurate knowledge of the end-effector mass, center of gravity, and the robot's acceleration.

For training manipulation policies, F/T data is typically synchronized with visual observations and action labels at a common timestep frequency. Since F/T sensors operate at much higher frequencies than cameras, the F/T data is usually downsampled or windowed. A common approach is to provide the policy with a short history of F/T measurements (e.g., the last 10 samples at 100 Hz, covering 100 ms) at each camera frame, giving the policy both the current force state and recent force dynamics.

The impact of including F/T data varies dramatically by task. For tasks where contact forces are the primary feedback signal — insertion, assembly, polishing, palpation — adding F/T observations can improve policy success rates by 30-50% compared to vision-only policies. For tasks where visual feedback is sufficient — bin picking, sorting, stacking — F/T data provides marginal improvement. Understanding which tasks benefit from F/T sensing guides data collection investments.

F/T Sensor Types Compared

Different force-torque sensing approaches for robot manipulation, with trade-offs in accuracy, cost, and integration complexity.

Sensor TypeAxesSample RateCostIntegration
Wrist-mounted F/T (ATI)6Up to 10 kHz$3K-10KBetween wrist and tool
Joint torque sensors7 (per joint)1 kHzBuilt into robotNative (Franka, KUKA iiwa)
Tactile skin arraysNormal per taxel100-1000 Hz$1K-5KFingertip or palm
Current-based estimationN per joint1 kHzNo extra hardwareSoftware only (less accurate)

Training Data Collection with F/T Sensors

Collecting F/T-rich training data introduces specific requirements beyond standard visual demonstration collection. Sensor calibration must be performed before each data collection session — this involves measuring the sensor bias (zero offset under no-load conditions) and verifying gravity compensation parameters. Uncalibrated F/T data can have systematic biases of 1-5 N, which is significant for precision tasks where contact forces are in the 0.5-10 N range.

Temporal synchronization between F/T data and other modalities (cameras, joint encoders) must be precise. Force events during contact occur on millisecond timescales — a 50 ms synchronization error means the policy learns the wrong association between visual observations and contact forces. Hardware-level synchronization using shared clocks or trigger signals is preferred over software-level timestamping, which can introduce variable latency.

Claru's data collection pipelines include calibrated F/T sensing for contact-rich manipulation tasks. Each dataset includes synchronized 6-axis force-torque measurements at 100+ Hz alongside calibrated camera streams and robot proprioception, with documented calibration procedures and sensor specifications. This enables direct consumption by multi-modal manipulation policies that leverage force feedback for contact-rich tasks.

Key References

  1. [1]Lee et al.. Making Sense of Vision and Touch: Self-Supervised Learning of Multimodal Representations for Contact-Rich Tasks.” ICRA 2019, 2019. Link
  2. [2]Luo et al.. Multisensory Robot Manipulation — Learning from Diverse Sensory Modalities.” arXiv 2024, 2024. Link
  3. [3]Haddadin et al.. Robot Collisions: A Survey on Detection, Isolation, and Identification.” IEEE T-RO 2017, 2017. Link

Frequently Asked Questions

For most learning-based manipulation tasks, joint torque sensing (available on robots like Franka Panda and KUKA iiwa) provides sufficient accuracy. Dedicated wrist-mounted F/T sensors offer higher accuracy and faster response times, which matter for precision assembly and delicate manipulation. If your task involves contact forces below 1 N or requires sub-millisecond force feedback, a dedicated sensor is recommended.

For most manipulation learning applications, 100-500 Hz is sufficient. This captures the dynamics of typical contact events (grasping, insertion, wiping) while keeping dataset sizes manageable. Tasks with very fast contact dynamics (hammering, catching) may benefit from 1 kHz logging. Logging above 1 kHz is rarely needed for learning applications.

Tasks with significant contact phases benefit most: peg-in-hole insertion (30-50% improvement), assembly (20-40%), surface finishing/polishing (25-35%), and deformable object manipulation (15-25%). Tasks that are primarily visual — bin picking, sorting by color, open-space reaching — show marginal improvement from F/T data (0-5%).

Standard practice includes gravity compensation (removing end-effector weight), low-pass filtering (typically 10-50 Hz cutoff for learning applications), and bias subtraction (recording the zero-load baseline before each session). For learning pipelines, normalizing F/T values to zero-mean unit-variance per axis prevents force channels from being dominated by gravity or bias artifacts.

Need Force-Torque Training Data?

Claru provides synchronized multi-modal datasets with calibrated F/T sensing for contact-rich manipulation tasks.