M2R2: MultiModal Robotic Representation for Temporal Action Segmentation

2026 IEEE International Conference on Robotics & Automation (ICRA)
1Technische Universität Wien (TU Wien)   2German Aerospace Center (DLR)

Abstract

Temporal action segmentation (TAS) has long been a key area of research in both robotics and computer vision. In robotics, algorithms have primarily focused on leveraging proprioceptive information to determine skill boundaries, with recent approaches in surgical robotics incorporating vision. In contrast, computer vision typically relies on exteroceptive sensors, such as cameras. Existing multimodal TAS models in robotics integrate feature fusion within the model, making it difficult to reuse learned features across different models. Meanwhile, pretrained vision-only feature extractors commonly used in computer vision struggle in scenarios with limited object visibility. In this work, we address these challenges by proposing M2R2, a multimodal feature extractor tailored for TAS, which combines information from both proprioceptive and exteroceptive sensors. We introduce a novel training strategy that enables the reuse of learned features across multiple TAS models. Our method sets a new state-of-the-art performance on three robotic datasets REASSEMBLE, (Im)PerfectPour, and JIGSAWS. Additionally, we conduct an extensive ablation study to evaluate the contribution of different modalities in robotic TAS tasks.

M2R2 Overview

Fig. 1: Many multimodal approaches follow an end-to-end paradigm (a). Modular approaches train the feature extractor and TAS model separately but rely on vision-only extractors (b). We propose the M2R2 feature extractor, which leverages multimodal information and integrates with diverse TAS models (c).

Contributions

  • Multimodal feature extractor — a deep-learning-based M2R2 feature extractor that fuses vision, audio, and proprioceptive data for robotic temporal action segmentation.
  • Novel training strategy — a modular training approach that decouples feature extraction from the TAS model, enabling reuse of learned features with any state-of-the-art TAS architecture (MSTCN, ASRF, DiffAct, …).
  • Extensive evaluation — state-of-the-art performance on three robotic datasets (REASSEMBLE, (Im)PerfectPour, JIGSAWS) and an ablation study quantifying the contribution of each sensor modality.

Method

M2R2 Feature Extractor Architecture


To compute the multimodal feature at time instant ti, we first process each modality separately: image features are extracted with an ActionCLIP visual transformer; audio features are extracted from the log-mel spectrogram via the Audio Spectrogram Transformer (AST); proprioceptive signals (end-effector pose, twist, gripper width, force/torque) are projected into a shared embedding space and fused over the temporal dimension using a learnable embedding. All modality tokens are then fused with a single self-attention Transformer encoder layer followed by an MLP.


M2R2 Architecture

Fig. 2: M2R2 Feature Extractor Architecture.

Training Strategy


We train the M2R2 feature extractor using two complementary objectives. First, we sample windows of three consecutive actions and maximize the cosine similarity between the window's average feature embedding and a CLIP text embedding of a sentence describing the action sequence — encouraging the model to capture action order. Second, a Boundary Regression Network predicts the probability of each frame being an action boundary, supervised with a smoothed ground-truth boundary sequence. The total loss is the sum of the action-contrastive loss and the boundary MSE loss.

After training, the Fusion Transformer, Boundary Regression Network, and CLIP text encoder are discarded. The resulting per-frame M2R2 features are used to train any downstream TAS model.
M2R2 Training

Fig. 3: M2R2 training objectives.

Experiments

REASSEMBLE Dataset

M2R2 features achieve state-of-the-art performance on the REASSEMBLE dataset across all TAS models. M2R2+ASRF attains an F1@50 of 82.4% on fine-grain labels and 97.1% on coarse labels, outperforming the best vision-only baseline by over 74 percentage points. Unsupervised proprioception-only baselines (BOCPD, AWE) are surpassed by at least 46.6 pp in F1@50.

Method Modalities Type Fine-grain Labels Coarse Labels
F1@10F1@25F1@50AccEDIT F1@10F1@25F1@50AccEDIT
BOCPDFTE2E 39.232.612.835.019.7
AWEPE2E 72.568.435.854.654.9
BrP+MSTCNVModular 12.910.98.47.918.4 32.128.220.065.034.0
BrP+ASRFVModular 12.811.18.66.419.4 57.644.523.354.367.5
BrP+DiffActVModular 12.010.15.15.721.4 62.951.822.554.873.4
M2R2+MSTCNV,A,P,FT,GModular 83.182.780.882.479.3 97.797.696.796.795.0
M2R2+ASRFV,A,P,FT,GModular 83.583.482.482.782.5 98.198.197.196.096.9
M2R2+DiffActV,A,P,FT,GModular 78.177.774.674.968.7 95.195.092.191.790.3

Table I: Results on REASSEMBLE. Bold = best, underline = second best. V – vision, A – audio, P – pose, FT – force/torque, G – gripper width.

(Im)PerfectPour Dataset — Generalization


We evaluate generalization by applying the M2R2 extractor trained on REASSEMBLE to the (Im)PerfectPour bartending dataset. M2R2+DiffAct achieves F1@50 of 89.7%, outperforming the best vision-only baseline by 16.1 percentage points. These results demonstrate that when available modalities are similar across datasets, M2R2 features transfer well to new task domains without retraining.
MethodModalitiesF1@10F1@25F1@50AccEDIT
BrP+MSTCNV81.073.963.676.888.9
BrP+ASRFV84.182.173.677.791.2
BrP+DiffActV81.679.971.874.483.1
M2R2+MSTCNV,P,FT,G92.689.884.693.988.9
M2R2+ASRFV,P,FT,G89.888.586.093.787.9
M2R2+DiffActV,P,FT,G90.589.789.793.383.0

Table II: Results on (Im)PerfectPour. M2R2 features trained on REASSEMBLE.

JIGSAWS Dataset — Cross-Embodiment

MethodModalitiesTypeF1@10F1@25F1@50AccEDIT
Lea et al.VE2En/rn/rn/r74.266.5
Weerasinghe et al.V,P,GE2E87.386.581.187.183.9
Atoum et al.V,PE2En/rn/rn/r90.389.0
BrP+MSTCNVModular53.241.723.038.776.8
BrP+ASRFVModular53.443.224.539.373.6
BrP+DiffActVModular59.649.630.244.581.9
M2R2+MSTCNV,P,GModular93.391.686.387.090.4
M2R2+ASRFV,P,GModular93.191.286.085.690.7
M2R2+DiffActV,P,GModular94.393.089.487.191.7

Table III: Results on JIGSAWS (Suturing, Leave-One-User-Out). n/r – not reported.


On the JIGSAWS surgical gesture recognition dataset, M2R2+DiffAct achieves an F1@50 of 89.4% and an EDIT score of 91.7%, surpassing prior state-of-the-art by 1.7% in EDIT. Unlike previous approaches that rely on additional binary gripper signals or invariant proprioceptive representations, M2R2 directly processes raw sensory readings. This demonstrates that M2R2 generalizes across different embodiments and task domains.

Qualitative Results

TAS Performance — REASSEMBLE, (Im)PerfectPour & JIGSAWS


REASSEMBLE — M2R2 features are less sensitive to high-frequency proprioceptive changes that cause heuristic approaches to over-segment, and more precise in differentiating visually similar objects (e.g., round vs. square pegs) compared to vision-only approaches. The segmentation bar shows ground truth (top) and predicted labels (bottom).

Qualitative evaluation

Fig. 4: Qualitative evaluation. M2R2 avoids over-segmentation from heuristics (left, middle) and differentiates objects better than vision-only approaches (right).



(Im)PerfectPour — M2R2 features trained on REASSEMBLE generalize to a new task domain (bartending) without retraining, achieving F1@50 of 89.7% and outperforming vision-only baselines by 16.1 pp.


JIGSAWS — M2R2 generalizes to a surgical robotics setting with a different embodiment (two-arm da Vinci). Using only raw pose, twist, and gripper width, M2R2+DiffAct achieves state-of-the-art F1@50 of 89.4% on the Suturing task.

Coarse vs fine-grain prediction

Fig. 5: Coarse vs. fine-grain prediction on REASSEMBLE. Large and medium gears are occasionally confused at fine-grain level, but coarse-level segmentation remains correct.

Ablation Study

Modality Contribution


We ablate the contribution of each sensor modality on the REASSEMBLE dataset using DiffAct. Key findings:
  • Vision alone performs worst (F1@50: 21.6%), as many objects are poorly visible due to their small size.
  • Audio alone achieves a high detection rate (80.1%) — action transitions are marked by distinct sounds — but struggles to identify objects.
  • Proprioception alone performs comparably to all modalities combined (F1@50: 74.5%), reflecting the specificity of the REASSEMBLE dataset.
  • Gripper width is the most informative proprioceptive signal — objects can often be distinguished by size.
  • All modalities combined yield the best detection rate (82.4%), showing that multimodal fusion improves boundary precision.
ConfigurationF1@10F1@25F1@50AccEDITDR
Only vision27.226.121.621.428.944.3
Only audio39.238.736.434.537.580.1
Only proprio78.077.574.575.469.280.3
No gripper67.967.264.265.159.282.2
No pose/vel69.969.165.164.661.880.5
No F/T72.972.469.970.465.280.6
All modalities78.177.774.674.968.782.4

Table IV: Modality ablation on REASSEMBLE with DiffAct (fine-grain labels). DR = Detection Rate.


Note: These ablation results are specific to the REASSEMBLE dataset, where objects are small and often identifiable by size or force profile. The relative importance of each modality is likely to differ in tasks with higher visual variability (e.g., meal preparation, where vision may be essential to distinguish similar-looking objects). Further investigation across more diverse datasets is needed to draw general conclusions about modality contributions in robotic TAS.

Audio-only ablation

(a) Some objects (e.g., waterproof connector) can be identified from audio alone.

No-gripper ablation

(b) Removing gripper information makes object size-based discrimination harder.

Fig. 6: Qualitative predictions for different modality combinations.

BibTeX

@inproceedings{sliwowski2025m2r2,
  title     = {{M2R2: MultiModal Robotic Representation for Temporal Action Segmentation}},
  author    = {Sliwowski, Daniel and Lee, Dongheui},
  booktitle = {2026 IEEE International Conference on Robotics & Automation (ICRA)},
  year      = {2026}
}