Bridging Perception and Action:
Spatially-Grounded Mid-Level Representations
for Robot Generalization

Mid-Level MoE Teaser

Bimanual, dexterous manipulation requires task-specific grounding. Different mid-level representations lead to different levels of improvement depending on the task.

Video Results

Mid-Level MoE: Preliminary Results

Demonstration of spatially-grounded mid-level representations for robot generalization across simulation and real-world dexterous bimanual manipulation tasks.

Abstract

In this work, we investigate how spatially-grounded auxiliary representations can provide both broad, high-level grounding, as well as direct, actionable information to help policy learning performance and generalization for dexterous tasks. We study these mid-level representations across four critical dimensions: object-centricity, motion-centricity, pose-awareness, and depth-awareness.

We use these interpretable mid-level representations to train specialist encoders via supervised learning, then use these representations as inputs to a diffusion policy to solve dexterous bimanual manipulation tasks in the real-world.

We propose a novel mixture-of-experts policy architecture that can combine multiple specialized expert models, each trained on a distinct mid-level representation, to improve the generalization of the policy. This method achieves an average of 11% higher success rate over a language-grounded baseline and a 24% higher success rate over a standard diffusion policy baseline for our evaluation tasks.

Furthermore, we find that leveraging mid-level representations as supervision signals for policy actions within a weighted imitation learning algorithm improves the precision with which the policy follows these representations, leading to an additional performance increase of 10%.

Method

Spatially-Grounded Mid-Level Representations

Our approach leverages four key types of mid-level representations that capture critical spatial and geometric properties for dexterous manipulation:

  • Object-Centric: Focus on object poses, sizes, shapes, and interaction points (e.g., grasp points, keypoints).
  • Motion-Centric: Capture dynamic aspects like object velocities, accelerations, and kinematic constraints.
  • Depth-Aware: Leverage depth information for spatial relationships and 3D reasoning.
  • Pose-Aware: Encode relative and absolute poses for precise manipulation tasks.

Architecture

Mid-Level MoE Architecture

Policy Architecture. Four images are passed into a transformer encoder. Each mid-level expert processes the image to generate specialized representations, which are combined through cross-attention.

Key Design Choices

  • Diverse Mid-level Experts: Four specialists corresponding to our representation axes
  • Attention-based Gating: Multi-headed attention for dynamic expert weighting
  • Cross-Attention Mechanism: Integration of visual and representation information

The Sensitivity-Robustness Tradeoff

Sensitivity-Robustness Tradeoff

The sensitivity-robustness tradeoff. Policies need to follow their mid-level representations while being robust to erroneous noise.

We identify a fundamental tradeoff between the sensitivity with which a robot follows its representations and its robustness to errors in these representations. Our architecture design balances:

  • Sensitivity: How well the policy adheres to provided mid-level representations
  • Robustness: Resilience to noise and perturbations in representations

Self-Consistency Training

Self-Consistency

Self-Consistency. We weight demonstrations based on how well the robot's trajectory matches its mid-level representation.

We introduce weighted imitation learning with self-consistency weights that emphasize demonstrations where the policy's actions align well with the mid-level expert outputs. This approach:

  • Improves policy precision in following representations
  • Reduces sensitivity while maintaining robustness
  • Creates a feedback loop for better expert utilization

Experimental Results

24%
improvement over standard diffusion policy
11%
improvement over language baseline
10%
additional improvement with self-consistency

Simulation Results

Simulation Results

Simulation Results. Mid-Level MoE achieves consistently high performance across different tasks by leveraging task-specific representations.

Task-Specific Benefits

Different representations excel at different tasks:

  • Motion-centric representations for insertion and stacking tasks
  • Object-centric representations for fruit bowl and kitchen organization
  • Pose-aware representations for assembly tasks
  • Depth-aware representations for handover and fine manipulation

Real-World Results

Real-World Results

Real-World Results. Clear differences in the benefits that different representations provide for real-world dexterous manipulation tasks.

Generalization to Real World

Our method demonstrates strong performance across diverse real-world scenarios:

  • Kitchen Stack: Organizing and stacking kitchen items
  • Cup Stacking: Combinatorial object arrangements
  • Shirt Hanging: Deformable object manipulation
  • Shoelace Tying: Fine motor control and dexterity

Self-Consistency Analysis

Self-Consistency Results

Architecture and Self-Consistency Ablation. Weighted Mid-Level MoE achieves 10% higher success rate than unweighted across real-world tasks.

Architectural Benefits

Our analysis reveals:

  • Cross-attention outperforms concatenation and early fusion
  • Lower sensitivity scores correlate with better performance
  • Higher robustness indices indicate stable performance
  • Self-consistency training provides consistent improvements

Paper & Resources

Research Paper

Read the full paper with detailed methodology, experiments, and analysis.

arXiv Paper

Source Code

Implementation of Mid-Level MoE architecture and training procedures.

GitHub Repository

Dataset

Training data and evaluation benchmarks for bimanual manipulation tasks.

Download Dataset

Citation

@article{yang2024midlevelmoe,
  title={Bridging Perception and Action: Spatially-Grounded Mid-Level Representations for Robot Generalization},
  author={Yang, Jonathan and Fu, Chuyuan Kelly and Shah, Dhruv and Sadigh, Dorsa and Xia, Fei and Zhang, Tingnan},
  year={2024},
  journal={arXiv preprint}
}