Skip to content

A curated list of recent robot learning papers incorporating diffusion models for robotics tasks.

Notifications You must be signed in to change notification settings

showlab/Awesome-Robotics-Diffusion

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

21 Commits
 
 

Repository files navigation

Awesome Robot Diffusion Awesome

A curated list of recent robot learning papers incorporating diffusion models for manipulation, navigation, planning etc.

Table of Contents

Benchmarks

Visual Pusher

Panda Arm

Dexdeform: Dexterous deformable object manipulation with human demonstrations and differentiable physics (To be checked)

Diffusion Policy

Manipulation

Is Conditional Generative Modeling all you need for Decision-Making? (Decision Diffuser)

Provable Guarantees for Generative Behavior Cloning: Bridging Low-Level Stability and High-Level Behavior (Highly Theoretical)

Flow Matching Imitation Learning for Multi-Support Manipulation (Flow matching)

Cold Diffusion on the Replay Buffer: Learning to Plan from Known Good States

DiMSam: Diffusion Models as Samplers for Task and Motion Planning under Partial Observability (With constraint)

Scaling Diffusion Policy in Transformer to 1 Billion Parameters for Robotic Manipulation

ViViDex: Learning Vision-based Dexterous Manipulation from Human Videos

3D-ViTac: Learning Fine-Grained Manipulation with Visuo-Tactile Sensing (Include tactile)

SplatSim: Zero-Shot Sim2Real Transfer of RGB Manipulation Policies Using Gaussian Splatting (Sim2Real)

Movement Primitive Diffusion: Learning Gentle Robotic Manipulation of Deformable Objects (Deformable)

SculptDiff: Learning Robotic Clay Sculpting from Humans with Goal Conditioned Diffusion Policy (3D deformable objects)

Pushing the Limits of Cross-Embodiment Learning for Manipulation and Navigation (Both manipulation and navigation)

Diffusion Co-Policy for Synergistic Human-Robot Collaborative Tasks

PianoMime: Learning a Generalist, Dexterous Piano Player from Internet Demonstrations

GenDP: 3D Semantic Fields for Category-Level Generalizable Diffusion Policy (single task generalizability)

Adaptive Compliance Policy: Learning Approximate Compliance for Diffusion Guided Control (considering compliance / forces during manipulation)

ForceMimic: Force-Centric Imitation Learning with Force-Motion Capture System for Contact-Rich Manipulation (forces centric)

Learning Diffusion Policies from Demonstrations For Compliant Contact-rich Manipulation (considering compliance / forces)

CAGE: Causal Attention Enables Data-Efficient Generalizable Robotic Manipulation (single task generalization)

Language-Guided Object-Centric Diffusion Policy for Collision-Aware Robotic Manipulation (3D, object centric, single task generalization)

Affordance-Centric Policy Learning: Sample Efficient and Generalisable Robot Policy Learning using Affordance-Centric Task Frames (object centric)

MaIL: Improving Imitation Learning with Selective State Space Models (using Mamba)

Crossway Diffusion: Improving Diffusion-based Visuomotor Policy via Self-supervised Learning

The Ingredients for Robotic Diffusion Transformers

Discrete Policy: Learning Disentangled Action Space for Multi-Task Robotic Manipulation

C3DM: Constrained-Context Conditional Diffusion Models for Imitation Learning (tackle spurious correlation)

DexGrasp-Diffusion: Diffusion-based Unified Functional Grasp Synthesis Method for Multi-Dexterous Robotic Hands

Sampling Constrained Trajectories Using Composable Diffusion Models (Trajectory optimization with constraints present)

SPOT: SE(3) Pose Trajectory Diffusion for Object-Centric Manipulation (Tracking object pose)

Unpacking Failure Modes of Generative Policies: Runtime Monitoring of Consistency and Progress (out-of-distribution scenarios, detect failures)

FREE FROM BELLMAN COMPLETENESS: TRAJECTORY STITCHING VIA MODEL-BASED RETURN-CONDITIONED SUPERVISED LEARNING

One-Shot Imitation under Mismatched Execution

LANGUAGE CONTROL DIFFUSION: EFFICIENTLY SCALING THROUGH SPACE, TIME, AND TASKS

Locomotion

Diffusion-based learning of contact plans for agile locomotion

DIDI: Diffusion-Guided Diversity for Offline Behavioral Generation (To be checked)

Preference Aligned Diffusion Planner for Quadrupedal Locomotion Control (out-of-distribution issue)

Grasping / Reorientation

Learning Visuotactile Skills with Two Multifingered Hands

Navigation / Task and Motion Planning

NoMaD: Goal Masked Diffusion Policies for Navigation and Exploration

DiPPeR: Diffusion-based 2D Path Planner applied on Legged Robots

SafeDiffuser: Safe Planning with Diffusion Probabilistic Models

LTLDoG: Satisfying Temporally-Extended Symbolic Constraints for Safe Diffusion-based Planning

LDP: A Local Diffusion Planner for Efficient Robot Navigation and Collision Avoidance

DTG : Diffusion-based Trajectory Generation for Mapless Global Navigation

DiffusionSeeder: Seeding Motion Optimization with Diffusion for Rapid Motion Planning (motion planning)

Potential Based Diffusion Motion Planning

Others

DroneDiffusion: Robust Quadrotor Dynamics Learning with Diffusion Models (Drones)

Diffusion Generation Models in Robot Learning

[unread]

GR-MG: Leveraging Partially-Annotated Data via Multi-Modal Goal-Conditioned Policy (generate goal image)

Generative Image as Action Models

Scaling Robot Learning with Semantically Imagined Experience

Large-Scale Actionless Video Pre-Training via Discrete Diffusion for Efficient Policy Learning

Diffusion model is an effective planner and data synthesizer for multi-task reinforcement learning

3D Vision-Language-Action Generative World Model

World models in robotics:

IRASim: Learning Interactive Real-Robot Action Simulators arXiv 2024.6

Structured World Models from Human Videos RSS 2023

HARP: Autoregressive Latent Video Prediction with High-Fidelity Image Generator ICIP 2022

DayDreamer: World Models for Physical Robot Learning CoRL 2022

Other work studying data generation in robotics (Not using diffusion)

MimicGen: A Data Generation System for Scalable Robot Learning using Human Demonstrations

Robot Learning Utilizing Diffusion Model Properties

PoCo: Policy Composition from and for Heterogeneous Robot Learning (To be checked)

Diffusion Forcing: Next-token Prediction Meets Full-Sequence Diffusion (To be checked)

Pre-trained Text-to-Image Diffusion Models Are Versatile Representation Learners for Control

Imitation Learning from Purified Demonstrations (using forward & reverse diffusion process to purify imperfect demonstrations)

About

A curated list of recent robot learning papers incorporating diffusion models for robotics tasks.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •