key iconAmortizing Trajectory Diffusion
with Keyed Drift Fields

  • Anonymous authors

Overview

  • Keyed Drifting Policies (KDP) is a simple and effective one-step trajectory planning method for offline RL and robotics, designed to recover diffusion-like planning behavior without the expensive iterative denoising loop.
  • KDP is conditioning-aware. Rather than matching trajectories in the full high-dimensional window space, KDP forms neighborhoods in a compact key space aligned with the condition (for example, the current state), which avoids collapse toward average, inert trajectories.
  • KDP is trained by drift-based refinement. It pulls generated trajectories toward condition-matched dataset windows and repels them from nearby generated samples, using a stop-gradient drifted target to amortize refinement into the generator during training.
  • KDP is fast and performant. At inference, it generates a full trajectory window in a single forward pass, preserving multimodal candidate planning while substantially reducing planning latency compared with diffusion sampling.

Challenge

Challenge visualization
  • Diffusion trajectory planners are expressive but too slow for real-time control. In receding-horizon planning, each control step may require \(T\) sequential denoising steps for every candidate trajectory, so latency grows quickly and can make closed-loop planning brittle under tight compute budgets.
  • One-step conditional trajectory generation is easy to get wrong. If we naively match trajectories in the full high-dimensional window space, distances are dominated by unconstrained future dimensions, which pulls samples toward average trajectories, collapses diversity, and leads to near-static behavior instead of useful condition-consistent plans.
Key observation: Conditional trajectory generation requires a conditioning-aware notion of neighborhood.

Abstract

Diffusion-based trajectory planners can synthesize rich, multimodal action sequences for offline reinforcement learning, but their iterative denoising incurs substantial inference-time cost, making closed-loop planning slow under tight compute budgets. We study the problem of achieving diffusion-like trajectory planning behavior with one-step inference, while retaining the ability to sample diverse candidate plans and condition on the current state in a receding-horizon control loop. Our key observation is that conditional trajectory generation fails under naïve distribution-matching objectives when the similarity measure used to align generated trajectories with the dataset is dominated by unconstrained future dimensions. In practice, this causes attraction toward average trajectories, collapses action diversity, and yields near-static behavior. Our key insight is that conditional generative planning requires a conditioning-aware notion of neighborhood: trajectory updates should be computed using distances in a compact key space that reflects the condition, while still applying updates in the full trajectory space. Building on this, we introduce key Keyed Drifting Policies (KDP), a one-step trajectory generator trained with a drift-field objective that attracts generated trajectories toward condition-matched dataset windows and repels them from nearby generated samples, using a stop-gradient drifted target to amortize iterative refinement into training. At inference, the resulting policy produces a full trajectory window in a single forward pass. Across standard RL benchmarks and real-time hardware deployments, KDP achieves strong performance with one-step inference and substantially lower planning latency than diffusion sampling.


Algorithm

KDP architecture

Quantitative Results

Short-horizon continuous control

TABLE I: Performance in short horizon control on D4RL Locomotion (ME=Medium-Expert, M=Medium).
HalfCheetah Hopper Walker2D Avg↑ NFE↓
MEM MEM MEM
Offline selection
BC55.242.652.552.9107.575.364.3
CQL91.644.0105.458.5108.872.580.1
IQL86.747.491.566.3109.678.380.0
DT86.842.6107.667.6108.174.081.1
TT95.046.9110.061.1101.979.082.3
MOPO63.342.323.728.044.617.836.6
MOReL53.342.1108.795.495.677.878.8
MBOP105.944.655.148.870.241.060.9
Generative trajectory planners
DQL96.869.1111.190.5110.187.091.15
CAC84.369.1100.480.7110.483.188.02
CBC32.731.090.671.7110.483.169.92
Diffuser88.942.8103.374.3106.979.682.620
KDP92.562.1103.690.3108.587.290.71
TABLE II: Analysis of planning cost. Hardware: 1xNVIDIA A100. (D=Diffuser, UR=Unranked, R=Ranked).
Method NFE↓ BEF↓ PL E2E
BC111.011.22
D-UR(K=1, T=20)2020149.32149.51
D-R(K=16, T=20)20320172.16172.34
D-R(K=64, T=20)201280279.98280.03
KDP-UR(K=1)111.982.11
KDP-R(K=16)1323.974.10
KDP-R(K=64)11284.764.96

Long-horizon, goal-conditioned planning

TABLE III: Long horizon manipulation analysis. PL is in s/step. (UM=U-Maze, Med.=Medium, Lar.=Large).
Method Maze2D AntMaze
UMMed.Lar.PL UMMed.PL
CQL5.75.012.584.053.7
IQL47.434.958.662.270.0
Diffuser113.9121.5123.03.84876.031.96.153
KDP122.3130.4133.00.03181.367.60.052
TABLE IV: Training objective ablations. (Locom.=Locomotion, Act. div.=Action diversity).
Ablation Locom.↑ Maze2D↑ Act. div.↑
Full method (KDP)90.7128.20.93
No keying2.919.60.14
Include self-negatives78.0891.30.93
Attraction-only (no repulsion)1.053.72.2e6
No drift normalization86.93131.50.93
Single τ92.37102.10.93

Dexterous manipulation and high-DoF control

TABLE V: Adroit results.
Task Diffuser KDP
ScoreT(s) ScoreT(s)
pen-clone10.71.63453.40.020
door-clone56.71.59861.80.024
hammer-clone53.11.53279.60.034
relocate-clone56.21.68562.80.036
TABLE VI: Action chunking.
Task L Diff. KDP
adroit-pen138%57%
adroit-pen841%73%
adroit-hammer147%62%
adroit-hammer841%63%
adroit-relocate135%65%
adroit-relocate864%71%

Real-world hardware experiments

TABLE VII: Real-world closed-loop results on a Crazyflie and an SO-100 arm. Setup: Diffuser (K=64, T=20); KDP (K=64).
Domain Method Succ. (%) TTS (s) Replan (Hz) Planner p50 E2E p50
NavigationDiffuser9212.54.1242312
Navigation KDP9412.038826
ManipulationDiffuser88353311410
Manipulation KDP903217947

Qualitative Results

Simulated benchmark results

Real-world hardware results