CarPlanner: Consistent Auto-regressive Trajectory Planning for Large-scale
Reinforcement Learning in Autonomous Driving
Dongkun Zhang1,2
Jiaming Liang2
Ke Guo2
Sha Lu1
Qi Wang2
Rong Xiong1,B
Zhenwei Miao2,†
Yue Wang1
1
Zhejiang University 2
Cainiao Network
1
{zhangdongkun, lusha, rxiong, ywang24}@zju.edu.cn
2
{liangjiaming.ljm, muguo.gk, ruifeng.wq, zhenwei.mzw}@alibaba-inc.com
Abstract
Trajectory planning is vital for autonomous driving, ensur-
ing safe and efficient navigation in complex environments.
While recent learning-based methods, particularly rein-
forcement learning (RL), have shown promise in specific
scenarios, RL planners struggle with training inefficiencies
and managing large-scale, real-world driving scenarios. In
this paper, we introduce CarPlanner, a Consistent auto-
regressive Planner that uses RL to generate multi-modal
trajectories. The auto-regressive structure enables efficient
large-scale RL training, while the incorporation of consis-
tency ensures stable policy learning by maintaining coher-
ent temporal consistency across time steps. Moreover, Car-
Planner employs a generation-selection framework with an
expert-guided reward function and an invariant-view mod-
ule, simplifying RL training and enhancing policy perfor-
mance. Extensive analysis demonstrates that our proposed
RL framework effectively addresses the challenges of train-
ing efficiency and performance enhancement, positioning
CarPlanner as a promising solution for trajectory planning
in autonomous driving. To the best of our knowledge, we are
the first to demonstrate that the RL-based planner can sur-
pass both IL- and rule-based state-of-the-arts (SOTAs) on
the challenging large-scale real-world dataset nuPlan. Our
proposed CarPlanner surpasses RL-, IL-, and rule-based
SOTA approaches within this demanding dataset.
1. Introduction
Trajectory planning [36] is essential in autonomous driv-
ing, utilizing outputs from perception and trajectory predic-
tion modules to generate future poses for the ego vehicle.
A controller tracks this planned trajectory, producing con-
trol commands for closed-loop driving. Recently, learning-
based trajectory planning has garnered attention due to its
†Project lead. B Corresponding author.
Figure 1. Frameworks for multi-step trajectory generation.
(a) Initialization-refinement that generates an initial trajectory and
refines it iteratively. (b) Vanilla auto-regressive models that decode
subsequent poses sequentially. (c) Our consistent auto-regressive
model that integrates time-consistent mode information.
potential to automate algorithm iteration, eliminate tedious
rule design, and ensure safety and comfort in diverse real-
world scenarios [36].
Most existing researches [11, 16, 29] employ imitation
learning (IL) to align planned trajectories with those of
human experts. However, this approach suffers from dis-
tribution shift [28] and causal confusion [8]. Reinforce-
ment learning (RL) offers a potential solution, addressing
these challenges and providing richer supervision through
reward functions. Although RL shows effectiveness in do-
mains such as games [34], robotics [19], and language mod-
els [25], it still struggles with training inefficiencies and per-
formance issues in the large-scale driving task. To the ex-
arXiv:2502.19908v3
[cs.RO]
24
Mar
2025
tent of our knowledge, no RL methods have yet achieved
competitive results on large-scale open datasets such as nu-
Plan [2], which features diverse real-world scenarios.
Thus, this paper aims to tackle two key challenges in
RL for trajectory planning: 1) training inefficiency and 2)
poor performance. Training inefficiency arises from the fact
that RL typically operates in a model-free setting, necessi-
tating an inefficient simulator running on a CPU to repeat-
edly roll out a policy for data collection. To overcome this
challenge, we propose an efficient model-based approach
utilizing neural networks as transition models. Our method
is optimized for execution on hardware accelerators such as
GPUs, rendering our time cost comparable to that of IL-
based methods.
To apply RL to solve the trajectory planning problem,
we formulate it as a multi-step sequential decision-making
task utilizing a Markov Decision Process (MDP). Existing
methods that generate the trajectory†
in multiple steps gen-
erally fall into two categories: initialization-refinement [17,
20, 33, 45] and auto-regressive models [27, 32, 41, 46].
The first category, illustrated in Fig. 1 (a), involves gen-
erating an initial trajectory estimate and subsequently refin-
ing it through iterative applications of RL. However, recent
studies, including Gen-Drive [18], suggest that it continues
to lag behind SOTA IL and rule-based planners. One no-
table limitation of this approach is its neglect of the tempo-
ral causality inherent in the trajectory planning task. Addi-
tionally, the complexity of direct optimization over high-
dimensional trajectory space can hinder the performance
of RL algorithms. The second category consists of auto-
regressive models, shown in Fig. 1 (b), which generate
the poses of the ego vehicle recurrently using a single-
step policy within a transition model. In this category, ego
poses at all time steps are consolidated to form the overall
planned trajectory. As taking temporal causality into ac-
count, current auto-regressive models allow for interactive
behaviors. However, a common limitation is their reliance
on auto-regressively random sampling from action distri-
butions to generate multi-modal trajectories. This vanilla
auto-regressive procedure may compromise long-term con-
sistency and unnecessarily expand the exploration space in
RL, leading to poor performance.
To address the limitations of auto-regressive models, we
introduce CarPlanner, a Consistent auto-regressive model
designed for efficient, large-scale RL-based Planner train-
ing (see Fig. 1 (c)). The key insight of CarPlanner is its
incorporation of consistent mode representation as condi-
tions for the auto-regressive model. Specifically, we lever-
age a longitudinal-lateral decomposed mode representation,
where the longitudinal mode is a scalar that captures aver-
†In this paper, the term “trajectory” refers to the future poses of the
ego vehicle or traffic agents. To avoid confusion, we use the term “state
(action) sequence” to refer to the “trajectory” in the RL community.
age speeds, and the lateral mode encompasses all possible
routes derived from the current state of the ego vehicle along
with map information. This mode remains constant across
time steps, providing stable and consistent guidance during
policy sampling.
Furthermore, we propose a universal reward function
that suits large-scale and diverse scenarios, eliminating the
need for scenario-specific reward designs. This function
consists of an expert-guided and task-oriented term. The
first term quantifies the displacement error between the ego-
planned trajectory and the expert’s trajectory, which, along
with the consistent mode representation, narrows down the
policy’s exploration space. The second term incorporates
common senses in driving tasks including the avoidance of
collision and adherence to the drivable area. Additionally,
we introduce an Invariant-View Module (IVM) to supply
invariant-view input for policy, with the aim of providing
time-agnostic policy input, easing the feature learning and
embracing generalization. To achieve this, IVM prepro-
cesses state and lateral mode by transforming agent, map,
and route information into the ego’s current coordinate and
by clipping information that is distant from the ego.
To our knowledge, we are the first to demonstrate that
RL-based planner outperforms state-of-the-art (SOTA) IL
and rule-based approaches on the challenging large-scale
nuPlan dataset. In summary, the key contributions of this
paper are highlighted as follows:
• We present CarPlanner, a consistent auto-regressive
planner that trains an RL policy to generate consistent
multi-modal trajectories.
• We introduce an expert-guided universal reward function
and IVM to simplify RL training and improve policy
generalization, leading to enhanced closed-loop perfor-
mance.
• We conduct a rigorous analysis on the characteristics of
IL and RL training, providing insights into their strengths
and limitations, while highlighting the advantages of RL
in tackling challenges such as distribution shift and causal
confusion.
• Our framework showcases exceptional performance, sur-
passing all RL-, IL-, and rule-based SOTAs on the nuPlan
benchmark. This underscores the potential of RL in nav-
igating complex real-world driving scenarios.
2. Related Work
2.1. Imitation-based Planning
The use of IL to train planners based on human demon-
strations has garnered significant interest recently. This
approach leverages the driving expertise of experienced
drivers who can safely and comfortably navigate a wide
range of real-world scenarios, along with the added advan-
tage of easily collectible driving data at scale [2, 9, 15]. Nu-
s0
Non-reactive
Transition Model
Mode Selector
c
Mode Scores
Previewed World s0
Consistent Mode
as Query
s1
a0 a1
...
...
Multi-modal
trajectories
Rule Selector
Ego-planned
trajectory
Trajectory Generator Rule-augmented Selector
NR Transition Model
Auto-regressive Policy
IVM
IVM
s2
Mode Selector
Figure 2. CarPlanner contains four parts. (1) The non-reactive transition model takes initial state s0 as input and predicts the future
trajectories of traffic agents. (2) The mode selector outputs scores based on the initial state and the modes c. (3) The trajectory generator
obeys an auto-regressive structure condition on the consistent mode and produces mode-aligned multi-modal trajectories. (4) The rule-
augmented selector compensates the mode scores by safety, comfort, and progress metrics.
merous studies [5, 17, 29] have focused on developing inno-
vative networks to enhance open-loop performance in this
domain. However, the ultimate challenge of autonomous
driving is achieving closed-loop operation, which is eval-
uated using driving-oriented metrics such as safety, adher-
ence to traffic rules, comfort, and progress. This reveals a
significant gap between the training and testing phases of
planners. Moreover, IL is particularly vulnerable to issues
such as distribution shift [28] and causal confusion [8]. The
first issue results in suboptimal decisions when the system
encounters scenarios that are not represented in the train-
ing data distribution. The second issue arises when net-
works inadvertently capture incorrect correlations and de-
velop shortcuts based on input information, primarily due to
the reliance on imitation loss from expert demonstrations.
Despite efforts in several studies [1, 3, 4, 42] to address
these challenges, the gap between training and testing re-
mains substantial.
2.2. RL in Autonomous Driving
In the field of autonomous driving, RL has demonstrated its
effectiveness in addressing specific scenarios such as high-
way driving [22, 39], lane changes [14, 23], and unprotected
left turns [22, 40]. Most methods directly learn policies
over the control space, which includes throttle, brake, and
steering commands. Due to the high frequency of control
command execution, the simulation process can be time-
consuming, and exploration can be inconsistent [40]. Sev-
eral works [40, 44] have proposed learning trajectory plan-
ners with actions defined as ego-planned trajectories, which
temporally extend the exploration space and improve train-
ing efficiency. However, a trade-off exists between the
trajectory horizon and training performance, as noted in
ASAP-RL [40]. Increasing the trajectory horizon results
in less reactive behaviors and a reduced amount of data,
while a smaller trajectory horizon leads to challenges sim-
ilar to those encountered in control space. Additionally,
these methods typically employ a model-free setting, mak-
ing them difficult to apply to the complex, diverse real-
world scenarios found in large-scale driving datasets. In
this paper, we propose adopting a model-based formulation
that can facilitate RL training on large-scale datasets. Under
this formulation, we aim to overcome the trajectory horizon
trade-off by using a transition model, which can provide a
preview of the world in which our policy can make multi-
step decisions during testing.
3. Method
3.1. Preliminaries
MDP is used to model sequential decision problems, for-
mulated as a tuple ⟨S, A, Pτ , R, ρ0, γ, T⟩. S is the state
space. A is the action space. Pτ : S × A → ∆(S) †
is the
state transition probability. R : S × A → R denotes the
reward function and is bounded. ρ0 ∈ ∆(S) is the initial
state distribution. T is the time horizon and γ is the dis-
count factor of future rewards. The state-action sequence is
defined as τ = (s0, a0, s1, a1, . . . , sT ), where st ∈ S and
at ∈ A are the state and action at time step t. The objective
of RL is to maximize the expected return:
  setlength {abovedisplayskip }{0pt} begin {split} max _{pi } mathbb {E}_{boldsymbol {s}_t sim {P_{tau }}, a_t sim pi } [sum _{t=0}^{T}gamma ^{t} R(boldsymbol {s}_t, a_t)]. label {equation:mdp} end {split} setlength {belowdisplayskip }{0pt}  (1)
Vectorized state representation. State st contains map
and agent information in vectorized representation [10].
Map information m includes the road network, traffic lights,
etc, which are represented by polylines and polygons.
Agent information includes the current and past poses of
ego vehicle and other traffic agents, which are represented
by polylines. The index of ego vehicle is 0 and the indices
of traffic agents range from 1 to N. For each agent i, its
history is denoted as si
t−H:t, i ∈ {0, 1, . . . , N}, where H is
the history time horizon.
†∆(X) denotes the set of probability distribution over set X.
3.2. Problem Formulation
We model the trajectory planning task as a sequential deci-
sion process and decouple the auto-regressive models into
policy and transition models. The key to connect trajec-
tory planning and auto-regressive models is to define the
action as the next pose of ego vehicle, i.e., at = s0
t+1.
Therefore, after forwarding the auto-regressive model, the
decoded pose is collected to be the ego-planned trajectory.
Specifically, we can reduce the state-action sequence to the
state sequence under this definition and vectorized repre-
sentation:
  setlength {abovedisplayskip }{0pt} begin {split} &P(boldsymbol {s}_0, a_0, boldsymbol {s}_1, a_1, dots , boldsymbol {s}_T)  &= P(m, s^{0:N}_{-H:0}, s^{0}_{1}, m, s^{0:N}_{1-H:1}, s^{0}_{2}, dots , m, s^{0:N}_{T-H:T})  &= P(m, s^{0:N}_{-H:0}, m, s^{0:N}_{1-H:1}, dots , m, s^{0:N}_{T-H:T})  &= P(boldsymbol {s}_0, boldsymbol {s}_1, dots , boldsymbol {s}_T). [-0.1cm] end {split} setlength {belowdisplayskip }{0pt} 
(2)
The state sequence can be further formulated in an auto-
regressive fashion and decomposed into policy and transi-
tion model:
  setlength {abovedisplayskip }{0pt} begin {split} &P(boldsymbol {s}_0, boldsymbol {s}_1, dots , boldsymbol {s}_T) = rho _{0}(boldsymbol {s}_0) prod _{t=0}^{T-1} P(boldsymbol {s}_{t+1} | boldsymbol {s}_t)  &= rho _{0}(boldsymbol {s}_0) prod _{t=0}^{T-1} P(s^0_{t+1}, s^{1:N}_{t+1} | boldsymbol {s}_t)  &= rho _{0}(boldsymbol {s}_0) prod _{t=0}^{T-1} underbrace {pi (a_t | boldsymbol {s}_t)}_{text {Policy}} underbrace {P_{tau }(s^{1:N}_{t+1} | boldsymbol {s}_t)}_{text {Transition Model}}. [-0.15cm] end {split} label {equation:incons} setlength {belowdisplayskip }{0pt} 
(3)
From Eq. (3), we can clearly identify the inherent problem
associated with the typical auto-regressive approach: incon-
sistent behaviors across time steps arise from the policy dis-
tribution, which depends on random sampling from the ac-
tion distribution.
To solve the above problem, we introduce consistent
mode information c that remains unchanged across time
steps into the auto-regressive fashion:
  setlength {abovedisplayskip }{0pt} begin {split} &P(boldsymbol {s}_0, boldsymbol {s}_1, dots , boldsymbol {s}_T) = int _{boldsymbol {c}} P(boldsymbol {s}_0, boldsymbol {s}_1, dots , boldsymbol {s}_T, boldsymbol {c}) dboldsymbol {c}  &= rho _0(boldsymbol {s}_0) int _{boldsymbol {c}} P(boldsymbol {c} | boldsymbol {s}_0) P(boldsymbol {s}_1, dots , boldsymbol {s}_T | boldsymbol {c}) dboldsymbol {c}  &= rho _0(boldsymbol {s}_0) prod _{t=0}^{T-1} underbrace {P_{tau }(s^{1:N}_{t+1} | boldsymbol {s}_t)}_{text {Transition Model}} int _{boldsymbol {c}} underbrace {P(boldsymbol {c} | boldsymbol {s}_0)}_{text {Mode Selector}} prod _{t=0}^{T-1} underbrace {pi (a_t | boldsymbol {s}_t, boldsymbol {c})}_{text {Policy}} dboldsymbol {c}. [-1.cm] label {equation:car} end {split} setlength {belowdisplayskip }{0pt} 
(4)
Since we focus on the ego trajectory planning, the consis-
tent mode c does not impact transition model.
This consistent auto-regressive formulation defined in
Eq. (4) reveals a generation-selection framework where the
mode selector scores each mode based on the initial state s0
and the trajectory generator generates multi-modal trajecto-
ries via sampling from the mode-conditioned policy.
Non-reactive transition model. The transition model for-
mulated in Eq. (4) needs to be employed in every time step
since it produces the poses of traffic agents at time step
t + 1 based on current state st. In practice, this process
is time-consuming and we do not observe a performance
improvement by using this transition model, therefore, we
use trajectory predictors P(s1:N
1:T |s0) as non-reactive transi-
tion model that produces all future poses of traffic agents in
one shot given initial state s0.
3.3. Planner Architecture
The framework of our proposed CarPlanner is illustrated
in Fig. 2, comprising four key components: 1) the non-
reactive transition model, 2) the mode selector, 3) the tra-
jectory generator, and 4) the rule-augmented selector.
Our planner operates within a generation-selection
framework. Given an initial state s0 and all possible Nmode
modes, the trajectory selector evaluates and assigns scores
to each mode. The trajectory generator then produces Nmode
trajectories that correspond to their respective modes. For
trajectory generator, the initial state s0 is replicated Nmode
times, each associated with one of the Nmode modes, effec-
tively creating Nmode parallel worlds. The policy is executed
within these previewed worlds. During the policy rollout, a
trajectory predictor acts as the state transition model, gener-
ating future poses of traffic agents across all time horizons.
3.3.1. Non-reactive Transition Model
This module takes the initial state s0 as input and outputs
the future trajectories of traffic agents. The initial state is
processed by agent and map encoders, followed by a self-
attention Transformer encoder [38] to fuse the agent and
map features. The agent features are then decoded into fu-
ture trajectories.
Agent and map encoders. The state s0 contains both map
and agent information. The map information m consists of
Nm,1 polylines and Nm,2 polygons. The polylines describe
lane centers and lane boundaries, with each polyline con-
taining 3Np points, where 3 corresponds to the lane center,
the left boundary, and the right boundary. Each point is with
dimension Dm = 9 and includes the following attributes: x,
y, heading, speed limit, and category. When concatenated,
the points of the left and right boundaries together with
the center point yield a dimension of Nm,1 × Np × 3Dm.
We leverage a PointNet [26] to extract features from the
points of each polyline, resulting in a dimensionality of
Nm,1 × D, where D represents the feature dimension. The
polygons represent intersections, crosswalks, stop lines, etc,
with each polygon containing Np points. We utilize another
PointNet to extract features from the points of each poly-
gon, producing a dimension of Nm,2 × D. We then con-
catenate the features from both polylines and polygons to
form the overall map features, resulting in a dimension of
Nm × D. The agent information A consists of N agents,
where each agent maintains poses for the past H time steps.
Each pose is with dimension Da = 10 and includes the
following attributes: x, y, heading, velocity, bounding box,
time step, and category. Consequently, the agent informa-
tion has a dimension of N × H × Da. We apply another
PointNet to extract features from the poses of each agent,
yielding an agent feature dimension of N × D.
3.3.2. Mode Selector
This module takes  s_0  and longitudinal-lateral decomposed
mode information as input and outputs the probability of
each mode. The number of modes Nmode = NlatNlon.
Route-speed decomposed mode. To capture the longitu-
dinal behaviors, we generate  N_{text {lon}}  modes that represent the
average speed of the trajectory associated with each mode.
Each longitudinal mode  c_{text {lon},j}  is defined as a scalar value of
 frac {j}{N_{text {lon}}} 
, repeated across a dimension  D . As a result, the dimen-
sionality of the longitudinal modes is  N_{text {lon}} times D . For lateral
behaviors, we identify  N_{text {lat}}  possible routes from the map
using a graph search algorithm. These routes correspond to
the lanes available for the ego vehicle. The dimensional-
ity of these routes is  N_{text {lat}} times N_r times D_m . We employ another
PointNet to aggregate the features of the  N_r  points along
each route, producing a lateral mode with a dimension of
 N_{text {lat}} times D . To create a comprehensive mode representation
c, we combine the lateral and longitudinal modes, resulting
in a combined dimension of  N_{text {lat}} times N_{text {lon}} times 2D . To align this
mode information with other feature dimensions, we pass it
through a linear layer, mapping it back to  N_{text {lat}} times N_{text {lon}} times D .
Query-based Transformer decoder. This decoder is em-
ployed to fuse the mode features with map and agent fea-
tures derived from  s_0 . In this framework, the mode serves
as the query, while the map and agent information act as the
keys and values. The updated mode features are decoded
through a multi-layer perceptron (MLP) to yield the scores
for each mode, which are subsequently normalized using
the softmax operator.
3.3.3. Trajectory Generator
This module operates in an auto-regressive manner, recur-
rently decoding the next pose of the ego vehicle at, given
the current state st, and consistent mode information c.
Invariant-view module (IVM). Before feeding the mode
and state into the network, we preprocess them to eliminate
time information. For the map and agent information in
state st, we select the K-nearest neighbors (KNN) to the
ego current pose and only feed these into the policy. K
is set to the half of map and agent elements respectively.
Regarding the routes that capture lateral behaviors, we filter
out the segments where the point closest to the current pose
of the ego vehicle is the starting point, retaining Kr points.
In this case, Kr is set to a quarter of Nr points in one route.
Finally, we transform the routes, agent, and map poses into
the coordinate frame of the ego vehicle at the current time
step t. We subtract the historical time steps t − H : t from
the current time step t, yielding time steps in range −H : 0.
Query-based Transformer decoder. We employ the same
backbone network architecture as the mode selector, but
with different query dimensions. Due to the IVM and the
fact that different modes yield distinct states, the map and
agent information cannot be shared among modes. As a re-
sult, we fuse information for each individual mode. Specifi-
cally, the query dimension is 1 times D, while the dimensions of
the keys and values are (N + N_{m}) times D. The output feature
dimension remains 1 times D. Note that Transformer decoder
can process information from multiple modes in parallel,
eliminating the need to handle each mode sequentially.
Policy output. The mode feature is processed by two dis-
tinct heads: a policy head and a value head. Each head com-
prises its own MLP to produce the parameters for the action
distribution and the corresponding value estimate. We em-
ploy a Gaussian distribution to model the action distribu-
tion, where actions are sampled from this distribution dur-
ing training. In contrast, during inference, we utilize the
mean of the distribution to determine the actions.
3.3.4. Rule-augmented Selector
This module is only utilized during inference and takes as
input the initial state  s_0 , the multi-modal ego-planned tra-
jectories, and the predicted future trajectories of agents. It
calculates driving-oriented metrics such as safety, progress,
comfort. A comprehensive score is obtained by the
weighted sum of rule-based scores and the mode scores pro-
vided by the mode selector. The ego-planned trajectory with
the highest score is selected as the output of the planner.
3.4. Training
We first train the non-reactive transition model and freeze
the weights during the training of the mode selector and tra-
jectory generator. Instead of feeding all modes to the gen-
erator, we apply a winner-takes-all strategy, wherein a posi-
tive mode is assigned based on the ego ground-truth trajec-
tory and serves as a condition for the trajectory generator.
Mode assignment. For the lateral mode, we assign the
route closest to the endpoint of ego ground-truth trajectory
as the positive lateral mode. For the longitudinal mode, we
partition the longitudinal space into N_{text {lon}} intervals and as-
sign the interval containing the endpoint of the ground-truth
trajectory as the positive longitudinal mode.
Reward function. To handle diverse scenarios, we use the
negative displacement error (DE) between the ego future
pose and the ground truth as a universal reward. We also in-
troduce additional terms to improve trajectory quality: col-
lision rate and drivable area compliance. If the future pose
collides or falls outside the drivable area, the reward is set
to -1; otherwise, it is 0.
Mode dropout. In some cases, there are no available routes
for ego to follow. However, since routes serve as queries in
Transformer, the absence of a route can lead to unstable or
hazardous outputs. To mitigate this issue, we implement a
mode dropout module during training that randomly masks
routes to prevent over-reliance on this information.
Loss function. For the selector, we use cross-entropy loss
that is the negative log-likelihood of the positive mode and a
side task that regresses the ego ground-truth trajectory. For
the generator, we use PPO [31] loss that consists of three
parts: policy improvement, value estimation, and entropy.
Full description can be found in supplementary.
4. Experiments
4.1. Experimental Setup
Dataset and simulator. We use nuPlan [2], a large-scale
closed-loop platform for studying trajectory planning in au-
tonomous driving, to evaluate the efficacy of our method.
The nuPlan dataset contains driving log data over 1,500
hours collected by human expert drivers across 4 diverse
cities. It includes complex, diverse scenarios such as lane
follow and change, left and right turn, traversing intersec-
tions and bus stops, roundabouts, interaction with pedestri-
ans, etc. As a closed-loop platform, nuPlan provides a sim-
ulator that uses scenarios from the dataset as initialization.
During the simulation, traffic agents are taken over by log-
replay (non-reactive) or an IDM [37] policy (reactive). The
ego vehicle is taken over by user-provided planners. The
simulator lasts for 15 seconds and runs at 10 Hz. At each
timestamp, the simulator queries the planner to plan a tra-
jectory, which is tracked by an LQR controller to generate
control commands to drive the ego vehicle.
Benchmarks and metrics. We use two benchmarks:
Test14-Random and Reduced-Val14 for comparing with
other methods and analyzing the design choices within
our method. The Test14-Random provided by PlanTF [4]
contains 261 scenarios. The Reduced-Val14 provided by
PDM [7] contains 318 scenarios.
We use the closed-loop score (CLS) provided by the of-
ficial nuPlan devkit†
to assess the performance of all meth-
ods. The CLS score comprehends different aspects such as
safety (S-CR, S-TTC), drivable area compliance (S-Area),
progress (S-PR), comfort, etc. Based on the different be-
havior types of traffic agents, CLS is detailed into CLS-NR
(non-reactive) and CLS-R (reactive).
Implementation details. We follow PDM [7] to construct
our training and validation splits. The size of the training
set is 176,218 where all available scenario types are used,
with a number of 4,000 scenarios per type. The size of the
validation set is 1,118 where 100 scenarios with 14 types are
selected. We train all models with 50 epochs in 2 NVIDIA
3090 GPUs. The batch size is 64 per GPU. We use AdamW
optimizer with an initial learning rate of 1e-4 and reduce the
learning rate when the validation loss stops decreasing with
a patience of 0 and decrease factor of 0.3. For RL training,
†https://github.com/motional/nuplan-devkit
Type Planners CLS-NR CLS-R
Rule IDM [37] 70.39 72.42
PDM-Closed [7] 90.05 91.64
IL
RasterModel [2] 69.66 67.54
UrbanDriver [29] 63.27 61.02
GC-PGP [12] 55.99 51.39
PDM-Open [7] 52.80 57.23
GameFormer [17] 80.80 79.31
PlanTF [4] 86.48 80.59
PEP [42] 91.45 89.74
PLUTO [3] 91.92 90.03
RL CarPlanner (Ours) 94.07 91.1
Table 1. Comparison with SOTAs in Test14-Random. Based on
the type of trajectory generator, all methods are categorized into
Rule, IL, and RL. The best result is in bold and the second best
result is underlined.
Type Planners CLS-NR S-CR S-PR
Rule PDM-Closed [7] 91.21 97.01 92.68
IL
GameFormer [17] 83.76 94.73 88.12
PlanTF [4] 83.66 94.02 92.67
Gen-Drive (Pretrain) [18] 85.12 93.65 86.64
RL
Gen-Drive (Finetune) [18] 87.53 95.72 89.94
CarPlanner (Ours) 91.45 96.38 95.37
Table 2. Comparison with SOTAs in Reduced-Val14 with non-
reactive traffic agents.
we set the discount γ = 0.1 and the GAE parameter λ =
0.9. The weights of value, policy, and entropy loss are set to
3, 100, and 0.001, respectively. The number of longitudinal
modes is set to 12 and a maximum number of lateral modes
are set to 5.
4.2. Comparison with SOTAs
SOTAs. We categorize the methods into Rule, IL, and RL
based on the type of trajectory generator. (1) PDM [7]
wins the nuPlan challenge 2023, its IL-based and rule-based
variants are denoted as PDM-Open and PDM-Closed, re-
spectively. PDM-Closed follows the generation-selection
framework where IDM is used to generate multiple candi-
date trajectories and rule-based selector considering safety,
progress, and comfort is used to select the best trajectory.
(2) PLUTO [3] also obeys the generation-selection frame-
work and uses contrastive IL to incorporate various data
augmentation techniques and trains the generator. (3) Gen-
Drive [18] is a concurrent work that follows a pretrain-
finetune pipeline where IL is used to pretrain a diffusion-
based planner and RL is used to finetune the denoising pro-
cess based on a reward model trained by AI preference.
Results. We compare our method with SOTAs in Test14-
Random and Reduced-Val14 benchmark as shown in Tab. 1
and Tab. 2. Overall, our CarPlanner demonstrates superior
performance, particularly in non-reactive environments.
In the non-reactive setting, our method achieves the
highest scores across all metrics, with an improvement of
4.02 and 2.15 compared to PDM-Closed and PLUTO, es-
tablishing the potential of RL and the superior performance
Design Choices Closed-loop metrics (↑) Open-loop losses (↓)
Reward
DE
Reward
Quality
Coord
Trans
KNN CLS-NR S-CR S-Area S-PR S-Comfort
Loss
Selector
Loss
Generator
✗ ✓ ✓ ✓ 31.79 95.74 98.45 33.10 48.84 1.03 30.3
✓ ✗ ✓ ✓ 90.44 97.49 96.91 93.33 90.73 0.99 1221.6
✓ ✓ ✗ ✓ 90.78 96.92 98.46 91.37 94.23 1.00 2130.7
✓ ✓ ✓ ✗ 92.73 98.07 98.46 94.69 93.44 1.03 2083.6
✓ ✓ ✓ ✓ 94.07 99.22 99.22 95.06 91.09 1.03 1624.5
Table 3. Ablation studies on the design choices in RL training. Results are in Test14-random non-reactive benchmark.
of our proposed framework. Moreover, CarPlanner reveals
substantial improvement in the progress metric S-PR com-
pared to PDM-Closed in Tab. 2 and comparable collision
metric S-CR, indicating the ability of our method to im-
proving driving efficiency while maintaining safe driving.
Importantly, we do not apply any techniques commonly
used in IL such as data augmentation [3, 4] and ego-history
masking [11], underscoring the intrinsic capability of our
approach to solving the closed-loop task.
In the reactive setting, while our method performs well,
it falls slightly short of PDM-Closed. This discrepancy
arises because our model was trained exclusively in non-
reactive settings and has not interacted with the IDM policy
used by reactive settings; as a result, our model is less robust
to disturbances generated by reactive agents during testing.
4.3. Ablation Studies
We investigate the effects of different design choices in RL
training. The results are shown in Tab. 3.
Influence of reward items. When using the quality re-
ward only, the planner tends to generate static trajectories
and achieves a low progress metric. This occurs because
the ego vehicle begins in a safe, drivable state, but mov-
ing forward is at risk of collisions or leaving the drivable
area. On the other hand, when the quality reward is incor-
porated alongside the DE reward, it leads to significant im-
provements in closed-loop metrics compared to using the
DE reward alone. For instance, the S-CR metric rises from
97.49 to 99.22, and the S-Area metric rises from 96.91 to
99.22. These improvements indicate that the quality reward
encourages safe and comfortable behaviors.
Effectiveness of IVM. The results show that the coordi-
nate transformation and KNN techniques in IVM notably
improve closed-loop metrics and generator loss. For in-
stance, with the coordinate transformation technique, the
overall closed-loop score increases from 90.78 to 94.07, and
S-PR rises from 91.37 to 95.06. These improvements are at-
tributed to the enhanced accuracy of value estimation in RL,
leading to generalized driving in closed-loop.
4.4. Extention to IL
In addition to designing for RL training, we also extend the
CarPlanner to incorporate IL. We conduct rigorous analysis
to compare the effects of various design choices in IL and
RL training, as summarized in Tab. 4. Our findings indicate
that while mode dropout and selector side task contribute to
both IL and RL training, ego-history dropout and backbone
sharing, often effective in IL, are less suitable for RL.
Ego-history dropout. Previous works [1, 3, 4, 11] sug-
gest that planners trained via IL may rely too heavily on
past poses and neglect environmental state information. To
counter this, we combine techniques from ChauffeurNet [1]
and PlanTF [4] into an ego-history dropout module, ran-
domly masking ego history poses and current velocity to
alleviate the causal confusion issue.
Our experiments confirm that ego-history dropout
benifits IL training, as it improves performance across
closed-loop metrics like S-CR and S-Area. However, in RL
training, we observe a negative impact on advantage esti-
mation due to ego-history dropout, which significantly af-
fects the value part of generator loss, leading to closed-loop
performance degradation. This suggests that RL training
naturally addresses the causal confusion problem inherent
in IL by uncovering causal relationships that align with the
reward signal, which explicitly encodes task-oriented pref-
erences. This capability highlights the potential of RL to
push the boundaries of learning-based planning.
Backbone sharing. This choice, often used in IL-based
multi-modal planners, promotes feature sharing across tasks
to improve generalization. While backbone sharing helps
IL by balancing losses across trajectory generator and se-
lector, we find it adversely affects RL training. Specifically,
backbone sharing leads to higher losses for both the trajec-
tory generator and selector in RL, indicating that gradients
from each task interfere. The divergent objectives in RL
for trajectory generation and selection tasks seem to con-
flict, reducing overall policy performance. Consequently,
we avoid backbone sharing in our RL framework to main-
tain task-specific gradient flow and improve policy quality.
4.5. Qualitative Results
We provide qualitative results as shown in Fig. 3. In
this scenario, ego vehicle is required to execute a right
turn while navigating around pedestrians. In this case,
Our method shows a smooth, efficient performance. From
tsim = 0s to tsim = 9s, all methods wait for the pedestrians
to cross the road. At tsim = 10s, an unexpected pedestrian
goes back and prepares to re-cross the road. PDM-Closed is
Design Choices Closed-loop metrics (↑) Open-loop losses (↓)
Loss
Type
Mode
Dropout
Selector
Side Task
Ego-history
Dropout
Backbone
Sharing
CLS-NR S-CR S-Area S-PR S-Comfort
Loss
Selector
Loss
Generator
IL
✗ ✗ ✗ ✗ 90.82 97.29 98.45 92.15 94.57 1.04 147.5
✓ ✗ ✗ ✗ 91.21 96.54 98.46 91.44 96.92 1.07 153.0
✓ ✓ ✗ ✗ 91.51 96.91 98.46 95.30 96.91 1.04 162.3
✓ ✓ ✓ ✗ 92.72 98.06 98.84 94.88 95.35 1.04 167.5
✓ ✓ ✓ ✓ 93.41 98.85 98.85 93.87 96.15 1.04 174.3
✗ ✗ ✗ ✗ 91.67 98.84 98.84 91.69 90.73 1.04 1812.6
✓ ✗ ✗ ✗ 93.46 98.07 99.61 94.26 92.28 1.09 2254.6
✓ ✓ ✗ ✗ 94.07 99.22 99.22 95.06 91.09 1.03 1624.5
RL
✓ ✓ ✓ ✗ 89.51 97.27 98.44 90.93 83.20 1.05 5424.3
✓ ✓ ✓ ✓ 88.66 95.54 98.84 92.82 86.05 1.21 1928.1
Table 4. Effect of different components on IL and RL loss using our CarPlanner. Results are in Test14-random non-reactive benchmark.
Metric Score
CLS-NR 47.40
S-CR 100.0
S-TTC 0.0
S-Area 100.0
S-PR 31.68
Metric Score
CLS-NR 47.50
S-CR 100.0
S-TTC 0.0
S-Area 100.0
S-PR 31.98
Metric Score
CLS-NR 95.14
S-CR 100.0
S-TTC 100.0
S-Area 100.0
S-PR 85.07
tsim = 11s tsim = 12s
tsim = 10s
tsim = 9s
tsim = 0s
CarPlanner-IL
(Ours)
CarPlanner
(Ours)
PDM-Closed
Figure 3. Qualitative comparison of PDM-Closed and our method in non-reactive environments. The scenario is annotated as
waiting for pedestrian to cross. In each frame shot, ego vehicle is marked as green. Traffic agents are marked as sky blue.
Lineplot with blue is the ego planned trajectory.
unaware of this situation and takes an emergency stop, but it
still intersects with this pedestrian. In contrast, our IL vari-
ant displays an awareness of the pedestrian’s movements
and consequently conducts a braking maneuver. However, it
still remains close to the pedestrian. Our RL method avoids
this hazard by starting up early up to tsim = 9s and achieves
the highest progress and safety metrics.
5. Conclusion
In this paper, we introduce CarPlanner, a consistent
auto-regressive planner aiming at large-scale RL training.
Thanks to the proposed framework, we train an RL-based
planner that outperforms existing RL-, IL-, and rule-based
SOTAs. Furthermore, we provide analysis indicating the
characteristics of IL and RL, highlighting the potential of
RL to take a further step toward learning-based planning.
Limitations and future work. RL needs delicate design
and is prone to input representation. RL can overfit its train-
ing environment and suffer from performance drop in un-
seen environments [21]. Our method leverages expert-aided
reward design to guide exploration. However, this approach
may constrain the full potential of RL, as it inherently relies
on expert demonstrations and may hinder the discovery of
solutions that surpass human expertise. Future work aims to
develop robust RL algorithms capable of overcoming these
limitations, enabling autonomous exploration and general-
ization across diverse environments.
6. Acknowledgement
Many thanks to Jingke Wang for helpful discussions and
all reviewers for improving the paper. This work was sup-
ported by Zhejiang Provincial Natural Science Foundation
of China under Grant No. LD24F030001, and by the Na-
tional Nature Science Foundation of China under Grant
62373322.
References
[1] Mayank Bansal, Alex Krizhevsky, and Abhijit Ogale. Chauf-
feurnet: Learning to drive by imitating the best and synthe-
sizing the worst. In Proc. of Robotics: Science and Systems,
2019. 3, 7
[2] Holger Caesar, Juraj Kabzan, Kok Seang Tan, Whye Kit
Fong, Eric Wolff, Alex Lang, Luke Fletcher, Oscar Beijbom,
and Sammy Omari. nuplan: A closed-loop ml-based plan-
ning benchmark for autonomous vehicles. arXiv preprint
arXiv:2106.11810, 2021. 2, 6, 12
[3] Jie Cheng, Yingbing Chen, and Qifeng Chen. Pluto: Push-
ing the limit of imitation learning-based planning for au-
tonomous driving. arXiv preprint arXiv:2404.14327, 2024.
3, 6, 7, 12, 13
[4] Jie Cheng, Yingbing Chen, Xiaodong Mei, Bowen Yang, Bo
Li, and Ming Liu. Rethinking imitation-based planners for
autonomous driving. In Proc. of the IEEE Intl. Conf. on
Robotics & Automation, pages 14123–14130. IEEE, 2024.
3, 6, 7
[5] Kashyap Chitta, Aditya Prakash, Bernhard Jaeger, Zehao Yu,
Katrin Renz, and Andreas Geiger. Transfuser: Imitation
with transformer-based sensor fusion for autonomous driv-
ing. IEEE Trans. on Pattern Analysis and Machine Intelli-
gence (TPAMI), 2022. 3
[6] Ignasi Clavera, Yao Fu, and Pieter Abbeel. Model-
augmented actor-critic: Backpropagating through paths. In
Proc. of the Int. Conf. on Learning Representations, 2020.
15
[7] Daniel Dauner, Marcel Hallgarten, Andreas Geiger, and
Kashyap Chitta. Parting with misconceptions about learning-
based vehicle motion planning. In Proc. of the Conf. on
Robot Learning, 2023. 6
[8] Pim De Haan, Dinesh Jayaraman, and Sergey Levine. Causal
confusion in imitation learning. Proc. of the Advances in
Neural Information Processing Systems, 32, 2019. 1, 3
[9] Scott Ettinger, Shuyang Cheng, Benjamin Caine, Chenxi
Liu, Hang Zhao, Sabeek Pradhan, Yuning Chai, Ben Sapp,
Charles R Qi, Yin Zhou, et al. Large scale interactive motion
forecasting for autonomous driving: The waymo open mo-
tion dataset. In Proc. of the IEEE/CVF Intl. Conf. on Com-
puter Vision, pages 9710–9719, 2021. 2, 12
[10] Jiyang Gao, Chen Sun, Hang Zhao, Yi Shen, Dragomir
Anguelov, Congcong Li, and Cordelia Schmid. Vectornet:
Encoding hd maps and agent dynamics from vectorized rep-
resentation. In Proc. of the IEEE/CVF Conf. on Computer
Vision and Pattern Recognition, pages 11525–11533, 2020.
3
[11] Ke Guo, Wei Jing, Junbo Chen, and Jia Pan. CCIL: Context-
conditioned imitation learning for urban driving. In Proc. of
Robotics: Science and Systems, 2023. 1, 7
[12] Marcel Hallgarten, Martin Stoll, and Andreas Zell. From
prediction to planning with goal conditioned lane graph
traversals. arXiv preprint arXiv:2302.07753, 2023. 6
[13] Nicklas A Hansen, Hao Su, and Xiaolong Wang. Tem-
poral difference learning for model predictive control. In
Proc. of the Intl. Conf. on Machine Learning, pages 8387–
8406. PMLR, 2022. 15
[14] Xiangkun He, Haohan Yang, Zhongxu Hu, and Chen Lv.
Robust lane change decision making for autonomous vehi-
cles: An observation adversarial reinforcement learning ap-
proach. IEEE Transactions on Intelligent Vehicles, 8(1):184–
193, 2022. 3
[15] John Houston, Guido Zuidhof, Luca Bergamini, Yawei
Ye, Long Chen, Ashesh Jain, Sammy Omari, Vladimir
Iglovikov, and Peter Ondruska. One thousand and one
hours: Self-driving motion prediction dataset. In Proc. of
the Conf. on Robot Learning, pages 409–418, 2021. 2
[16] Yihan Hu, Jiazhi Yang, Li Chen, Keyu Li, Chonghao Sima,
Xizhou Zhu, Siqi Chai, Senyao Du, Tianwei Lin, Wenhai
Wang, et al. Planning-oriented autonomous driving. In
Proc. of the IEEE/CVF Conf. on Computer Vision and Pat-
tern Recognition, pages 17853–17862, 2023. 1
[17] Zhiyu Huang, Haochen Liu, and Chen Lv. Gameformer:
Game-theoretic modeling and learning of transformer-based
interactive prediction and planning for autonomous driving.
In Proc. of the IEEE/CVF Intl. Conf. on Computer Vision,
pages 3903–3913, 2023. 2, 3, 6, 13
[18] Zhiyu Huang, Xinshuo Weng, Maximilian Igl, Yuxiao Chen,
Yulong Cao, Boris Ivanovic, Marco Pavone, and Chen Lv.
Gen-drive: Enhancing diffusion generative driving poli-
cies with reward modeling and reinforcement learning fine-
tuning. arXiv preprint arXiv:2410.05582, 2024. 2, 6, 12
[19] Julian Ibarz, Jie Tan, Chelsea Finn, Mrinal Kalakrishnan,
Peter Pastor, and Sergey Levine. How to train your robot
with deep reinforcement learning: lessons we have learned.
Intl. Journal of Robotics Research, 40(4-5):698–721, 2021.
1
[20] Chiyu Jiang, Andre Cornman, Cheolho Park, Benjamin
Sapp, Yin Zhou, Dragomir Anguelov, et al. Motiondiffuser:
Controllable multi-agent motion prediction using diffusion.
In Proc. of the IEEE/CVF Conf. on Computer Vision and Pat-
tern Recognition, pages 9644–9653, 2023. 2
[21] Robert Kirk, Amy Zhang, Edward Grefenstette, and Tim
Rocktäschel. A survey of zero-shot generalisation in deep
reinforcement learning. Journal of Artificial Intelligence Re-
search, 76:201–264, 2023. 8
[22] Edouard Leurent and Jean Mercat. Social attention for au-
tonomous decision-making in dense traffic. arXiv preprint
arXiv:1911.12250, 2019. 3
[23] Guofa Li, Yifan Yang, Shen Li, Xingda Qu, Nengchao Lyu,
and Shengbo Eben Li. Decision making of autonomous ve-
hicles in lane change scenarios: Deep reinforcement learn-
ing approaches with risk awareness. Transportation research
part C: emerging technologies, 134:103452, 2022. 3
[24] Quanyi Li, Zhenghao Mark Peng, Lan Feng, Zhizheng Liu,
Chenda Duan, Wenjie Mo, and Bolei Zhou. Scenarionet:
Open-source platform for large-scale traffic scenario simula-
tion and modeling. Proc. of the Advances in Neural Infor-
mation Processing Systems, 36:3894–3920, 2023. 12
[25] Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Car-
roll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini
Agarwal, Katarina Slama, Alex Ray, et al. Training language
models to follow instructions with human feedback. Proc. of
the Advances in Neural Information Processing Systems, 35:
27730–27744, 2022. 1
[26] Charles R Qi, Hao Su, Kaichun Mo, and Leonidas J Guibas.
Pointnet: Deep learning on point sets for 3d classification
and segmentation. In Proc. of the IEEE/CVF Conf. on Com-
puter Vision and Pattern Recognition, pages 652–660, 2017.
4
[27] Nicholas Rhinehart, Rowan McAllister, Kris Kitani, and
Sergey Levine. Precog: Prediction conditioned on goals
in visual multi-agent settings. In Proc. of the IEEE/CVF
Intl. Conf. on Computer Vision, pages 2821–2830, 2019. 2
[28] Stéphane Ross, Geoffrey Gordon, and Drew Bagnell. A re-
duction of imitation learning and structured prediction to no-
regret online learning. In Proceedings of the fourteenth inter-
national conference on artificial intelligence and statistics,
pages 627–635, 2011. 1, 3
[29] Oliver Scheel, Luca Bergamini, Maciej Wolczyk, Błażej
Osiński, and Peter Ondruska. Urban driver: Learning to
drive from real-world demonstrations using policy gradients.
In Proc. of the Conf. on Robot Learning, pages 718–728,
2022. 1, 3, 6
[30] John Schulman, Philipp Moritz, Sergey Levine, Michael Jor-
dan, and Pieter Abbeel. High-dimensional continuous con-
trol using generalized advantage estimation. arXiv preprint
arXiv:1506.02438, 2015. 11, 12
[31] John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Rad-
ford, and Oleg Klimov. Proximal policy optimization algo-
rithms. arXiv preprint arXiv:1707.06347, 2017. 6, 11
[32] Ari Seff, Brian Cera, Dian Chen, Mason Ng, Aurick Zhou,
Nigamaa Nayakanti, Khaled S Refaat, Rami Al-Rfou, and
Benjamin Sapp. Motionlm: Multi-agent motion forecast-
ing as language modeling. In Proc. of the IEEE/CVF
Intl. Conf. on Computer Vision, pages 8579–8590, 2023. 2
[33] Shaoshuai Shi, Li Jiang, Dengxin Dai, and Bernt Schiele.
Motion transformer with global intention localization and lo-
cal movement refinement. Proc. of the Advances in Neural
Information Processing Systems, 35:6531–6543, 2022. 2
[34] David Silver, Aja Huang, Chris J Maddison, Arthur Guez,
Laurent Sifre, George Van Den Driessche, Julian Schrit-
twieser, Ioannis Antonoglou, Veda Panneershelvam, Marc
Lanctot, et al. Mastering the game of go with deep neural
networks and tree search. nature, 529(7587):484–489, 2016.
1
[35] Simon Suo, Sebastian Regalado, Sergio Casas, and Raquel
Urtasun. Trafficsim: Learning to simulate realistic multi-
agent behaviors. In Proc. of the IEEE/CVF Conf. on Com-
puter Vision and Pattern Recognition, pages 10400–10409,
2021. 13
[36] Ardi Tampuu, Tambet Matiisen, Maksym Semikin, Dmytro
Fishman, and Naveed Muhammad. A survey of end-to-end
driving: Architectures and training methods. IEEE Trans-
actions on Neural Networks and Learning Systems, 33(4):
1364–1384, 2020. 1
[37] Martin Treiber, Ansgar Hennecke, and Dirk Helbing. Con-
gested traffic states in empirical observations and micro-
scopic simulations. Physical review E, 62(2):1805, 2000.
6
[38] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszko-
reit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia
Polosukhin. Attention is all you need. Proc. of the Advances
in Neural Information Processing Systems, 30, 2017. 4
[39] Huanjie Wang, Shihua Yuan, Mengyu Guo, Xueyuan Li, and
Wei Lan. A deep reinforcement learning-based approach
for autonomous driving in highway on-ramp merge. Pro-
ceedings of the Institution of Mechanical engineers, Part D:
Journal of Automobile engineering, 235(10-11):2726–2739,
2021. 3
[40] Letian Wang, Jie Liu, Hao Shao, Wenshuo Wang, Ruobing
Chen, Yu Liu, and Steven L Waslander. Efficient Reinforce-
ment Learning for Autonomous Driving with Parameterized
Skills and Priors. In Proc. of Robotics: Science and Systems,
Daegu, Republic of Korea, 2023. 3
[41] Wei Wu, Xiaoxin Feng, Ziyan Gao, and Yuheng Kan. Smart:
Scalable multi-agent real-time simulation via next-token pre-
diction. arXiv preprint arXiv:2405.15677, 2024. 2
[42] Dongkun Zhang, Jiaming Liang, Sha Lu, Ke Guo, Qi Wang,
Rong Xiong, Zhenwei Miao, and Yue Wang. Pep: Policy-
embedded trajectory planning for autonomous driving. IEEE
Robotics and Automation Letters, 2024. 3, 6
[43] Zhejun Zhang, Alexander Liniger, Christos Sakaridis, Fisher
Yu, and Luc V Gool. Real-time motion prediction via het-
erogeneous polyline transformer with relative pose encoding.
Proc. of the Advances in Neural Information Processing Sys-
tems, 36, 2024. 13
[44] Tong Zhou, Letian Wang, Ruobing Chen, Wenshuo Wang,
and Yu Liu. Accelerating reinforcement learning for au-
tonomous driving using task-agnostic and ego-centric mo-
tion skills. In Proc. of the IEEE/RSJ Intl. Conf. on Intelligent
Robots and Systems, pages 11289–11296. IEEE, 2023. 3
[45] Yang Zhou, Hao Shao, Letian Wang, Steven L Waslander,
Hongsheng Li, and Yu Liu. Smartrefine: A scenario-adaptive
refinement framework for efficient motion prediction. In
Proc. of the IEEE/CVF Conf. on Computer Vision and Pat-
tern Recognition, pages 15281–15290, 2024. 2
[46] Zikang Zhou, Haibo Hu, Xinhong Chen, Jianping Wang,
Nan Guan, Kui Wu, Yung-Hui Li, Yu-Kai Huang, and
Chun Jason Xue. Behaviorgpt: Smart agent simulation
for autonomous driving with next-patch prediction. arXiv
preprint arXiv:2405.17372, 2024. 2
CarPlanner: Consistent Auto-regressive Trajectory Planning for Large-scale
Reinforcement Learning in Autonomous Driving
Supplementary Material
A. Training Procedure
Algorithm 1 outlines the training process for the CarPlanner
framework. Notably, during the training of the trajectory
generator, we have the flexibility to employ either RL or IL,
but, in this work, we do not combine RL and IL simultane-
ously, opting instead to explore their distinct characteristics
separately. The definitions of the loss functions are given in
the following.
Loss of non-reactive transition model. The non-reactive
transition model β is trained to simulate agent trajecto-
ries based on the initial state s0. For each data sample
(s0, s1:N,gt
1:T ) ∈ D, the model predicts trajectories s1:N
1:T =
β(s0), and the training objective minimizes the L1 loss:
  L_{text {tm}} = frac {1}{T} sum _{t=1}^T sum _{n=1}^N left | s_t^n - s_t^{n,text {gt}} right |_1.  (5)
Mode selector loss. This contains two parts: cross-entropy
and side-task loss. The cross-entropy loss is defined as:
  text {CrossEntropyLoss}(boldsymbol {sigma }, c^*) = -sum _{i=1}^{N_{text {mode}}} mathbb {I}(c_i = c^*) log sigma _i,  (6)
where σi is the assigned score for mode ci, Nmode is the
number of candidate modes, and I is the indicator function.
The side-task loss is defined as:
  text {SideTaskLoss}(bar {s}^{0}_{1:T}, s^{0,text {gt}}_{1:T}) = frac {1}{T} sum _{t=1}^T left | bar {s}_t^0 - s_t^{0,text {gt}} right |_1,  (7)
where s̄0
t is the output ego future trajectory.
Generator loss with RL. The PPO [31] loss consists of
three parts: policy, value, and entropy loss. The policy loss
is defined as:
  begin {split} &text {PolicyLoss}(a_{0:T-1}, d_{0:T-1, text {new}}, d_{0:T-1}, A_{0:T-1})  = &- frac {1}{T} sum _{t=0}^{T-1} min left ( r_t A_t, text {clip}(r_t, 1-epsilon , 1+epsilon ) A_t right ), end {split} 
(8)
Algorithm 1 Training Procedure of CarPlanner
1: Input: Dataset D containing initial states s0 and ground-truth
trajectories s0:N,gt
1:T , longitudinal modes clon, discount factor γ,
GAE parameter λ, update interval I.
2: Require: Non-reactive transition model β, mode selector
fselector, policy π, policy old πold.
3: Step 1: Training Transition Model
4: for (s0, s1:N,gt
1:T ) ∈ D do
5: Simulate agent trajectories s1:N
1:T ← β(s0)
6: Calculate loss Ltm ← L1Loss(s1:N
1:T , s1:N,gt
1:T )
7: Backpropagate and update β using Ltm
8: end for
9: Step 2: Training Selector and Generator
10: Initialize training step ← 0
11: Initialize policy old πold ← π
12: for (s0, s0,gt
1:T ) ∈ D do
13: Non-Reactive Transition Model:
14: Simulate agent trajectories s1:N
1:T ← β(s0)
15: Mode Assignment:
16: Determine clat based on s0
17: Concatenate clat and clon to get c
18: Determine positive mode c∗
based on s0,gt
1:T and c
19: Mode Selector Loss:
20: Compute scores σ, s̄0
1:T ← fselector(s0, c)
21: Lselector ← CrossEntropyLoss(σ, c∗
) +
SideTaskLoss(s̄0
1:T , s0,gt
1:T )
22: Generator Loss:
23: if Reinforcement Learning (RL) Training then
24: Use πold, s0, c∗
, and s1:N
1:T to collect rollout data
(s0:T −1, a0:T −1, d0:T −1, V0:T −1, R0:T −1)
25: Compute advantage A0:T −1 and return R̂0:T −1 using
GAE [30]: A0:T −1, R̂0:T −1 ← GAE(R0:T −1, V0:T −1, γ, λ)
26: Compute policy distribution and value estimates:
(d0:T −1,new, V0:T −1,new) ← π(s0:T −1, a0:T −1, c∗
)
27: Lgenerator ← ValueLoss(V0:T −1,new, R̂0:T −1) +
PolicyLoss(d0:T −1,new, d0:T −1, A0:T −1) −
Entropy(d0:T −1,new)
28: else if Imitation Learning (IL) Training then
29: Use π, s0, c∗
, and s1:N
1:T to collect action sequence
a0:T −1
30: Stack action sequence as ego-planned trajectory
s0
1:T ← Stack(a0:T −1)
31: Lgenerator ← L1Loss(s0
1:T , s0,gt
1:T )
32: end if
33: Overall Loss:
34: L ← Lselector + Lgenerator
35: Backpropagate and update fselector, π using L
36: Policy Update:
37: Increment training step ← training step +1
38: if training step % I == 0 then
39: Update πold ← π
40: end if
41: end for
where the ratio rt is given by rt =
Prob(at,dt,new)
Prob(at,dt) , dt,new
and dt are the policy distributions (mean and standard devi-
ation of Gaussian distribution) at time step t induced by π
and πold respectively, the function Prob(a, d) calculates the
probability of a given action a under a distribution d, and
At is the advantage estimated using GAE [30]. The value
and entropy loss are defined as:
  text {ValueLoss}(V_{0:T-1, text {new}}, hat {R}_{0:T-1}) = frac {1}{T} sum _{t=0}^{T-1} left | V_{t,text {new}} - hat {R}_t right |_2^2, 
(9)
  text {Entropy}(d_{0:T-1, text {new}}) = frac {1}{T} sum _{t=0}^{T-1} mathcal {H}(d_{t,text {new}}),  (10)
where Vt,new and R̂t are the predicted and actual returns,
and H represents the entropy of the policy distribution d.
Generator loss with IL. In IL, the generator minimizes the
trajectory error between the ego-planned trajectory s0
1:T and
the ground-truth trajectory s0,gt
1:T . The loss is defined as:
  L_{text {generator}} = frac {1}{T} sum _{t=1}^T left | s_t^0 - s_t^{0,text {gt}} right |_1.  (11)
B. Implementation Details
The hyperparameters of model architecture, PPO-related
parameters, and loss weights are summarized in Tab. 5. The
magnitudes of value, policy, and entropy loss are 103
, 100
,
and 10−3
, respectively. The trajectory generator generates
trajectories with a time horizon of 8 seconds at 1-second
intervals, corresponding to time horizon T = 8. During
testing, these trajectories are interpolated to 0.1-second in-
tervals. The weight of scores generated by the rule and
mode selectors is set to a ratio of 1 : 0.3. In cases where
no ego candidate trajectory satisfies the safety criteria eval-
uated by the rule selector, an emergency stop is triggered.
For the Test14-Random benchmark, a replanning frequency
of 10Hz is employed, adhering to the official nuPlan sim-
ulation configuration. In contrast, for the Reduced-Val14
benchmark, a replanning frequency of 1Hz is used to en-
sure a fair comparison with Gen-Drive [18].
C. Ablation Study on RL Training
In this part, we examine the training efficiency of CarPlan-
ner, performance of vanilla and consistent auto-regressive
frameworks, the use of reactive and non-reactive model in
RL training, and the impact of varying the time horizon.
Training efficiency. We compare the efficiency of our
model-based framework with that of ScenarioNet [24],
which is an open-source platform for model-free RL train-
ing in real-world datasets [2, 9]. As shown in Tab. 6, Car-
Planner achieves a remarkable improvement in sampling ef-
Parameter Value
Feature dimension D 256
Static point dimension Dm 9
Agent pose dimension Da 10
Activation ReLU
Number of layers 3
Number of attention heads 8
Dropout 0.1
discount factor γ 0.1
GAE parameter λ 0.9
Clip range ϵ 0.2
Update interval I 8
Weight of selector loss 1
Weight of value loss 3
Weight of policy loss 100
Weight of entropy loss 0.001
Weight of IL loss 1
Table 5. Hyperparameters of model architecture, PPO-related pa-
rameters, and loss weights.
Planner CLS-NR (↑)
Efficiency
(samples/sec, ↑)
Num. Samples Train Time
ScenarioNet [24] 55.60 25.72 7,798,472 3d12h11m38s
CarPlanner-IL 93.41 1,181.46 70,487,200 16h34m12s
CarPlanner 94.07 1,632.25 70,487,200 11h59m44s
Table 6. Comparison of training efficiency with model-free set-
tings. Experimental results are based on the Test14-Random non-
reactive benchmark.
Design Choices Closed-loop metrics (↑)
Model Type
Random
Sample
Guide
Reward
CLS-NR S-CR S-Area S-PR S-Comfort
Vanilla
✓ Progress 67.56 ± 0.38 90.97 ± 0.78 94.64 ± 1.72 72.17 ± 0.21 64.21 ± 1.29
✓ DE 86.89 ± 0.28 97.34 ± 0.37 96.36 ± 0.18 89.90 ± 0.11 94.03 ± 0.65
Consistent
✗ FE 88.14 96.86 98.43 91.39 73.73
✗ DE 94.07 99.22 99.22 95.06 91.09
Table 7. Comparison of vanilla and consistent auto-regressive
frameworks with different guide reward design. Experimental re-
sults are based on the Test14-Random non-reactive benchmark.
Model Loss CLS-NR (↑)
Consistent Ratio
Lat (↑)
Consistent Ratio
Lon (↑)
Vanilla RL 86.89 ± 0.28 20.00 ± 0.10 8.33 ± 0.00
PLUTO [3] IL 91.92 62.45 41.80
Consistent IL 93.41 68.26 43.01
Consistent RL 94.07 79.58 43.03
Table 8. Comparison for consistency. Experimental results are
based on the Test14-Random non-reactive benchmark.
ficiency, outperforming ScenarioNet by two orders of mag-
nitude. Furthermore, CarPlanner not only excels in ef-
Closed-loop metrics (↑)
Transition Model CLS-NR S-CR S-Area S-PR S-Comfort
Reactive 91.03 96.92 99.23 91.28 90.00
Non-reactive 94.07 99.22 99.22 95.06 91.09
Table 9. Comparison of the usage of reactive and non-reactive
transition models. Experimental results are based on the Test14-
Random non-reactive benchmark.
   
7LPH+RUL]RQ7UDLQ
7L
P
H+RUL
]
RQ7HVW
/615
Figure 4. Performance of different training time horizons under
different testing time horizons. The value in each cell is the CLS-
NR metric on the Test14-Random non-reactive benchmark.
ficiency but also achieves SOTA performance, surpassing
ScenarioNet by a wide margin.
Vanilla vs. consistent auto-regressive framework. The
results are shown in Tabs. 7 and 8. The consistent auto-
regressive framework generates multi-modal trajectories by
conditioning on mode representations. In contrast, the
vanilla framework relies on random sampling from the ac-
tion Gaussian distribution to produce multi-modal trajec-
tories. To ensure comparability in the number of modes
generated by both frameworks, we sample 60 trajectories in
parallel for the vanilla framework. Given that random sam-
pling introduces variability, we average the results across 3
random seeds. For the consistent framework, we use dis-
placement error (DE) and final error (FE) as guide func-
tions to assist the policy in generating mode-aligned trajec-
tories. For the vanilla framework, DE is compared against a
progress reward, which encourages longitudinal movement
along the route while discouraging excessive lateral devia-
tions that move the vehicle too far from any possible route.
The consistent ratio computes the ratio of generated trajec-
tories that fall in their corresponding modes in longitudinal
and lateral directions separately.
Overall, our proposed consistent framework outperforms
the vanilla framework in terms of closed-loop performance,
highlighting the benefits of incorporating consistency. Fur-
thermore, RL provides more consistant trajectories than the
vanilla framework and IL-based methods. Additionally, we
find that DE serves as an effective guide function for policy
training, further enhancing closed-loop performance.
Reactive vs. non-reactive transition model. We compare
the performance of the CarPlanner framework when trained
with reactive and non-reactive transition models. The reac-
tive transition model shares a similar architecture with the
auto-regressive planner for the ego vehicle, utilizing rela-
tive pose encoding [43] as the backbone network to extract
features of traffic agents and predict their subsequent poses.
The training loss and hyperparameters are consistent with
those used for the non-reactive transition model. As shown
in Tab. 9, except for the S-Area metric, using non-reactive
transition model outperforms the reactive transition model
in our current implementation. The primary difference lies
in the assumptions about traffic agents: the reactive transi-
tion model assumes that the ego vehicle can negotiate with
traffic agents and share the same priority, whereas in the
non-reactive model, traffic agents do not respond to the ego
vehicle, effectively assigning them higher priority. A repre-
sentative example is presented in Fig. 5. When trained with
the reactive transition model, the planner assumes pedestri-
ans will yield to the vehicle, leading it to attempt to move
forward. However, at tsim = 12s, the planner collides with
pedestrians, triggering an emergency brake, which nega-
tively impacts safety, progress, and comfort metrics. Al-
though the performance of using reactive transition model
is not satisfied currently, it is a more realistic assumption
and we will further investigate this in future work.
Time horizon. We evaluate the CarPlanner framework by
training it with different time horizons, including 1, 3, 5,
and 8 seconds, and testing the planners in each time hori-
zon. The results in Fig. 4 confirm that increasing the time
horizon has a positive effect on the performance for both
training and testing. A special case is when the training time
horizon is set to 1, all tested time horizons exhibit poor per-
formance, highlighting the importance of multi-step learn-
ing in RL. Additionally, the observation that increasing the
training time horizon enhances closed-loop performance
suggests the potential for further improvements by extend-
ing the time horizon beyond 8 seconds. However, due to
current limitations in data preparation, which is designed
for horizons up to 8 seconds, expanding the time horizon
would not provide map information or ground-truth trajec-
tories, hindering further analysis. Consequently, we leave
this exploration for future work.
D. Comparison with Differentiable Loss
In typical IL setting, the supervision signal provided to
the trajectory generator is the displacement error (DE) be-
tween the ego-planned trajectory and the ground-truth tra-
jectory. Several works [3, 17, 35] propose to convert non-

How Are Learning-Based Methods Reshaping Trajectory Planning in Autonomous Driving?

  • 1.
    CarPlanner: Consistent Auto-regressiveTrajectory Planning for Large-scale Reinforcement Learning in Autonomous Driving Dongkun Zhang1,2 Jiaming Liang2 Ke Guo2 Sha Lu1 Qi Wang2 Rong Xiong1,B Zhenwei Miao2,† Yue Wang1 1 Zhejiang University 2 Cainiao Network 1 {zhangdongkun, lusha, rxiong, ywang24}@zju.edu.cn 2 {liangjiaming.ljm, muguo.gk, ruifeng.wq, zhenwei.mzw}@alibaba-inc.com Abstract Trajectory planning is vital for autonomous driving, ensur- ing safe and efficient navigation in complex environments. While recent learning-based methods, particularly rein- forcement learning (RL), have shown promise in specific scenarios, RL planners struggle with training inefficiencies and managing large-scale, real-world driving scenarios. In this paper, we introduce CarPlanner, a Consistent auto- regressive Planner that uses RL to generate multi-modal trajectories. The auto-regressive structure enables efficient large-scale RL training, while the incorporation of consis- tency ensures stable policy learning by maintaining coher- ent temporal consistency across time steps. Moreover, Car- Planner employs a generation-selection framework with an expert-guided reward function and an invariant-view mod- ule, simplifying RL training and enhancing policy perfor- mance. Extensive analysis demonstrates that our proposed RL framework effectively addresses the challenges of train- ing efficiency and performance enhancement, positioning CarPlanner as a promising solution for trajectory planning in autonomous driving. To the best of our knowledge, we are the first to demonstrate that the RL-based planner can sur- pass both IL- and rule-based state-of-the-arts (SOTAs) on the challenging large-scale real-world dataset nuPlan. Our proposed CarPlanner surpasses RL-, IL-, and rule-based SOTA approaches within this demanding dataset. 1. Introduction Trajectory planning [36] is essential in autonomous driv- ing, utilizing outputs from perception and trajectory predic- tion modules to generate future poses for the ego vehicle. A controller tracks this planned trajectory, producing con- trol commands for closed-loop driving. Recently, learning- based trajectory planning has garnered attention due to its †Project lead. B Corresponding author. Figure 1. Frameworks for multi-step trajectory generation. (a) Initialization-refinement that generates an initial trajectory and refines it iteratively. (b) Vanilla auto-regressive models that decode subsequent poses sequentially. (c) Our consistent auto-regressive model that integrates time-consistent mode information. potential to automate algorithm iteration, eliminate tedious rule design, and ensure safety and comfort in diverse real- world scenarios [36]. Most existing researches [11, 16, 29] employ imitation learning (IL) to align planned trajectories with those of human experts. However, this approach suffers from dis- tribution shift [28] and causal confusion [8]. Reinforce- ment learning (RL) offers a potential solution, addressing these challenges and providing richer supervision through reward functions. Although RL shows effectiveness in do- mains such as games [34], robotics [19], and language mod- els [25], it still struggles with training inefficiencies and per- formance issues in the large-scale driving task. To the ex- arXiv:2502.19908v3 [cs.RO] 24 Mar 2025
  • 2.
    tent of ourknowledge, no RL methods have yet achieved competitive results on large-scale open datasets such as nu- Plan [2], which features diverse real-world scenarios. Thus, this paper aims to tackle two key challenges in RL for trajectory planning: 1) training inefficiency and 2) poor performance. Training inefficiency arises from the fact that RL typically operates in a model-free setting, necessi- tating an inefficient simulator running on a CPU to repeat- edly roll out a policy for data collection. To overcome this challenge, we propose an efficient model-based approach utilizing neural networks as transition models. Our method is optimized for execution on hardware accelerators such as GPUs, rendering our time cost comparable to that of IL- based methods. To apply RL to solve the trajectory planning problem, we formulate it as a multi-step sequential decision-making task utilizing a Markov Decision Process (MDP). Existing methods that generate the trajectory† in multiple steps gen- erally fall into two categories: initialization-refinement [17, 20, 33, 45] and auto-regressive models [27, 32, 41, 46]. The first category, illustrated in Fig. 1 (a), involves gen- erating an initial trajectory estimate and subsequently refin- ing it through iterative applications of RL. However, recent studies, including Gen-Drive [18], suggest that it continues to lag behind SOTA IL and rule-based planners. One no- table limitation of this approach is its neglect of the tempo- ral causality inherent in the trajectory planning task. Addi- tionally, the complexity of direct optimization over high- dimensional trajectory space can hinder the performance of RL algorithms. The second category consists of auto- regressive models, shown in Fig. 1 (b), which generate the poses of the ego vehicle recurrently using a single- step policy within a transition model. In this category, ego poses at all time steps are consolidated to form the overall planned trajectory. As taking temporal causality into ac- count, current auto-regressive models allow for interactive behaviors. However, a common limitation is their reliance on auto-regressively random sampling from action distri- butions to generate multi-modal trajectories. This vanilla auto-regressive procedure may compromise long-term con- sistency and unnecessarily expand the exploration space in RL, leading to poor performance. To address the limitations of auto-regressive models, we introduce CarPlanner, a Consistent auto-regressive model designed for efficient, large-scale RL-based Planner train- ing (see Fig. 1 (c)). The key insight of CarPlanner is its incorporation of consistent mode representation as condi- tions for the auto-regressive model. Specifically, we lever- age a longitudinal-lateral decomposed mode representation, where the longitudinal mode is a scalar that captures aver- †In this paper, the term “trajectory” refers to the future poses of the ego vehicle or traffic agents. To avoid confusion, we use the term “state (action) sequence” to refer to the “trajectory” in the RL community. age speeds, and the lateral mode encompasses all possible routes derived from the current state of the ego vehicle along with map information. This mode remains constant across time steps, providing stable and consistent guidance during policy sampling. Furthermore, we propose a universal reward function that suits large-scale and diverse scenarios, eliminating the need for scenario-specific reward designs. This function consists of an expert-guided and task-oriented term. The first term quantifies the displacement error between the ego- planned trajectory and the expert’s trajectory, which, along with the consistent mode representation, narrows down the policy’s exploration space. The second term incorporates common senses in driving tasks including the avoidance of collision and adherence to the drivable area. Additionally, we introduce an Invariant-View Module (IVM) to supply invariant-view input for policy, with the aim of providing time-agnostic policy input, easing the feature learning and embracing generalization. To achieve this, IVM prepro- cesses state and lateral mode by transforming agent, map, and route information into the ego’s current coordinate and by clipping information that is distant from the ego. To our knowledge, we are the first to demonstrate that RL-based planner outperforms state-of-the-art (SOTA) IL and rule-based approaches on the challenging large-scale nuPlan dataset. In summary, the key contributions of this paper are highlighted as follows: • We present CarPlanner, a consistent auto-regressive planner that trains an RL policy to generate consistent multi-modal trajectories. • We introduce an expert-guided universal reward function and IVM to simplify RL training and improve policy generalization, leading to enhanced closed-loop perfor- mance. • We conduct a rigorous analysis on the characteristics of IL and RL training, providing insights into their strengths and limitations, while highlighting the advantages of RL in tackling challenges such as distribution shift and causal confusion. • Our framework showcases exceptional performance, sur- passing all RL-, IL-, and rule-based SOTAs on the nuPlan benchmark. This underscores the potential of RL in nav- igating complex real-world driving scenarios. 2. Related Work 2.1. Imitation-based Planning The use of IL to train planners based on human demon- strations has garnered significant interest recently. This approach leverages the driving expertise of experienced drivers who can safely and comfortably navigate a wide range of real-world scenarios, along with the added advan- tage of easily collectible driving data at scale [2, 9, 15]. Nu-
  • 3.
    s0 Non-reactive Transition Model Mode Selector c ModeScores Previewed World s0 Consistent Mode as Query s1 a0 a1 ... ... Multi-modal trajectories Rule Selector Ego-planned trajectory Trajectory Generator Rule-augmented Selector NR Transition Model Auto-regressive Policy IVM IVM s2 Mode Selector Figure 2. CarPlanner contains four parts. (1) The non-reactive transition model takes initial state s0 as input and predicts the future trajectories of traffic agents. (2) The mode selector outputs scores based on the initial state and the modes c. (3) The trajectory generator obeys an auto-regressive structure condition on the consistent mode and produces mode-aligned multi-modal trajectories. (4) The rule- augmented selector compensates the mode scores by safety, comfort, and progress metrics. merous studies [5, 17, 29] have focused on developing inno- vative networks to enhance open-loop performance in this domain. However, the ultimate challenge of autonomous driving is achieving closed-loop operation, which is eval- uated using driving-oriented metrics such as safety, adher- ence to traffic rules, comfort, and progress. This reveals a significant gap between the training and testing phases of planners. Moreover, IL is particularly vulnerable to issues such as distribution shift [28] and causal confusion [8]. The first issue results in suboptimal decisions when the system encounters scenarios that are not represented in the train- ing data distribution. The second issue arises when net- works inadvertently capture incorrect correlations and de- velop shortcuts based on input information, primarily due to the reliance on imitation loss from expert demonstrations. Despite efforts in several studies [1, 3, 4, 42] to address these challenges, the gap between training and testing re- mains substantial. 2.2. RL in Autonomous Driving In the field of autonomous driving, RL has demonstrated its effectiveness in addressing specific scenarios such as high- way driving [22, 39], lane changes [14, 23], and unprotected left turns [22, 40]. Most methods directly learn policies over the control space, which includes throttle, brake, and steering commands. Due to the high frequency of control command execution, the simulation process can be time- consuming, and exploration can be inconsistent [40]. Sev- eral works [40, 44] have proposed learning trajectory plan- ners with actions defined as ego-planned trajectories, which temporally extend the exploration space and improve train- ing efficiency. However, a trade-off exists between the trajectory horizon and training performance, as noted in ASAP-RL [40]. Increasing the trajectory horizon results in less reactive behaviors and a reduced amount of data, while a smaller trajectory horizon leads to challenges sim- ilar to those encountered in control space. Additionally, these methods typically employ a model-free setting, mak- ing them difficult to apply to the complex, diverse real- world scenarios found in large-scale driving datasets. In this paper, we propose adopting a model-based formulation that can facilitate RL training on large-scale datasets. Under this formulation, we aim to overcome the trajectory horizon trade-off by using a transition model, which can provide a preview of the world in which our policy can make multi- step decisions during testing. 3. Method 3.1. Preliminaries MDP is used to model sequential decision problems, for- mulated as a tuple ⟨S, A, Pτ , R, ρ0, γ, T⟩. S is the state space. A is the action space. Pτ : S × A → ∆(S) † is the state transition probability. R : S × A → R denotes the reward function and is bounded. ρ0 ∈ ∆(S) is the initial state distribution. T is the time horizon and γ is the dis- count factor of future rewards. The state-action sequence is defined as τ = (s0, a0, s1, a1, . . . , sT ), where st ∈ S and at ∈ A are the state and action at time step t. The objective of RL is to maximize the expected return: setlength {abovedisplayskip }{0pt} begin {split} max _{pi } mathbb {E}_{boldsymbol {s}_t sim {P_{tau }}, a_t sim pi } [sum _{t=0}^{T}gamma ^{t} R(boldsymbol {s}_t, a_t)]. label {equation:mdp} end {split} setlength {belowdisplayskip }{0pt} (1) Vectorized state representation. State st contains map and agent information in vectorized representation [10]. Map information m includes the road network, traffic lights, etc, which are represented by polylines and polygons. Agent information includes the current and past poses of ego vehicle and other traffic agents, which are represented by polylines. The index of ego vehicle is 0 and the indices of traffic agents range from 1 to N. For each agent i, its history is denoted as si t−H:t, i ∈ {0, 1, . . . , N}, where H is the history time horizon. †∆(X) denotes the set of probability distribution over set X.
  • 4.
    3.2. Problem Formulation Wemodel the trajectory planning task as a sequential deci- sion process and decouple the auto-regressive models into policy and transition models. The key to connect trajec- tory planning and auto-regressive models is to define the action as the next pose of ego vehicle, i.e., at = s0 t+1. Therefore, after forwarding the auto-regressive model, the decoded pose is collected to be the ego-planned trajectory. Specifically, we can reduce the state-action sequence to the state sequence under this definition and vectorized repre- sentation: setlength {abovedisplayskip }{0pt} begin {split} &P(boldsymbol {s}_0, a_0, boldsymbol {s}_1, a_1, dots , boldsymbol {s}_T) &= P(m, s^{0:N}_{-H:0}, s^{0}_{1}, m, s^{0:N}_{1-H:1}, s^{0}_{2}, dots , m, s^{0:N}_{T-H:T}) &= P(m, s^{0:N}_{-H:0}, m, s^{0:N}_{1-H:1}, dots , m, s^{0:N}_{T-H:T}) &= P(boldsymbol {s}_0, boldsymbol {s}_1, dots , boldsymbol {s}_T). [-0.1cm] end {split} setlength {belowdisplayskip }{0pt} (2) The state sequence can be further formulated in an auto- regressive fashion and decomposed into policy and transi- tion model: setlength {abovedisplayskip }{0pt} begin {split} &P(boldsymbol {s}_0, boldsymbol {s}_1, dots , boldsymbol {s}_T) = rho _{0}(boldsymbol {s}_0) prod _{t=0}^{T-1} P(boldsymbol {s}_{t+1} | boldsymbol {s}_t) &= rho _{0}(boldsymbol {s}_0) prod _{t=0}^{T-1} P(s^0_{t+1}, s^{1:N}_{t+1} | boldsymbol {s}_t) &= rho _{0}(boldsymbol {s}_0) prod _{t=0}^{T-1} underbrace {pi (a_t | boldsymbol {s}_t)}_{text {Policy}} underbrace {P_{tau }(s^{1:N}_{t+1} | boldsymbol {s}_t)}_{text {Transition Model}}. [-0.15cm] end {split} label {equation:incons} setlength {belowdisplayskip }{0pt} (3) From Eq. (3), we can clearly identify the inherent problem associated with the typical auto-regressive approach: incon- sistent behaviors across time steps arise from the policy dis- tribution, which depends on random sampling from the ac- tion distribution. To solve the above problem, we introduce consistent mode information c that remains unchanged across time steps into the auto-regressive fashion: setlength {abovedisplayskip }{0pt} begin {split} &P(boldsymbol {s}_0, boldsymbol {s}_1, dots , boldsymbol {s}_T) = int _{boldsymbol {c}} P(boldsymbol {s}_0, boldsymbol {s}_1, dots , boldsymbol {s}_T, boldsymbol {c}) dboldsymbol {c} &= rho _0(boldsymbol {s}_0) int _{boldsymbol {c}} P(boldsymbol {c} | boldsymbol {s}_0) P(boldsymbol {s}_1, dots , boldsymbol {s}_T | boldsymbol {c}) dboldsymbol {c} &= rho _0(boldsymbol {s}_0) prod _{t=0}^{T-1} underbrace {P_{tau }(s^{1:N}_{t+1} | boldsymbol {s}_t)}_{text {Transition Model}} int _{boldsymbol {c}} underbrace {P(boldsymbol {c} | boldsymbol {s}_0)}_{text {Mode Selector}} prod _{t=0}^{T-1} underbrace {pi (a_t | boldsymbol {s}_t, boldsymbol {c})}_{text {Policy}} dboldsymbol {c}. [-1.cm] label {equation:car} end {split} setlength {belowdisplayskip }{0pt} (4) Since we focus on the ego trajectory planning, the consis- tent mode c does not impact transition model. This consistent auto-regressive formulation defined in Eq. (4) reveals a generation-selection framework where the mode selector scores each mode based on the initial state s0 and the trajectory generator generates multi-modal trajecto- ries via sampling from the mode-conditioned policy. Non-reactive transition model. The transition model for- mulated in Eq. (4) needs to be employed in every time step since it produces the poses of traffic agents at time step t + 1 based on current state st. In practice, this process is time-consuming and we do not observe a performance improvement by using this transition model, therefore, we use trajectory predictors P(s1:N 1:T |s0) as non-reactive transi- tion model that produces all future poses of traffic agents in one shot given initial state s0. 3.3. Planner Architecture The framework of our proposed CarPlanner is illustrated in Fig. 2, comprising four key components: 1) the non- reactive transition model, 2) the mode selector, 3) the tra- jectory generator, and 4) the rule-augmented selector. Our planner operates within a generation-selection framework. Given an initial state s0 and all possible Nmode modes, the trajectory selector evaluates and assigns scores to each mode. The trajectory generator then produces Nmode trajectories that correspond to their respective modes. For trajectory generator, the initial state s0 is replicated Nmode times, each associated with one of the Nmode modes, effec- tively creating Nmode parallel worlds. The policy is executed within these previewed worlds. During the policy rollout, a trajectory predictor acts as the state transition model, gener- ating future poses of traffic agents across all time horizons. 3.3.1. Non-reactive Transition Model This module takes the initial state s0 as input and outputs the future trajectories of traffic agents. The initial state is processed by agent and map encoders, followed by a self- attention Transformer encoder [38] to fuse the agent and map features. The agent features are then decoded into fu- ture trajectories. Agent and map encoders. The state s0 contains both map and agent information. The map information m consists of Nm,1 polylines and Nm,2 polygons. The polylines describe lane centers and lane boundaries, with each polyline con- taining 3Np points, where 3 corresponds to the lane center, the left boundary, and the right boundary. Each point is with dimension Dm = 9 and includes the following attributes: x, y, heading, speed limit, and category. When concatenated, the points of the left and right boundaries together with the center point yield a dimension of Nm,1 × Np × 3Dm. We leverage a PointNet [26] to extract features from the points of each polyline, resulting in a dimensionality of Nm,1 × D, where D represents the feature dimension. The polygons represent intersections, crosswalks, stop lines, etc, with each polygon containing Np points. We utilize another PointNet to extract features from the points of each poly- gon, producing a dimension of Nm,2 × D. We then con- catenate the features from both polylines and polygons to form the overall map features, resulting in a dimension of Nm × D. The agent information A consists of N agents, where each agent maintains poses for the past H time steps. Each pose is with dimension Da = 10 and includes the
  • 5.
    following attributes: x,y, heading, velocity, bounding box, time step, and category. Consequently, the agent informa- tion has a dimension of N × H × Da. We apply another PointNet to extract features from the poses of each agent, yielding an agent feature dimension of N × D. 3.3.2. Mode Selector This module takes s_0 and longitudinal-lateral decomposed mode information as input and outputs the probability of each mode. The number of modes Nmode = NlatNlon. Route-speed decomposed mode. To capture the longitu- dinal behaviors, we generate N_{text {lon}} modes that represent the average speed of the trajectory associated with each mode. Each longitudinal mode c_{text {lon},j} is defined as a scalar value of frac {j}{N_{text {lon}}} , repeated across a dimension D . As a result, the dimen- sionality of the longitudinal modes is N_{text {lon}} times D . For lateral behaviors, we identify N_{text {lat}} possible routes from the map using a graph search algorithm. These routes correspond to the lanes available for the ego vehicle. The dimensional- ity of these routes is N_{text {lat}} times N_r times D_m . We employ another PointNet to aggregate the features of the N_r points along each route, producing a lateral mode with a dimension of N_{text {lat}} times D . To create a comprehensive mode representation c, we combine the lateral and longitudinal modes, resulting in a combined dimension of N_{text {lat}} times N_{text {lon}} times 2D . To align this mode information with other feature dimensions, we pass it through a linear layer, mapping it back to N_{text {lat}} times N_{text {lon}} times D . Query-based Transformer decoder. This decoder is em- ployed to fuse the mode features with map and agent fea- tures derived from s_0 . In this framework, the mode serves as the query, while the map and agent information act as the keys and values. The updated mode features are decoded through a multi-layer perceptron (MLP) to yield the scores for each mode, which are subsequently normalized using the softmax operator. 3.3.3. Trajectory Generator This module operates in an auto-regressive manner, recur- rently decoding the next pose of the ego vehicle at, given the current state st, and consistent mode information c. Invariant-view module (IVM). Before feeding the mode and state into the network, we preprocess them to eliminate time information. For the map and agent information in state st, we select the K-nearest neighbors (KNN) to the ego current pose and only feed these into the policy. K is set to the half of map and agent elements respectively. Regarding the routes that capture lateral behaviors, we filter out the segments where the point closest to the current pose of the ego vehicle is the starting point, retaining Kr points. In this case, Kr is set to a quarter of Nr points in one route. Finally, we transform the routes, agent, and map poses into the coordinate frame of the ego vehicle at the current time step t. We subtract the historical time steps t − H : t from the current time step t, yielding time steps in range −H : 0. Query-based Transformer decoder. We employ the same backbone network architecture as the mode selector, but with different query dimensions. Due to the IVM and the fact that different modes yield distinct states, the map and agent information cannot be shared among modes. As a re- sult, we fuse information for each individual mode. Specifi- cally, the query dimension is 1 times D, while the dimensions of the keys and values are (N + N_{m}) times D. The output feature dimension remains 1 times D. Note that Transformer decoder can process information from multiple modes in parallel, eliminating the need to handle each mode sequentially. Policy output. The mode feature is processed by two dis- tinct heads: a policy head and a value head. Each head com- prises its own MLP to produce the parameters for the action distribution and the corresponding value estimate. We em- ploy a Gaussian distribution to model the action distribu- tion, where actions are sampled from this distribution dur- ing training. In contrast, during inference, we utilize the mean of the distribution to determine the actions. 3.3.4. Rule-augmented Selector This module is only utilized during inference and takes as input the initial state s_0 , the multi-modal ego-planned tra- jectories, and the predicted future trajectories of agents. It calculates driving-oriented metrics such as safety, progress, comfort. A comprehensive score is obtained by the weighted sum of rule-based scores and the mode scores pro- vided by the mode selector. The ego-planned trajectory with the highest score is selected as the output of the planner. 3.4. Training We first train the non-reactive transition model and freeze the weights during the training of the mode selector and tra- jectory generator. Instead of feeding all modes to the gen- erator, we apply a winner-takes-all strategy, wherein a posi- tive mode is assigned based on the ego ground-truth trajec- tory and serves as a condition for the trajectory generator. Mode assignment. For the lateral mode, we assign the route closest to the endpoint of ego ground-truth trajectory as the positive lateral mode. For the longitudinal mode, we partition the longitudinal space into N_{text {lon}} intervals and as- sign the interval containing the endpoint of the ground-truth trajectory as the positive longitudinal mode. Reward function. To handle diverse scenarios, we use the negative displacement error (DE) between the ego future pose and the ground truth as a universal reward. We also in- troduce additional terms to improve trajectory quality: col- lision rate and drivable area compliance. If the future pose collides or falls outside the drivable area, the reward is set to -1; otherwise, it is 0. Mode dropout. In some cases, there are no available routes for ego to follow. However, since routes serve as queries in Transformer, the absence of a route can lead to unstable or hazardous outputs. To mitigate this issue, we implement a
  • 6.
    mode dropout moduleduring training that randomly masks routes to prevent over-reliance on this information. Loss function. For the selector, we use cross-entropy loss that is the negative log-likelihood of the positive mode and a side task that regresses the ego ground-truth trajectory. For the generator, we use PPO [31] loss that consists of three parts: policy improvement, value estimation, and entropy. Full description can be found in supplementary. 4. Experiments 4.1. Experimental Setup Dataset and simulator. We use nuPlan [2], a large-scale closed-loop platform for studying trajectory planning in au- tonomous driving, to evaluate the efficacy of our method. The nuPlan dataset contains driving log data over 1,500 hours collected by human expert drivers across 4 diverse cities. It includes complex, diverse scenarios such as lane follow and change, left and right turn, traversing intersec- tions and bus stops, roundabouts, interaction with pedestri- ans, etc. As a closed-loop platform, nuPlan provides a sim- ulator that uses scenarios from the dataset as initialization. During the simulation, traffic agents are taken over by log- replay (non-reactive) or an IDM [37] policy (reactive). The ego vehicle is taken over by user-provided planners. The simulator lasts for 15 seconds and runs at 10 Hz. At each timestamp, the simulator queries the planner to plan a tra- jectory, which is tracked by an LQR controller to generate control commands to drive the ego vehicle. Benchmarks and metrics. We use two benchmarks: Test14-Random and Reduced-Val14 for comparing with other methods and analyzing the design choices within our method. The Test14-Random provided by PlanTF [4] contains 261 scenarios. The Reduced-Val14 provided by PDM [7] contains 318 scenarios. We use the closed-loop score (CLS) provided by the of- ficial nuPlan devkit† to assess the performance of all meth- ods. The CLS score comprehends different aspects such as safety (S-CR, S-TTC), drivable area compliance (S-Area), progress (S-PR), comfort, etc. Based on the different be- havior types of traffic agents, CLS is detailed into CLS-NR (non-reactive) and CLS-R (reactive). Implementation details. We follow PDM [7] to construct our training and validation splits. The size of the training set is 176,218 where all available scenario types are used, with a number of 4,000 scenarios per type. The size of the validation set is 1,118 where 100 scenarios with 14 types are selected. We train all models with 50 epochs in 2 NVIDIA 3090 GPUs. The batch size is 64 per GPU. We use AdamW optimizer with an initial learning rate of 1e-4 and reduce the learning rate when the validation loss stops decreasing with a patience of 0 and decrease factor of 0.3. For RL training, †https://github.com/motional/nuplan-devkit Type Planners CLS-NR CLS-R Rule IDM [37] 70.39 72.42 PDM-Closed [7] 90.05 91.64 IL RasterModel [2] 69.66 67.54 UrbanDriver [29] 63.27 61.02 GC-PGP [12] 55.99 51.39 PDM-Open [7] 52.80 57.23 GameFormer [17] 80.80 79.31 PlanTF [4] 86.48 80.59 PEP [42] 91.45 89.74 PLUTO [3] 91.92 90.03 RL CarPlanner (Ours) 94.07 91.1 Table 1. Comparison with SOTAs in Test14-Random. Based on the type of trajectory generator, all methods are categorized into Rule, IL, and RL. The best result is in bold and the second best result is underlined. Type Planners CLS-NR S-CR S-PR Rule PDM-Closed [7] 91.21 97.01 92.68 IL GameFormer [17] 83.76 94.73 88.12 PlanTF [4] 83.66 94.02 92.67 Gen-Drive (Pretrain) [18] 85.12 93.65 86.64 RL Gen-Drive (Finetune) [18] 87.53 95.72 89.94 CarPlanner (Ours) 91.45 96.38 95.37 Table 2. Comparison with SOTAs in Reduced-Val14 with non- reactive traffic agents. we set the discount γ = 0.1 and the GAE parameter λ = 0.9. The weights of value, policy, and entropy loss are set to 3, 100, and 0.001, respectively. The number of longitudinal modes is set to 12 and a maximum number of lateral modes are set to 5. 4.2. Comparison with SOTAs SOTAs. We categorize the methods into Rule, IL, and RL based on the type of trajectory generator. (1) PDM [7] wins the nuPlan challenge 2023, its IL-based and rule-based variants are denoted as PDM-Open and PDM-Closed, re- spectively. PDM-Closed follows the generation-selection framework where IDM is used to generate multiple candi- date trajectories and rule-based selector considering safety, progress, and comfort is used to select the best trajectory. (2) PLUTO [3] also obeys the generation-selection frame- work and uses contrastive IL to incorporate various data augmentation techniques and trains the generator. (3) Gen- Drive [18] is a concurrent work that follows a pretrain- finetune pipeline where IL is used to pretrain a diffusion- based planner and RL is used to finetune the denoising pro- cess based on a reward model trained by AI preference. Results. We compare our method with SOTAs in Test14- Random and Reduced-Val14 benchmark as shown in Tab. 1 and Tab. 2. Overall, our CarPlanner demonstrates superior performance, particularly in non-reactive environments. In the non-reactive setting, our method achieves the highest scores across all metrics, with an improvement of 4.02 and 2.15 compared to PDM-Closed and PLUTO, es- tablishing the potential of RL and the superior performance
  • 7.
    Design Choices Closed-loopmetrics (↑) Open-loop losses (↓) Reward DE Reward Quality Coord Trans KNN CLS-NR S-CR S-Area S-PR S-Comfort Loss Selector Loss Generator ✗ ✓ ✓ ✓ 31.79 95.74 98.45 33.10 48.84 1.03 30.3 ✓ ✗ ✓ ✓ 90.44 97.49 96.91 93.33 90.73 0.99 1221.6 ✓ ✓ ✗ ✓ 90.78 96.92 98.46 91.37 94.23 1.00 2130.7 ✓ ✓ ✓ ✗ 92.73 98.07 98.46 94.69 93.44 1.03 2083.6 ✓ ✓ ✓ ✓ 94.07 99.22 99.22 95.06 91.09 1.03 1624.5 Table 3. Ablation studies on the design choices in RL training. Results are in Test14-random non-reactive benchmark. of our proposed framework. Moreover, CarPlanner reveals substantial improvement in the progress metric S-PR com- pared to PDM-Closed in Tab. 2 and comparable collision metric S-CR, indicating the ability of our method to im- proving driving efficiency while maintaining safe driving. Importantly, we do not apply any techniques commonly used in IL such as data augmentation [3, 4] and ego-history masking [11], underscoring the intrinsic capability of our approach to solving the closed-loop task. In the reactive setting, while our method performs well, it falls slightly short of PDM-Closed. This discrepancy arises because our model was trained exclusively in non- reactive settings and has not interacted with the IDM policy used by reactive settings; as a result, our model is less robust to disturbances generated by reactive agents during testing. 4.3. Ablation Studies We investigate the effects of different design choices in RL training. The results are shown in Tab. 3. Influence of reward items. When using the quality re- ward only, the planner tends to generate static trajectories and achieves a low progress metric. This occurs because the ego vehicle begins in a safe, drivable state, but mov- ing forward is at risk of collisions or leaving the drivable area. On the other hand, when the quality reward is incor- porated alongside the DE reward, it leads to significant im- provements in closed-loop metrics compared to using the DE reward alone. For instance, the S-CR metric rises from 97.49 to 99.22, and the S-Area metric rises from 96.91 to 99.22. These improvements indicate that the quality reward encourages safe and comfortable behaviors. Effectiveness of IVM. The results show that the coordi- nate transformation and KNN techniques in IVM notably improve closed-loop metrics and generator loss. For in- stance, with the coordinate transformation technique, the overall closed-loop score increases from 90.78 to 94.07, and S-PR rises from 91.37 to 95.06. These improvements are at- tributed to the enhanced accuracy of value estimation in RL, leading to generalized driving in closed-loop. 4.4. Extention to IL In addition to designing for RL training, we also extend the CarPlanner to incorporate IL. We conduct rigorous analysis to compare the effects of various design choices in IL and RL training, as summarized in Tab. 4. Our findings indicate that while mode dropout and selector side task contribute to both IL and RL training, ego-history dropout and backbone sharing, often effective in IL, are less suitable for RL. Ego-history dropout. Previous works [1, 3, 4, 11] sug- gest that planners trained via IL may rely too heavily on past poses and neglect environmental state information. To counter this, we combine techniques from ChauffeurNet [1] and PlanTF [4] into an ego-history dropout module, ran- domly masking ego history poses and current velocity to alleviate the causal confusion issue. Our experiments confirm that ego-history dropout benifits IL training, as it improves performance across closed-loop metrics like S-CR and S-Area. However, in RL training, we observe a negative impact on advantage esti- mation due to ego-history dropout, which significantly af- fects the value part of generator loss, leading to closed-loop performance degradation. This suggests that RL training naturally addresses the causal confusion problem inherent in IL by uncovering causal relationships that align with the reward signal, which explicitly encodes task-oriented pref- erences. This capability highlights the potential of RL to push the boundaries of learning-based planning. Backbone sharing. This choice, often used in IL-based multi-modal planners, promotes feature sharing across tasks to improve generalization. While backbone sharing helps IL by balancing losses across trajectory generator and se- lector, we find it adversely affects RL training. Specifically, backbone sharing leads to higher losses for both the trajec- tory generator and selector in RL, indicating that gradients from each task interfere. The divergent objectives in RL for trajectory generation and selection tasks seem to con- flict, reducing overall policy performance. Consequently, we avoid backbone sharing in our RL framework to main- tain task-specific gradient flow and improve policy quality. 4.5. Qualitative Results We provide qualitative results as shown in Fig. 3. In this scenario, ego vehicle is required to execute a right turn while navigating around pedestrians. In this case, Our method shows a smooth, efficient performance. From tsim = 0s to tsim = 9s, all methods wait for the pedestrians to cross the road. At tsim = 10s, an unexpected pedestrian goes back and prepares to re-cross the road. PDM-Closed is
  • 8.
    Design Choices Closed-loopmetrics (↑) Open-loop losses (↓) Loss Type Mode Dropout Selector Side Task Ego-history Dropout Backbone Sharing CLS-NR S-CR S-Area S-PR S-Comfort Loss Selector Loss Generator IL ✗ ✗ ✗ ✗ 90.82 97.29 98.45 92.15 94.57 1.04 147.5 ✓ ✗ ✗ ✗ 91.21 96.54 98.46 91.44 96.92 1.07 153.0 ✓ ✓ ✗ ✗ 91.51 96.91 98.46 95.30 96.91 1.04 162.3 ✓ ✓ ✓ ✗ 92.72 98.06 98.84 94.88 95.35 1.04 167.5 ✓ ✓ ✓ ✓ 93.41 98.85 98.85 93.87 96.15 1.04 174.3 ✗ ✗ ✗ ✗ 91.67 98.84 98.84 91.69 90.73 1.04 1812.6 ✓ ✗ ✗ ✗ 93.46 98.07 99.61 94.26 92.28 1.09 2254.6 ✓ ✓ ✗ ✗ 94.07 99.22 99.22 95.06 91.09 1.03 1624.5 RL ✓ ✓ ✓ ✗ 89.51 97.27 98.44 90.93 83.20 1.05 5424.3 ✓ ✓ ✓ ✓ 88.66 95.54 98.84 92.82 86.05 1.21 1928.1 Table 4. Effect of different components on IL and RL loss using our CarPlanner. Results are in Test14-random non-reactive benchmark. Metric Score CLS-NR 47.40 S-CR 100.0 S-TTC 0.0 S-Area 100.0 S-PR 31.68 Metric Score CLS-NR 47.50 S-CR 100.0 S-TTC 0.0 S-Area 100.0 S-PR 31.98 Metric Score CLS-NR 95.14 S-CR 100.0 S-TTC 100.0 S-Area 100.0 S-PR 85.07 tsim = 11s tsim = 12s tsim = 10s tsim = 9s tsim = 0s CarPlanner-IL (Ours) CarPlanner (Ours) PDM-Closed Figure 3. Qualitative comparison of PDM-Closed and our method in non-reactive environments. The scenario is annotated as waiting for pedestrian to cross. In each frame shot, ego vehicle is marked as green. Traffic agents are marked as sky blue. Lineplot with blue is the ego planned trajectory. unaware of this situation and takes an emergency stop, but it still intersects with this pedestrian. In contrast, our IL vari- ant displays an awareness of the pedestrian’s movements and consequently conducts a braking maneuver. However, it still remains close to the pedestrian. Our RL method avoids this hazard by starting up early up to tsim = 9s and achieves the highest progress and safety metrics. 5. Conclusion In this paper, we introduce CarPlanner, a consistent auto-regressive planner aiming at large-scale RL training. Thanks to the proposed framework, we train an RL-based planner that outperforms existing RL-, IL-, and rule-based SOTAs. Furthermore, we provide analysis indicating the characteristics of IL and RL, highlighting the potential of RL to take a further step toward learning-based planning. Limitations and future work. RL needs delicate design and is prone to input representation. RL can overfit its train- ing environment and suffer from performance drop in un- seen environments [21]. Our method leverages expert-aided reward design to guide exploration. However, this approach may constrain the full potential of RL, as it inherently relies on expert demonstrations and may hinder the discovery of solutions that surpass human expertise. Future work aims to develop robust RL algorithms capable of overcoming these limitations, enabling autonomous exploration and general- ization across diverse environments.
  • 9.
    6. Acknowledgement Many thanksto Jingke Wang for helpful discussions and all reviewers for improving the paper. This work was sup- ported by Zhejiang Provincial Natural Science Foundation of China under Grant No. LD24F030001, and by the Na- tional Nature Science Foundation of China under Grant 62373322. References [1] Mayank Bansal, Alex Krizhevsky, and Abhijit Ogale. Chauf- feurnet: Learning to drive by imitating the best and synthe- sizing the worst. In Proc. of Robotics: Science and Systems, 2019. 3, 7 [2] Holger Caesar, Juraj Kabzan, Kok Seang Tan, Whye Kit Fong, Eric Wolff, Alex Lang, Luke Fletcher, Oscar Beijbom, and Sammy Omari. nuplan: A closed-loop ml-based plan- ning benchmark for autonomous vehicles. arXiv preprint arXiv:2106.11810, 2021. 2, 6, 12 [3] Jie Cheng, Yingbing Chen, and Qifeng Chen. Pluto: Push- ing the limit of imitation learning-based planning for au- tonomous driving. arXiv preprint arXiv:2404.14327, 2024. 3, 6, 7, 12, 13 [4] Jie Cheng, Yingbing Chen, Xiaodong Mei, Bowen Yang, Bo Li, and Ming Liu. Rethinking imitation-based planners for autonomous driving. In Proc. of the IEEE Intl. Conf. on Robotics & Automation, pages 14123–14130. IEEE, 2024. 3, 6, 7 [5] Kashyap Chitta, Aditya Prakash, Bernhard Jaeger, Zehao Yu, Katrin Renz, and Andreas Geiger. Transfuser: Imitation with transformer-based sensor fusion for autonomous driv- ing. IEEE Trans. on Pattern Analysis and Machine Intelli- gence (TPAMI), 2022. 3 [6] Ignasi Clavera, Yao Fu, and Pieter Abbeel. Model- augmented actor-critic: Backpropagating through paths. In Proc. of the Int. Conf. on Learning Representations, 2020. 15 [7] Daniel Dauner, Marcel Hallgarten, Andreas Geiger, and Kashyap Chitta. Parting with misconceptions about learning- based vehicle motion planning. In Proc. of the Conf. on Robot Learning, 2023. 6 [8] Pim De Haan, Dinesh Jayaraman, and Sergey Levine. Causal confusion in imitation learning. Proc. of the Advances in Neural Information Processing Systems, 32, 2019. 1, 3 [9] Scott Ettinger, Shuyang Cheng, Benjamin Caine, Chenxi Liu, Hang Zhao, Sabeek Pradhan, Yuning Chai, Ben Sapp, Charles R Qi, Yin Zhou, et al. Large scale interactive motion forecasting for autonomous driving: The waymo open mo- tion dataset. In Proc. of the IEEE/CVF Intl. Conf. on Com- puter Vision, pages 9710–9719, 2021. 2, 12 [10] Jiyang Gao, Chen Sun, Hang Zhao, Yi Shen, Dragomir Anguelov, Congcong Li, and Cordelia Schmid. Vectornet: Encoding hd maps and agent dynamics from vectorized rep- resentation. In Proc. of the IEEE/CVF Conf. on Computer Vision and Pattern Recognition, pages 11525–11533, 2020. 3 [11] Ke Guo, Wei Jing, Junbo Chen, and Jia Pan. CCIL: Context- conditioned imitation learning for urban driving. In Proc. of Robotics: Science and Systems, 2023. 1, 7 [12] Marcel Hallgarten, Martin Stoll, and Andreas Zell. From prediction to planning with goal conditioned lane graph traversals. arXiv preprint arXiv:2302.07753, 2023. 6 [13] Nicklas A Hansen, Hao Su, and Xiaolong Wang. Tem- poral difference learning for model predictive control. In Proc. of the Intl. Conf. on Machine Learning, pages 8387– 8406. PMLR, 2022. 15 [14] Xiangkun He, Haohan Yang, Zhongxu Hu, and Chen Lv. Robust lane change decision making for autonomous vehi- cles: An observation adversarial reinforcement learning ap- proach. IEEE Transactions on Intelligent Vehicles, 8(1):184– 193, 2022. 3 [15] John Houston, Guido Zuidhof, Luca Bergamini, Yawei Ye, Long Chen, Ashesh Jain, Sammy Omari, Vladimir Iglovikov, and Peter Ondruska. One thousand and one hours: Self-driving motion prediction dataset. In Proc. of the Conf. on Robot Learning, pages 409–418, 2021. 2 [16] Yihan Hu, Jiazhi Yang, Li Chen, Keyu Li, Chonghao Sima, Xizhou Zhu, Siqi Chai, Senyao Du, Tianwei Lin, Wenhai Wang, et al. Planning-oriented autonomous driving. In Proc. of the IEEE/CVF Conf. on Computer Vision and Pat- tern Recognition, pages 17853–17862, 2023. 1 [17] Zhiyu Huang, Haochen Liu, and Chen Lv. Gameformer: Game-theoretic modeling and learning of transformer-based interactive prediction and planning for autonomous driving. In Proc. of the IEEE/CVF Intl. Conf. on Computer Vision, pages 3903–3913, 2023. 2, 3, 6, 13 [18] Zhiyu Huang, Xinshuo Weng, Maximilian Igl, Yuxiao Chen, Yulong Cao, Boris Ivanovic, Marco Pavone, and Chen Lv. Gen-drive: Enhancing diffusion generative driving poli- cies with reward modeling and reinforcement learning fine- tuning. arXiv preprint arXiv:2410.05582, 2024. 2, 6, 12 [19] Julian Ibarz, Jie Tan, Chelsea Finn, Mrinal Kalakrishnan, Peter Pastor, and Sergey Levine. How to train your robot with deep reinforcement learning: lessons we have learned. Intl. Journal of Robotics Research, 40(4-5):698–721, 2021. 1 [20] Chiyu Jiang, Andre Cornman, Cheolho Park, Benjamin Sapp, Yin Zhou, Dragomir Anguelov, et al. Motiondiffuser: Controllable multi-agent motion prediction using diffusion. In Proc. of the IEEE/CVF Conf. on Computer Vision and Pat- tern Recognition, pages 9644–9653, 2023. 2 [21] Robert Kirk, Amy Zhang, Edward Grefenstette, and Tim Rocktäschel. A survey of zero-shot generalisation in deep reinforcement learning. Journal of Artificial Intelligence Re- search, 76:201–264, 2023. 8 [22] Edouard Leurent and Jean Mercat. Social attention for au- tonomous decision-making in dense traffic. arXiv preprint arXiv:1911.12250, 2019. 3 [23] Guofa Li, Yifan Yang, Shen Li, Xingda Qu, Nengchao Lyu, and Shengbo Eben Li. Decision making of autonomous ve- hicles in lane change scenarios: Deep reinforcement learn- ing approaches with risk awareness. Transportation research part C: emerging technologies, 134:103452, 2022. 3
  • 10.
    [24] Quanyi Li,Zhenghao Mark Peng, Lan Feng, Zhizheng Liu, Chenda Duan, Wenjie Mo, and Bolei Zhou. Scenarionet: Open-source platform for large-scale traffic scenario simula- tion and modeling. Proc. of the Advances in Neural Infor- mation Processing Systems, 36:3894–3920, 2023. 12 [25] Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Car- roll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. Training language models to follow instructions with human feedback. Proc. of the Advances in Neural Information Processing Systems, 35: 27730–27744, 2022. 1 [26] Charles R Qi, Hao Su, Kaichun Mo, and Leonidas J Guibas. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proc. of the IEEE/CVF Conf. on Com- puter Vision and Pattern Recognition, pages 652–660, 2017. 4 [27] Nicholas Rhinehart, Rowan McAllister, Kris Kitani, and Sergey Levine. Precog: Prediction conditioned on goals in visual multi-agent settings. In Proc. of the IEEE/CVF Intl. Conf. on Computer Vision, pages 2821–2830, 2019. 2 [28] Stéphane Ross, Geoffrey Gordon, and Drew Bagnell. A re- duction of imitation learning and structured prediction to no- regret online learning. In Proceedings of the fourteenth inter- national conference on artificial intelligence and statistics, pages 627–635, 2011. 1, 3 [29] Oliver Scheel, Luca Bergamini, Maciej Wolczyk, Błażej Osiński, and Peter Ondruska. Urban driver: Learning to drive from real-world demonstrations using policy gradients. In Proc. of the Conf. on Robot Learning, pages 718–728, 2022. 1, 3, 6 [30] John Schulman, Philipp Moritz, Sergey Levine, Michael Jor- dan, and Pieter Abbeel. High-dimensional continuous con- trol using generalized advantage estimation. arXiv preprint arXiv:1506.02438, 2015. 11, 12 [31] John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Rad- ford, and Oleg Klimov. Proximal policy optimization algo- rithms. arXiv preprint arXiv:1707.06347, 2017. 6, 11 [32] Ari Seff, Brian Cera, Dian Chen, Mason Ng, Aurick Zhou, Nigamaa Nayakanti, Khaled S Refaat, Rami Al-Rfou, and Benjamin Sapp. Motionlm: Multi-agent motion forecast- ing as language modeling. In Proc. of the IEEE/CVF Intl. Conf. on Computer Vision, pages 8579–8590, 2023. 2 [33] Shaoshuai Shi, Li Jiang, Dengxin Dai, and Bernt Schiele. Motion transformer with global intention localization and lo- cal movement refinement. Proc. of the Advances in Neural Information Processing Systems, 35:6531–6543, 2022. 2 [34] David Silver, Aja Huang, Chris J Maddison, Arthur Guez, Laurent Sifre, George Van Den Driessche, Julian Schrit- twieser, Ioannis Antonoglou, Veda Panneershelvam, Marc Lanctot, et al. Mastering the game of go with deep neural networks and tree search. nature, 529(7587):484–489, 2016. 1 [35] Simon Suo, Sebastian Regalado, Sergio Casas, and Raquel Urtasun. Trafficsim: Learning to simulate realistic multi- agent behaviors. In Proc. of the IEEE/CVF Conf. on Com- puter Vision and Pattern Recognition, pages 10400–10409, 2021. 13 [36] Ardi Tampuu, Tambet Matiisen, Maksym Semikin, Dmytro Fishman, and Naveed Muhammad. A survey of end-to-end driving: Architectures and training methods. IEEE Trans- actions on Neural Networks and Learning Systems, 33(4): 1364–1384, 2020. 1 [37] Martin Treiber, Ansgar Hennecke, and Dirk Helbing. Con- gested traffic states in empirical observations and micro- scopic simulations. Physical review E, 62(2):1805, 2000. 6 [38] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszko- reit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Proc. of the Advances in Neural Information Processing Systems, 30, 2017. 4 [39] Huanjie Wang, Shihua Yuan, Mengyu Guo, Xueyuan Li, and Wei Lan. A deep reinforcement learning-based approach for autonomous driving in highway on-ramp merge. Pro- ceedings of the Institution of Mechanical engineers, Part D: Journal of Automobile engineering, 235(10-11):2726–2739, 2021. 3 [40] Letian Wang, Jie Liu, Hao Shao, Wenshuo Wang, Ruobing Chen, Yu Liu, and Steven L Waslander. Efficient Reinforce- ment Learning for Autonomous Driving with Parameterized Skills and Priors. In Proc. of Robotics: Science and Systems, Daegu, Republic of Korea, 2023. 3 [41] Wei Wu, Xiaoxin Feng, Ziyan Gao, and Yuheng Kan. Smart: Scalable multi-agent real-time simulation via next-token pre- diction. arXiv preprint arXiv:2405.15677, 2024. 2 [42] Dongkun Zhang, Jiaming Liang, Sha Lu, Ke Guo, Qi Wang, Rong Xiong, Zhenwei Miao, and Yue Wang. Pep: Policy- embedded trajectory planning for autonomous driving. IEEE Robotics and Automation Letters, 2024. 3, 6 [43] Zhejun Zhang, Alexander Liniger, Christos Sakaridis, Fisher Yu, and Luc V Gool. Real-time motion prediction via het- erogeneous polyline transformer with relative pose encoding. Proc. of the Advances in Neural Information Processing Sys- tems, 36, 2024. 13 [44] Tong Zhou, Letian Wang, Ruobing Chen, Wenshuo Wang, and Yu Liu. Accelerating reinforcement learning for au- tonomous driving using task-agnostic and ego-centric mo- tion skills. In Proc. of the IEEE/RSJ Intl. Conf. on Intelligent Robots and Systems, pages 11289–11296. IEEE, 2023. 3 [45] Yang Zhou, Hao Shao, Letian Wang, Steven L Waslander, Hongsheng Li, and Yu Liu. Smartrefine: A scenario-adaptive refinement framework for efficient motion prediction. In Proc. of the IEEE/CVF Conf. on Computer Vision and Pat- tern Recognition, pages 15281–15290, 2024. 2 [46] Zikang Zhou, Haibo Hu, Xinhong Chen, Jianping Wang, Nan Guan, Kui Wu, Yung-Hui Li, Yu-Kai Huang, and Chun Jason Xue. Behaviorgpt: Smart agent simulation for autonomous driving with next-patch prediction. arXiv preprint arXiv:2405.17372, 2024. 2
  • 11.
    CarPlanner: Consistent Auto-regressiveTrajectory Planning for Large-scale Reinforcement Learning in Autonomous Driving Supplementary Material A. Training Procedure Algorithm 1 outlines the training process for the CarPlanner framework. Notably, during the training of the trajectory generator, we have the flexibility to employ either RL or IL, but, in this work, we do not combine RL and IL simultane- ously, opting instead to explore their distinct characteristics separately. The definitions of the loss functions are given in the following. Loss of non-reactive transition model. The non-reactive transition model β is trained to simulate agent trajecto- ries based on the initial state s0. For each data sample (s0, s1:N,gt 1:T ) ∈ D, the model predicts trajectories s1:N 1:T = β(s0), and the training objective minimizes the L1 loss: L_{text {tm}} = frac {1}{T} sum _{t=1}^T sum _{n=1}^N left | s_t^n - s_t^{n,text {gt}} right |_1. (5) Mode selector loss. This contains two parts: cross-entropy and side-task loss. The cross-entropy loss is defined as: text {CrossEntropyLoss}(boldsymbol {sigma }, c^*) = -sum _{i=1}^{N_{text {mode}}} mathbb {I}(c_i = c^*) log sigma _i, (6) where σi is the assigned score for mode ci, Nmode is the number of candidate modes, and I is the indicator function. The side-task loss is defined as: text {SideTaskLoss}(bar {s}^{0}_{1:T}, s^{0,text {gt}}_{1:T}) = frac {1}{T} sum _{t=1}^T left | bar {s}_t^0 - s_t^{0,text {gt}} right |_1, (7) where s̄0 t is the output ego future trajectory. Generator loss with RL. The PPO [31] loss consists of three parts: policy, value, and entropy loss. The policy loss is defined as: begin {split} &text {PolicyLoss}(a_{0:T-1}, d_{0:T-1, text {new}}, d_{0:T-1}, A_{0:T-1}) = &- frac {1}{T} sum _{t=0}^{T-1} min left ( r_t A_t, text {clip}(r_t, 1-epsilon , 1+epsilon ) A_t right ), end {split} (8) Algorithm 1 Training Procedure of CarPlanner 1: Input: Dataset D containing initial states s0 and ground-truth trajectories s0:N,gt 1:T , longitudinal modes clon, discount factor γ, GAE parameter λ, update interval I. 2: Require: Non-reactive transition model β, mode selector fselector, policy π, policy old πold. 3: Step 1: Training Transition Model 4: for (s0, s1:N,gt 1:T ) ∈ D do 5: Simulate agent trajectories s1:N 1:T ← β(s0) 6: Calculate loss Ltm ← L1Loss(s1:N 1:T , s1:N,gt 1:T ) 7: Backpropagate and update β using Ltm 8: end for 9: Step 2: Training Selector and Generator 10: Initialize training step ← 0 11: Initialize policy old πold ← π 12: for (s0, s0,gt 1:T ) ∈ D do 13: Non-Reactive Transition Model: 14: Simulate agent trajectories s1:N 1:T ← β(s0) 15: Mode Assignment: 16: Determine clat based on s0 17: Concatenate clat and clon to get c 18: Determine positive mode c∗ based on s0,gt 1:T and c 19: Mode Selector Loss: 20: Compute scores σ, s̄0 1:T ← fselector(s0, c) 21: Lselector ← CrossEntropyLoss(σ, c∗ ) + SideTaskLoss(s̄0 1:T , s0,gt 1:T ) 22: Generator Loss: 23: if Reinforcement Learning (RL) Training then 24: Use πold, s0, c∗ , and s1:N 1:T to collect rollout data (s0:T −1, a0:T −1, d0:T −1, V0:T −1, R0:T −1) 25: Compute advantage A0:T −1 and return R̂0:T −1 using GAE [30]: A0:T −1, R̂0:T −1 ← GAE(R0:T −1, V0:T −1, γ, λ) 26: Compute policy distribution and value estimates: (d0:T −1,new, V0:T −1,new) ← π(s0:T −1, a0:T −1, c∗ ) 27: Lgenerator ← ValueLoss(V0:T −1,new, R̂0:T −1) + PolicyLoss(d0:T −1,new, d0:T −1, A0:T −1) − Entropy(d0:T −1,new) 28: else if Imitation Learning (IL) Training then 29: Use π, s0, c∗ , and s1:N 1:T to collect action sequence a0:T −1 30: Stack action sequence as ego-planned trajectory s0 1:T ← Stack(a0:T −1) 31: Lgenerator ← L1Loss(s0 1:T , s0,gt 1:T ) 32: end if 33: Overall Loss: 34: L ← Lselector + Lgenerator 35: Backpropagate and update fselector, π using L 36: Policy Update: 37: Increment training step ← training step +1 38: if training step % I == 0 then 39: Update πold ← π 40: end if 41: end for
  • 12.
    where the ratiort is given by rt = Prob(at,dt,new) Prob(at,dt) , dt,new and dt are the policy distributions (mean and standard devi- ation of Gaussian distribution) at time step t induced by π and πold respectively, the function Prob(a, d) calculates the probability of a given action a under a distribution d, and At is the advantage estimated using GAE [30]. The value and entropy loss are defined as: text {ValueLoss}(V_{0:T-1, text {new}}, hat {R}_{0:T-1}) = frac {1}{T} sum _{t=0}^{T-1} left | V_{t,text {new}} - hat {R}_t right |_2^2, (9) text {Entropy}(d_{0:T-1, text {new}}) = frac {1}{T} sum _{t=0}^{T-1} mathcal {H}(d_{t,text {new}}), (10) where Vt,new and R̂t are the predicted and actual returns, and H represents the entropy of the policy distribution d. Generator loss with IL. In IL, the generator minimizes the trajectory error between the ego-planned trajectory s0 1:T and the ground-truth trajectory s0,gt 1:T . The loss is defined as: L_{text {generator}} = frac {1}{T} sum _{t=1}^T left | s_t^0 - s_t^{0,text {gt}} right |_1. (11) B. Implementation Details The hyperparameters of model architecture, PPO-related parameters, and loss weights are summarized in Tab. 5. The magnitudes of value, policy, and entropy loss are 103 , 100 , and 10−3 , respectively. The trajectory generator generates trajectories with a time horizon of 8 seconds at 1-second intervals, corresponding to time horizon T = 8. During testing, these trajectories are interpolated to 0.1-second in- tervals. The weight of scores generated by the rule and mode selectors is set to a ratio of 1 : 0.3. In cases where no ego candidate trajectory satisfies the safety criteria eval- uated by the rule selector, an emergency stop is triggered. For the Test14-Random benchmark, a replanning frequency of 10Hz is employed, adhering to the official nuPlan sim- ulation configuration. In contrast, for the Reduced-Val14 benchmark, a replanning frequency of 1Hz is used to en- sure a fair comparison with Gen-Drive [18]. C. Ablation Study on RL Training In this part, we examine the training efficiency of CarPlan- ner, performance of vanilla and consistent auto-regressive frameworks, the use of reactive and non-reactive model in RL training, and the impact of varying the time horizon. Training efficiency. We compare the efficiency of our model-based framework with that of ScenarioNet [24], which is an open-source platform for model-free RL train- ing in real-world datasets [2, 9]. As shown in Tab. 6, Car- Planner achieves a remarkable improvement in sampling ef- Parameter Value Feature dimension D 256 Static point dimension Dm 9 Agent pose dimension Da 10 Activation ReLU Number of layers 3 Number of attention heads 8 Dropout 0.1 discount factor γ 0.1 GAE parameter λ 0.9 Clip range ϵ 0.2 Update interval I 8 Weight of selector loss 1 Weight of value loss 3 Weight of policy loss 100 Weight of entropy loss 0.001 Weight of IL loss 1 Table 5. Hyperparameters of model architecture, PPO-related pa- rameters, and loss weights. Planner CLS-NR (↑) Efficiency (samples/sec, ↑) Num. Samples Train Time ScenarioNet [24] 55.60 25.72 7,798,472 3d12h11m38s CarPlanner-IL 93.41 1,181.46 70,487,200 16h34m12s CarPlanner 94.07 1,632.25 70,487,200 11h59m44s Table 6. Comparison of training efficiency with model-free set- tings. Experimental results are based on the Test14-Random non- reactive benchmark. Design Choices Closed-loop metrics (↑) Model Type Random Sample Guide Reward CLS-NR S-CR S-Area S-PR S-Comfort Vanilla ✓ Progress 67.56 ± 0.38 90.97 ± 0.78 94.64 ± 1.72 72.17 ± 0.21 64.21 ± 1.29 ✓ DE 86.89 ± 0.28 97.34 ± 0.37 96.36 ± 0.18 89.90 ± 0.11 94.03 ± 0.65 Consistent ✗ FE 88.14 96.86 98.43 91.39 73.73 ✗ DE 94.07 99.22 99.22 95.06 91.09 Table 7. Comparison of vanilla and consistent auto-regressive frameworks with different guide reward design. Experimental re- sults are based on the Test14-Random non-reactive benchmark. Model Loss CLS-NR (↑) Consistent Ratio Lat (↑) Consistent Ratio Lon (↑) Vanilla RL 86.89 ± 0.28 20.00 ± 0.10 8.33 ± 0.00 PLUTO [3] IL 91.92 62.45 41.80 Consistent IL 93.41 68.26 43.01 Consistent RL 94.07 79.58 43.03 Table 8. Comparison for consistency. Experimental results are based on the Test14-Random non-reactive benchmark. ficiency, outperforming ScenarioNet by two orders of mag- nitude. Furthermore, CarPlanner not only excels in ef-
  • 13.
    Closed-loop metrics (↑) TransitionModel CLS-NR S-CR S-Area S-PR S-Comfort Reactive 91.03 96.92 99.23 91.28 90.00 Non-reactive 94.07 99.22 99.22 95.06 91.09 Table 9. Comparison of the usage of reactive and non-reactive transition models. Experimental results are based on the Test14- Random non-reactive benchmark. 7LPH+RUL]RQ7UDLQ
  • 14.
  • 15.
    /615 Figure 4. Performanceof different training time horizons under different testing time horizons. The value in each cell is the CLS- NR metric on the Test14-Random non-reactive benchmark. ficiency but also achieves SOTA performance, surpassing ScenarioNet by a wide margin. Vanilla vs. consistent auto-regressive framework. The results are shown in Tabs. 7 and 8. The consistent auto- regressive framework generates multi-modal trajectories by conditioning on mode representations. In contrast, the vanilla framework relies on random sampling from the ac- tion Gaussian distribution to produce multi-modal trajec- tories. To ensure comparability in the number of modes generated by both frameworks, we sample 60 trajectories in parallel for the vanilla framework. Given that random sam- pling introduces variability, we average the results across 3 random seeds. For the consistent framework, we use dis- placement error (DE) and final error (FE) as guide func- tions to assist the policy in generating mode-aligned trajec- tories. For the vanilla framework, DE is compared against a progress reward, which encourages longitudinal movement along the route while discouraging excessive lateral devia- tions that move the vehicle too far from any possible route. The consistent ratio computes the ratio of generated trajec- tories that fall in their corresponding modes in longitudinal and lateral directions separately. Overall, our proposed consistent framework outperforms the vanilla framework in terms of closed-loop performance, highlighting the benefits of incorporating consistency. Fur- thermore, RL provides more consistant trajectories than the vanilla framework and IL-based methods. Additionally, we find that DE serves as an effective guide function for policy training, further enhancing closed-loop performance. Reactive vs. non-reactive transition model. We compare the performance of the CarPlanner framework when trained with reactive and non-reactive transition models. The reac- tive transition model shares a similar architecture with the auto-regressive planner for the ego vehicle, utilizing rela- tive pose encoding [43] as the backbone network to extract features of traffic agents and predict their subsequent poses. The training loss and hyperparameters are consistent with those used for the non-reactive transition model. As shown in Tab. 9, except for the S-Area metric, using non-reactive transition model outperforms the reactive transition model in our current implementation. The primary difference lies in the assumptions about traffic agents: the reactive transi- tion model assumes that the ego vehicle can negotiate with traffic agents and share the same priority, whereas in the non-reactive model, traffic agents do not respond to the ego vehicle, effectively assigning them higher priority. A repre- sentative example is presented in Fig. 5. When trained with the reactive transition model, the planner assumes pedestri- ans will yield to the vehicle, leading it to attempt to move forward. However, at tsim = 12s, the planner collides with pedestrians, triggering an emergency brake, which nega- tively impacts safety, progress, and comfort metrics. Al- though the performance of using reactive transition model is not satisfied currently, it is a more realistic assumption and we will further investigate this in future work. Time horizon. We evaluate the CarPlanner framework by training it with different time horizons, including 1, 3, 5, and 8 seconds, and testing the planners in each time hori- zon. The results in Fig. 4 confirm that increasing the time horizon has a positive effect on the performance for both training and testing. A special case is when the training time horizon is set to 1, all tested time horizons exhibit poor per- formance, highlighting the importance of multi-step learn- ing in RL. Additionally, the observation that increasing the training time horizon enhances closed-loop performance suggests the potential for further improvements by extend- ing the time horizon beyond 8 seconds. However, due to current limitations in data preparation, which is designed for horizons up to 8 seconds, expanding the time horizon would not provide map information or ground-truth trajec- tories, hindering further analysis. Consequently, we leave this exploration for future work. D. Comparison with Differentiable Loss In typical IL setting, the supervision signal provided to the trajectory generator is the displacement error (DE) be- tween the ego-planned trajectory and the ground-truth tra- jectory. Several works [3, 17, 35] propose to convert non-
  • 16.
    tsim = 11stsim = 12s tsim = 10s tsim = 0s tsim = 13s Non-reactive Reactive Metric Score CLS-NR 0.0 S-CR 0.0 S-TTC 0.0 S-Area 100.0 S-PR 61.16 S-Comfort 0.0 Metric Score CLS-NR 96.78 S-CR 100.0 S-TTC 100.0 S-Area 100.0 S-PR 91.7 S-Comfort 100.0 Figure 5. Qualitative comparison of using reactive and non-reactive transition model in non-reactive environments. The scenario is annotated as waiting for pedestrian to cross. In each frame shot, ego vehicle is marked as green. Traffic agents are marked as sky blue. Lineplot with blue is the ego planned trajectory. Supervision Signals Closed-loop metrics (↑) Open-loop metrics (↓) Loss Type DE Col Area CLS-NR S-CR S-Area S-PR S-Comfort Col Mean [Min, Max] Area Mean [Min, Max] IL ✓ ✗ ✗ 93.41 98.85 98.85 93.87 96.15 0.17 [0.00, 0.47] 0.09 [0.00, 0.40] ✓ ✓ ✗ 93.67 99.23 98.85 94.63 94.23 0.16 [0.00, 0.43] 0.07 [0.00, 0.27] ✓ ✗ ✓ 93.12 98.46 98.84 92.88 94.21 0.15 [0.00, 0.44] 0.08 [0.00, 0.30] ✓ ✓ ✓ 93.32 98.46 98.46 94.05 95.77 0.15 [0.00, 0.43] 0.09 [0.00, 0.39] RL ✓ ✗ ✗ 90.44 97.49 96.91 93.33 90.73 0.17 [0.00, 0.49] 0.14 [0.00, 0.51] ✓ ✓ ✓ 94.07 99.22 99.22 95.06 91.09 0.12 [0.00, 0.39] 0.05 [0.00, 0.22] Table 10. Comparison with different loss types and supervision signals. Closed-loop results are based on the Test14-Random non-reactive benchmark. Open-loop results are on validation set. differentiable metrics, such as avoiding collision (Col) and adherence to drivable area (Area), into differentiable loss functions that can directly backpropagate to the generator. In contrast, CarPlanner leverages an RL framework, which introduces surrogate objectives to indirectly optimize these non-differentiable metrics. In this part, we compare these two approaches which provide rich supervision signals to the trajectory generator. The results are summarized in Tab. 10. In IL training, the Col and Area metrics are converted into differentiable loss functions, whereas in RL training, Col and Area are treated as reward functions, contributing to the quality reward as described in the main paper. It is important to note that the implementations for differentiable loss functions and re- ward functions are identical, except that gradient flow is en- abled for differentiable loss functions. The open-loop met- rics compute the Col and Area values across all candidate multi-modal trajectories, with the Mean, Min, and Max re- ferring to the mean, minimum, and maximum values of the Col and Area metrics within the candidate trajectory set. Our findings suggest that incorporating Col loss bene- fits the open-loop Col metric and improves the closed-loop S-CR metrics, thereby enhancing closed-loop performance. However, incorporating Area loss results in better open- loop Area metrics but deteriorates closed-loop performance. Compared to differentiable loss functions, RL with Col and Area as quality rewards yields the trajectory set with the highest overall quality, as evidenced by smaller Mean and Max metrics in open-loop metrics. This improvement can be attributed to RL’s ability to optimize the reward-to-go using surrogate objectives that account for future rewards, while differentiable loss functions are limited to timewise- aligned optimization in our current implementation. This distinction is illustrated in Fig. 6: in (a), the loss at time step t is directly computed from s0 t , meaning that during backward propagation, the loss at time step t cannot influ- ence the optimization of prior time steps. In (b), however, the non-differentiable reward is aggregated into a return (reward-to-go), which serves as a reference for computing the loss at time step t. Through this process, the reward at
  • 17.
    Design Choices Closed-loopmetrics (↑) Loss Type Model Type Mode Type Mode Dropout Scorer Side Task Ego-history Dropout Backbone Sharing CLS-NR S-CR S-Area S-PR S-Comfort IL Vanilla - - - ✓ - 86.48 97.09 97.29 88.05 94.19 Consistent Lon ✗ ✓ ✓ ✓ 88.79 96.67 96.08 89.63 94.90 Consistent Lon-Lat ✓ ✓ ✓ ✓ 93.41 98.85 98.85 93.87 96.15 RL Vanilla - - - ✗ - 85.56 97.27 95.70 89.17 93.36 Consistent Lon ✗ ✓ ✗ ✗ 90.57 97.30 97.68 92.20 94.59 Consistent Lon-Lat ✓ ✓ ✗ ✗ 94.07 99.22 99.22 95.06 91.09 Table 11. Effect of different mode representations. Experimental results are based on the Test14-Random non-reactive benchmark. Planner Planner s0 1 L1 s0 2 L2 … … s0 T LT s0 1 R̂1 R1 L1 s0 2 R̂2 R2 L2 … … … … s0 T R̂T RT LT (a) Differentiable loss (b) RL Forward (with gradient) Forward (without gradient) Backward Figure 6. The computational graph of differentiable loss (a) and RL (b) framework for optimizing same metrics such as displace- ment errors, collision avoidance, and adherence to drivable area. time step t can influence the trajectory at earlier time steps t′ (t′ t). In the future, we aim to combine the advantages of differentiable loss which can provide low-variance gra- dients, and RL which can provide long-term foresight, by model-based RL optimization techniques [6, 13]. E. Effect of Mode Representation In this part, we examine the impact of mode representations on performance. The results are presented in Tab. 11. For both the vanilla and consistent frameworks, we disable the use of random sampling to focus solely on mode-aligned trajectories. As a result, the vanilla framework can only generate single-modal trajectories, leading to the lowest performance. In the consistent framework, we explore two types of mode representations: Lon and Lon-Lat. The Lon representation assigns modes based on longitudinal move- ments along the route, whereas the Lon-Lat representation decomposes modes by both longitudinal and lateral move- ments. Aligned with the main paper, we use ego-history dropout and backbone sharing only for IL training. For the Lon representation, we close mode dropout since it does not rely on any map or agent representation in initial state. The results indicate that introducing consistency provides greater benefits to RL training, with the Lon-Lat represen- tation proving to be more effective than the Lon representa- tion. This suggests that decomposing mode representations into both longitudinal and lateral components enhances the model’s ability by providing more explicit mode informa- tion.