SparseDrive: End-to-End Autonomous Driving via Sparse Scene Representation
:page_facing_up: Read on arXiv
Overview
SparseDrive by Sun et al. (ICRA 2025) proposes a paradigm shift from dense BEV-based end-to-end driving to fully sparse scene representations. The core argument is that dense BEV grids are wasteful: most cells are empty, and the computation spent on them doesn't contribute to planning quality. SparseDrive represents the entire driving scene using sparse instance features, anchor boxes, and polylines -- handling detection, tracking, prediction, and planning in a unified sparse framework.
The key insight is that motion prediction and planning share three overlooked similarities: both require modeling high-order bidirectional interactions among agents, both need rich semantic and geometric scene understanding, and both are inherently multi-modal problems with uncertainty. SparseDrive exploits these parallels through a symmetric sparse perception module and a parallel motion planner that handles prediction and planning jointly.
SparseDrive surpasses prior SOTA on nuScenes across all metrics, especially the safety-critical collision rate, while maintaining much higher training and inference efficiency than dense alternatives.
Key Contributions
- Fully sparse scene representation: Replaces dense BEV grids with sparse instance features (anchor boxes for agents, polylines for map elements), scaling computation with scene complexity rather than spatial resolution
- Symmetric sparse perception: A unified architecture that handles 3D detection, online mapping, and multi-object tracking through symmetric query-based modules
- Parallel motion planner: Jointly handles motion prediction and ego planning through a shared architecture, exploiting the structural similarities between forecasting other agents and planning ego motion
- Hierarchical planning selection: Multi-modal trajectory generation with safety-aware selection that considers collision avoidance and traffic rule compliance
- State-of-the-art efficiency: Faster training and inference than dense BEV alternatives while achieving better performance
Architecture / Method

Multi-Camera Images (6x)
│
▼
┌──────────────────┐
│ Image Backbone │ (ResNet-50/101)
│ (Multi-View) │
└────────┬─────────┘
│ multi-view features
▼
┌──────────────────────────────────────────────────┐
│ Symmetric Sparse Perception │
│ │
│ ┌──────────────────┐ ┌──────────────────────┐ │
│ │ Agent Queries │ │ Map Queries │ │
│ │ (anchor boxes) │ │ (anchor polylines) │ │
│ └────────┬─────────┘ └────────┬─────────────┘ │
│ │ deformable │ deformable │
│ │ cross-attn │ cross-attn │
│ ▼ ▼ │
│ ┌────────────────┐ ┌────────────────────────┐ │
│ │ 3D Detection │ │ Online Mapping │ │
│ │ + Tracking │ │ (lanes, edges, xwalks) │ │
│ └────────┬───────┘ └────────┬───────────────┘ │
└───────────┼───────────────────┼──────────────────┘
└─────────┬─────────┘
▼
┌──────────────────────────────────────────────────┐
│ Parallel Motion Planner │
│ │
│ ┌────────────────┐ ┌────────────────────┐ │
│ │ Agent Motion │ │ Ego Planning │ │
│ │ Prediction │ │ (multi-modal) │ │
│ │ (multi-modal) │ │ │ │
│ └────────┬───────┘ └────────┬───────────┘ │
│ │ │ │
│ └──────────┬───────────┘ │
│ ▼ │
│ ┌─────────────────────┐ │
│ │ Hierarchical Select │ │
│ │ (safety + progress │ │
│ │ + rule compliance) │ │
│ └────────┬────────────┘ │
└────────────────────┼─────────────────────────────┘
▼
Final Trajectory
SparseDrive consists of three main modules:
1. Image Backbone + Sparse Feature Extraction: Multi-camera images are processed by a backbone (ResNet-50/101). Instead of constructing a dense BEV grid, sparse 3D queries (anchor boxes for agents, anchor polylines for map elements) attend to multi-view image features via deformable cross-attention. Each query extracts features relevant to its spatial location from the images directly.
2. Symmetric Sparse Perception: Agent queries and map queries are processed through symmetric transformer layers. Agent queries predict 3D bounding boxes, velocities, and attributes. Map queries predict polylines (lane boundaries, road edges, crosswalks). Both use the same attention mechanism, allowing weight sharing and unified training. Temporal information is incorporated by propagating queries across frames for tracking.
3. Parallel Motion Planner: This is the core innovation. Rather than predicting agent futures and then planning separately, SparseDrive uses a shared architecture: - Agent motion prediction: Agent queries attend to map context and other agents, producing multi-modal future trajectory distributions - Ego planning: Ego queries attend to the same scene context and produce multi-modal ego trajectory candidates - Hierarchical selection: Multiple trajectory candidates are scored based on safety (collision checking against predicted agent trajectories), progress (distance toward goal), and traffic rule compliance. The best candidate is selected as the final plan
The parallel design means prediction and planning share intermediate representations and can be trained jointly end-to-end.
Results
nuScenes Benchmark
SparseDrive-B (ResNet101, 512×1408 input):
| Metric | UniAD | VAD | SparseDrive-B |
|---|---|---|---|
| Detection mAP | 38.0 | -- | 49.6 (+11.6 vs UniAD) |
| Multi-Object Tracking AMOTA | 35.9 | -- | 50.1 (+14.2 vs UniAD) |
| Online Mapping mAP | 47.6 | 47.6 | 56.2 (+8.6 vs VAD) |
| Planning L2 (m) | 1.03 | 0.72 | 0.58 |
| Collision Rate | 0.31% | 0.21% | 0.06% |
SparseDrive-S (ResNet50, 256×704): Achieves 9.0 FPS versus 1.8 FPS for UniAD.
SparseDrive achieves the best results across all metrics, with a particularly notable 71.4% reduction in collision rate compared to VAD. Training is 4.8× faster (SparseDrive-B) and 7.2× faster (SparseDrive-S) than UniAD, with inference 4.1× and 5.0× faster respectively.
Limitations
- Sparse representation relies on good query initialization; if initial anchors miss objects, they cannot be recovered
- The parallel planner assumes that prediction and planning share sufficient structure for joint modeling; adversarial or highly unpredictable scenarios may violate this assumption
- Evaluated primarily in open-loop on nuScenes; closed-loop performance needs validation
- Multi-modal trajectory selection uses hand-designed scoring criteria rather than learned selection
Connections
- Extends the vectorized paradigm from Vad Vectorized Scene Representation For Efficient Autonomous Driving to a fully sparse framework
- Directly evolved into Sparsedrivev2 End To End Autonomous Driving Via Sparse Scene Representation which adds trajectory vocabulary scaling
- Query-based perception builds on Bevformer Learning Birds Eye View Representation From Multi Camera Images Via Spatiotemporal Transformers (deformable attention)
- Joint prediction-planning design relates to Planning Oriented Autonomous Driving (UniAD) but is fully sparse
- Map encoding connects to Vectornet Encoding Hd Maps And Agent Dynamics From Vectorized Representation (vectorized map representation)
- Evaluated on Nuscenes A Multimodal Dataset For Autonomous Driving
- Benchmarked on Navsim V2 Pseudo Simulation For Autonomous Driving (NAVSIM) in V2