File size: 3,132 Bytes
ba0dc92 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 | ---
license: cc-by-nc-sa-4.0
language:
- en
tags:
- video-generation
- vision-language-navigation
- embodied-ai
- pytorch
---

# SparseVideoNav: Sparse Video Generation Propels Real-World Beyond-the-View Vision-Language Navigation
## Model Details
### Model Description
SparseVideoNav introduces video generation models to real-world beyond-the-view vision-language navigation for the first time. It pioneers a paradigm shift from continuous to sparse video generation for longer prediction horizons. By guiding trajectory inference with a generated sparse future spanning a 20-second horizon, it achieves sub-second inference (a 27× speed-up). It also marks the first realization of beyond-the-view navigation in challenging night scenes.
- **Developed by:** Hai Zhang, Siqi Liang, Li Chen, Yuxian Li, Yukuan Xu, Yichao Zhong, Fu Zhang, Hongyang Li
- **Shared by:** The University of Hong Kong & OpenDriveLab
- **Model type:** Video Generation-based Model for Vision-Language Navigation
- **Language(s) (NLP):** English (Instruction prompts)
- **License:** CC BY-NC-SA 4.0
- **Finetuned from model:** Based on UMT5-XXL (text encoder) and Wan2.1 VAE.
### Model Sources
- **Repository:** [https://github.com/OpenDriveLab/SparseVideoNav](https://github.com/OpenDriveLab/SparseVideoNav)
- **Paper:** [arXiv:2602.05827](https://arxiv.org/abs/2602.05827)
- **Project Page:** [https://opendrivelab.com/SparseVideoNav](https://opendrivelab.com/SparseVideoNav)
## Uses
### Direct Use
The model is designed for generating sparse future video frames based on a current visual observation (video) and a natural language instruction (e.g., "turn right"). It is primarily intended for research in Embodied AI, specifically Vision-Language Navigation (VLN) in real-world environments.
### Out-of-Scope Use
The model is a research prototype and is not intended for deployment in safety-critical real-world autonomous driving or robotic navigation systems without further extensive testing, safety validation, and fallback mechanisms.
## How to Get Started with the Model
Use the code below to get started with the model using our custom pipeline.
Ensure you have cloned the [GitHub repository](https://github.com/OpenDriveLab/SparseVideoNav) and installed the requirements.
```python
from omegaconf import OmegaConf
from inference import SVNPipeline
# Load configuration
cfg = OmegaConf.load("config/inference.yaml")
cfg.ckpt_path = "/path/to/models/SparseVideoNav-Models" # Path to your downloaded weights
cfg.inference.device = "cuda:0"
# Initialize pipeline
pipeline = SVNPipeline.from_pretrained(cfg)
# Run inference (Returns np.ndarray (T, H, W, C) uint8)
video = pipeline(video="/path/to/input.mp4", text="turn right")
```
## BibTeX
```python
@article{zhang2026sparse,
title={Sparse Video Generation Propels Real-World Beyond-the-View Vision-Language Navigation},
author={Zhang, Hai and Liang, Siqi and Chen, Li and Li, Yuxian and Xu, Yukuan and Zhong, Yichao and Zhang, Fu and Li, Hongyang},
journal={arXiv preprint arXiv:2602.05827},
year={2026}
} |