--- language: - en license: mit size_categories: - 10K CUA-Suite Logo

VideoCUA

The largest open, human annotated video corpus for desktop computer use
Part of CUA-Suite: Massive Human-annotated Video Demonstrations for Computer-Use Agents

PaperProject PageGitHubUI-VisionGroundCUA

CUA-Suite Teaser

## Overview **VideoCUA** is the largest open expert video corpus for desktop computer use, comprising **~10K tasks**, **55 hours** of continuous 30 fps screen recordings, and **6 million frames** across **87 professional desktop applications** spanning 12 categories. Unlike sparse screenshot datasets, VideoCUA preserves the full temporal dynamics of human interaction — every mouse movement, click, drag, scroll, and keystroke is logged with millisecond precision alongside continuous video. This enables research in action prediction, imitation learning, visual world models, and video-based reward modeling. VideoCUA is part of [CUA-Suite](https://cua-suite.github.io/), a unified ecosystem that also includes: - [**UI-Vision**](https://uivision.github.io/) — A desktop-centric benchmark evaluating element grounding, layout understanding, and action prediction. - [**GroundCUA**](https://groundcua.github.io/) — A large-scale pixel-precise UI grounding dataset with 5M+ human-verified element annotations. ## Usage To process the raw video data and action logs into trajectories for training or evaluation, you can use the synthesis pipeline provided in the [GitHub repository](https://github.com/ServiceNow/GroundCUA/tree/main/VideoCUA). ### 1. Download & Extract ```bash bash download_data.sh --repo ServiceNow/VideoCUA --output_dir ./VideoCUA ``` ### 2. Convert to Trace Format To extract video frames at each action timestamp and produce standardized trajectories: ```bash python convert_videocua.py \ --data_dir ./VideoCUA/data \ --output_dir ./videocua_processed \ --num_workers 4 ``` ### 3. Generate CoT Annotations ```bash python gen_cot.py \ --task_list_path ./videocua_processed/task_list.json \ --model claude-sonnet-4.5 \ --num_threads 4 \ --suffix cot_v1 ``` ## Repository Structure ``` . ├── assets/ │ ├── cua-suite-logo.png │ └── cua-suite-teaser.png ├── raw_data/ # One zip per application (87 total) │ ├── 7-Zip.zip │ ├── Affine.zip │ ├── Anki.zip │ ├── ... │ └── draw.io.zip └── README.md ``` ## Data Format Each application zip in `raw_data/` contains multiple task folders identified by numeric task IDs. Each task folder has the following structure: ``` / ├── action_log.json # Task metadata and timestamped actions └── video/ ├── video.mp4 # Continuous 30 fps screen recording (1920×1080) └── video_metadata.json # Video properties (fps, duration, resolution, etc.) ``` ### `action_log.json` ```json { "task_id": 45525, "task_instruction": "Open test.7z present in archive it and see the contents", "platform": "7-Zip", "action_log": [ { "action_type": "CLICK", "timestamp": 2.581, "action_params": { "x": 47, "y": 242, "numClicks": 2 }, "groundcua_id": "9a7daeed..." } ] } ``` Each action entry includes a `groundcua_id` field — this is the unique identifier for the corresponding screenshot in the [GroundCUA](https://huggingface.co/datasets/ServiceNow/GroundCUA) repository. Using this ID, you can look up the fully annotated screenshot in GroundCUA, linking the video action trajectory to dense UI grounding annotations. ## Citation If you find VideoCUA or any of the other works in CUA-Suite useful for your research, please cite our works: ```bibtex @inproceedings{ jian2026cuasuite, title={{CUA}-Suite: Massive Human-annotated Video Demonstrations for Computer-Use Agents}, author={Xiangru Jian and Shravan Nayak and Kevin Qinghong Lin and Aarash Feizi and Kaixin Li and Patrice Bechard and Spandana Gella and Sai Rajeswar}, booktitle={ICLR 2026 Workshop on Lifelong Agents: Learning, Aligning, Evolving}, year={2026}, url={https://openreview.net/forum?id=IgTUGrZfMr} } @inproceedings{ feizi2026grounding, title={Grounding Computer Use Agents on Human Demonstrations}, author={Aarash Feizi and Shravan Nayak and Xiangru Jian and Kevin Qinghong Lin and Kaixin Li and Rabiul Awal and Xing Han L{\`u} and Johan Obando-Ceron and Juan A. Rodriguez and Nicolas Chapados and David Vazquez and Adriana Romero-Soriano and Reihaneh Rabbany and Perouz Taslakian and Christopher Pal and Spandana Gella and Sai Rajeswar}, booktitle={The Fourteenth International Conference on Learning Representations}, year={2026}, url={https://openreview.net/forum?id=9WiPZy3Kro} } @inproceedings{ nayak2025uivision, title={{UI}-Vision: A Desktop-centric {GUI} Benchmark for Visual Perception and Interaction}, author={Shravan Nayak and Xiangru Jian and Kevin Qinghong Lin and Juan A. Rodriguez and Montek Kalsi and Nicolas Chapados and M. Tamer {\"O}zsu and Aishwarya Agrawal and David Vazquez and Christopher Pal and Perouz Taslakian and Spandana Gella and Sai Rajeswar}, booktitle={Forty-second International Conference on Machine Learning}, year={2025}, url={https://openreview.net/forum?id=5Rtj4mYH1C} } ``` ## License This dataset is released under the [MIT License](LICENSE).