pith. machine review for the scientific record. sign in

arxiv: 2604.07991 · v1 · submitted 2026-04-09 · 💻 cs.CV · cs.MM

Recognition: 2 theorem links

· Lean Theorem

MotionScape: A Large-Scale Real-World Highly Dynamic UAV Video Dataset for World Models

Enze Zhu, Kan Wei, Lei Wang, Xiaoxuan Liu, Yongkang Zou, Zhan Chen, Zile Guo

Authors on Pith no claims yet

Pith reviewed 2026-05-10 17:07 UTC · model grok-4.3

classification 💻 cs.CV cs.MM
keywords UAV video datasetworld models6-DoF trajectoriesdynamic camera motionvideo predictionaerial navigationsemantic annotationembodied intelligence
0
0 comments X

The pith

A new dataset of highly dynamic UAV videos with 6-DoF trajectories and language annotations improves world models' simulation of complex 3D aerial dynamics.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces MotionScape, a collection of over 30 hours of 4K real-world UAV videos featuring rapid 6-DoF camera motions that differ from the smoother patterns in most existing training sets. These videos come paired with automatically recovered camera trajectories and natural language descriptions to create aligned training samples. Experiments demonstrate that models trained with this data better predict physical dynamics and preserve consistency when viewpoints shift sharply, which matters for UAV navigation and planning where sudden movements are common. The work addresses a distribution gap in current data that limits world models from handling unconstrained aerial environments effectively.

Core claim

MotionScape supplies over 30 hours of 4K UAV-view videos with more than 4.5 million frames, each tightly coupled to accurate 6-DoF camera trajectories and fine-grained natural language descriptions. An automated multi-stage pipeline performs CLIP-based filtering, robust visual SLAM for trajectory recovery, temporal segmentation, and large-language-model semantic annotation to produce the aligned samples. When existing world models incorporate this data, they gain improved ability to simulate complex 3D dynamics and handle large viewpoint shifts, supporting better decision-making for UAV agents.

What carries the argument

The MotionScape dataset, built through an automated pipeline that couples raw UAV videos with 6-DoF trajectories recovered via visual SLAM and semantic annotations from language models.

Load-bearing premise

That the main barrier for world models on UAV tasks is the absence of high-dynamic 6-DoF motion patterns in prior training data, and that the automatically generated trajectories and annotations are accurate enough to close the gap without adding new errors.

What would settle it

Retrain a baseline world model on MotionScape versus standard datasets alone, then evaluate prediction error on held-out UAV sequences using metrics for 3D spatiotemporal consistency under rapid viewpoint changes; absence of measurable gains would falsify the improvement claim.

Figures

Figures reproduced from arXiv: 2604.07991 by Enze Zhu, Kan Wei, Lei Wang, Xiaoxuan Liu, Yongkang Zou, Zhan Chen, Zile Guo.

Figure 1
Figure 1. Figure 1: Example output sequence of Cosmos 2.5-2B for video continuation in a highly dynamic UAV scenario. The zoomed-in [PITH_FULL_IMAGE:figures/full_fig_p001_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Distribution of video resolutions and environmen [PITH_FULL_IMAGE:figures/full_fig_p002_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Representative samples illustrating the scene and weather diversity of our dataset, including mountain, indoor, [PITH_FULL_IMAGE:figures/full_fig_p004_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Distribution of video resolutions and environmental conditions in our dataset. [PITH_FULL_IMAGE:figures/full_fig_p006_4.png] view at source ↗
read the original abstract

Recent advances in world models have demonstrated strong capabilities in simulating physical reality, making them an increasingly important foundation for embodied intelligence. For UAV agents in particular, accurate prediction of complex 3D dynamics is essential for autonomous navigation and robust decision-making in unconstrained environments. However, under the highly dynamic camera trajectories typical of UAV views, existing world models often struggle to maintain spatiotemporal physical consistency. A key reason lies in the distribution bias of current training data: most existing datasets exhibit restricted 2.5D motion patterns, such as ground-constrained autonomous driving scenes or relatively smooth human-centric egocentric videos, and therefore lack realistic high-dynamic 6-DoF UAV motion priors. To address this gap, we present MotionScape, a large-scale real-world UAV-view video dataset with highly dynamic motion for world modeling. MotionScape contains over 30 hours of 4K UAV-view videos, totaling more than 4.5M frames. This novel dataset features semantically and geometrically aligned training samples, where diverse real-world UAV videos are tightly coupled with accurate 6-DoF camera trajectories and fine-grained natural language descriptions. To build the dataset, we develop an automated multi-stage processing pipeline that integrates CLIP-based relevance filtering, temporal segmentation, robust visual SLAM for trajectory recovery, and large-language-model-driven semantic annotation. Extensive experiments show that incorporating such semantically and geometrically aligned annotations effectively improves the ability of existing world models to simulate complex 3D dynamics and handle large viewpoint shifts, thereby benefiting decision-making and planning for UAV agents in complex environments. The dataset is publicly available at https://github.com/Thelegendzz/MotionScape

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript introduces MotionScape, a dataset comprising over 30 hours of 4K UAV-view videos (more than 4.5M frames) with semantically and geometrically aligned annotations, including 6-DoF camera trajectories recovered via visual SLAM and fine-grained natural language descriptions generated by LLMs. The automated pipeline combines CLIP-based filtering, temporal segmentation, SLAM trajectory recovery, and LLM annotation. The central claim is that training world models on these aligned samples improves simulation of complex 3D dynamics and handling of large viewpoint shifts for UAV agents.

Significance. If the trajectories prove accurate and the claimed improvements hold under rigorous evaluation, the dataset would address a clear gap in high-dynamic 6-DoF training data for world models, potentially aiding UAV planning and decision-making. The public release of the dataset and processing pipeline constitutes a concrete, reusable contribution.

major comments (2)
  1. [Abstract] Abstract: the statement that 'extensive experiments show that incorporating such semantically and geometrically aligned annotations effectively improves...' is unsupported by any reported quantitative metrics, baselines, ablations, or error analysis, leaving the central empirical claim unverified.
  2. [Dataset construction pipeline] Dataset construction pipeline: the 'robust visual SLAM' step for 6-DoF trajectory recovery reports no quantitative validation (e.g., ATE, RPE, scale consistency, or comparison against GPS/IMU logs), which is load-bearing because monocular SLAM in high-dynamic UAV footage is prone to drift, tracking loss, and scale ambiguity; without these checks the geometric alignment cannot be assumed sufficient to teach genuine 3D dynamics rather than artifacts.
minor comments (2)
  1. [Abstract] Abstract: the total number of distinct trajectories or annotated segments is not stated, which would help readers gauge dataset diversity.
  2. [Dataset availability] Ensure the public GitHub repository includes the full processing code and any filtering thresholds so that the automated pipeline is fully reproducible.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive and detailed review. We address each major comment point by point below, indicating where revisions will be made to strengthen the manuscript.

read point-by-point responses
  1. Referee: [Abstract] Abstract: the statement that 'extensive experiments show that incorporating such semantically and geometrically aligned annotations effectively improves...' is unsupported by any reported quantitative metrics, baselines, ablations, or error analysis, leaving the central empirical claim unverified.

    Authors: We agree that the abstract's empirical claim would be stronger with explicit references to quantitative results. The current manuscript presents experimental results on world model training, but we acknowledge these could be expanded with clearer metrics and ablations. In the revised version we will update the abstract to reference specific findings and augment the experiments section with additional quantitative metrics, baselines, ablations, and error analysis to fully support the claim. revision: yes

  2. Referee: [Dataset construction pipeline] Dataset construction pipeline: the 'robust visual SLAM' step for 6-DoF trajectory recovery reports no quantitative validation (e.g., ATE, RPE, scale consistency, or comparison against GPS/IMU logs), which is load-bearing because monocular SLAM in high-dynamic UAV footage is prone to drift, tracking loss, and scale ambiguity; without these checks the geometric alignment cannot be assumed sufficient to teach genuine 3D dynamics rather than artifacts.

    Authors: We concur that quantitative validation of the recovered 6-DoF trajectories is essential. The manuscript describes the use of a robust visual SLAM pipeline but does not report ATE, RPE, scale consistency checks, or comparisons to GPS/IMU. Because synchronized ground-truth sensor logs were not collected for the majority of sequences, direct quantitative comparison is not possible across the full dataset. In the revision we will add a dedicated subsection on trajectory quality, including qualitative validation (visual inspection, smoothness, and cross-sequence consistency), and an explicit limitations discussion covering potential drift, tracking loss, and scale ambiguity in monocular SLAM under high-dynamic UAV motion. revision: partial

Circularity Check

0 steps flagged

No circularity: dataset release with external empirical validation

full rationale

The paper's core contribution is the release of MotionScape, a new UAV video dataset constructed via an automated pipeline (CLIP filtering, visual SLAM trajectory recovery, LLM semantic annotation). The central claim—that the dataset improves world models on 3D dynamics and viewpoint shifts—is supported by reported experiments measuring performance gains on held-out tasks. No equations, fitted parameters renamed as predictions, or self-citation chains appear in the provided text. The derivation chain is the pipeline itself, which produces new data rather than reducing any result to its own inputs by construction. This is a standard non-circular dataset paper.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

This is a dataset construction paper; the central claim rests on the utility of the collected data and pipeline rather than any mathematical axioms, free parameters, or newly postulated entities. Standard tools (CLIP, visual SLAM, LLMs) are invoked without new assumptions beyond their established performance.

pith-pipeline@v0.9.0 · 5616 in / 1246 out tokens · 75996 ms · 2026-05-10T17:07:08.785862+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

Reference graph

Works this paper leans on

36 extracted references · 11 canonical work pages · 8 internal anchors

  1. [1]

    Arslan Ali, Junjie Bai, Maciej Bala, Yogesh Balaji, Aaron Blakeman, Tiffany Cai, Jiaxin Cao, Tianshi Cao, Elizabeth Cha, Yu-Wei Chao, et al. 2025. World Simulation with Video Foundation Models for Physical AI. arXiv:2511.00062 [cs.CV] https: //arxiv.org/abs/2511.00062

  2. [2]

    Amado Antonini, Winter Guerra, Varun Murali, Thomas Sayre-McCord, and Sertac Karaman. 2018. The blackbird dataset: A large-scale dataset for uav perception in aggressive flight. InInternational Symposium on Experimental Robotics. Springer, 130–139

  3. [3]

    Holger Caesar, Varun Bankiti, Alex H Lang, Sourabh Vora, Venice Erin Liong, Qiang Xu, Anush Krishnan, Yu Pan, Giancarlo Baldan, and Oscar Beijbom. 2020. nuscenes: A multimodal dataset for autonomous driving. 11621–11631 pages

  4. [4]

    Carlos Campos, Richard Elvira, Juan J Gómez Rodríguez, José MM Montiel, and Juan D Tardós. 2021. Orb-slam3: An accurate open-source library for visual, visual–inertial, and multimap slam.IEEE transactions on robotics37, 6 (2021), 1874–1890

  5. [5]

    Haoxin Chen, Yong Zhang, Xiaodong Cun, Menghan Xia, Xintao Wang, Chao Weng, and Ying Shan. 2024. Videocrafter2: Overcoming data limitations for high-quality video diffusion models. 7310–7320 pages

  6. [6]

    Tsai-Shien Chen, Aliaksandr Siarohin, Willi Menapace, Ekaterina Deyneka, Hsiang-wei Chao, Byung Eun Jeon, Yuwei Fang, Hsin-Ying Lee, Jian Ren, Ming- Hsuan Yang, et al . 2024. Panda-70m: Captioning 70m videos with multiple cross-modality teachers. 13320–13331 pages

  7. [7]

    Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xi- aohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al . 2020. An image is worth 16x16 words: Trans- formers for image recognition at scale.arXiv preprint arXiv:2010.11929(2020). https://arxiv.org/abs/2010.11929

  8. [8]

    Michael Fonder and Marc Van Droogenbroeck. 2019. Mid-air: A multi-modal dataset for extremely low altitude drone flights. InProceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops. 0–0

  9. [9]

    Kristen Grauman, Andrew Westbury, Eugene Byrne, Zachary Chavis, Antonino Furnari, Rohit Girdhar, Jackson Hamburger, Hao Jiang, Miao Liu, Xingyu Liu, et al. 2022. Ego4d: Around the world in 3,000 hours of egocentric video. 18995– 19012 pages

  10. [10]

    David Ha and Jürgen Schmidhuber. 2018. World models.arXiv preprint arXiv:1803.101222, 3 (2018), 440

  11. [11]

    2003.Multiple view geometry in computer vision

    Richard Hartley and Andrew Zisserman. 2003.Multiple view geometry in computer vision. Cambridge university press

  12. [12]

    Hao He, Yinghao Xu, Yuwei Guo, Gordon Wetzstein, Bo Dai, Hongsheng Li, and Ceyuan Yang. 2024. Cameractrl: Enabling camera control for text-to-video generation.arXiv preprint arXiv:2404.02101(2024). https://arxiv.org/abs/2404. 02101

  13. [13]

    Will Kay, Joao Carreira, Karen Simonyan, et al. 2017. The Kinetics Human Action Video Dataset. arXiv:1705.06950 [cs.CV] https://arxiv.org/abs/1705.06950

  14. [14]

    Chen-Hsuan Lin, Wei-Chiu Ma, Antonio Torralba, and Simon Lucey. 2021. Barf: Bundle-adjusting neural radiance fields. InProceedings of the IEEE/CVF interna- tional conference on computer vision. 5741–5751

  15. [15]

    Andrew Liu, Richard Tucker, Varun Jampani, Ameesh Makadia, Noah Snavely, and Angjoo Kanazawa. 2021. Infinite nature: Perpetual view generation of natural scenes from a single image. InProceedings of the IEEE/CVF International Conference on Computer Vision. 14458–14467

  16. [16]

    Yixin Liu, Kai Zhang, Yuan Li, et al . 2024. Sora: A Review on Back- ground, Technology, Limitations, and Opportunities of Large Vision Models. arXiv:2402.17177 [cs.CV] https://arxiv.org/abs/2402.17177

  17. [17]

    Antonio Loquercio, Elia Kaufmann, René Ranftl, Matthias Müller, Vladlen Koltun, and Davide Scaramuzza. 2021. Learning high-speed flight in the wild.Science Robotics6, 59 (2021), eabg5810

  18. [18]

    Ye Lyu, George Vosselman, Gui-Song Xia, Alper Yilmaz, and Michael Ying Yang

  19. [19]

    UAVid: A semantic segmentation dataset for UAV imagery.ISPRS journal of photogrammetry and remote sensing165 (2020), 108–119

  20. [20]

    András L Majdik, Charles Till, and Davide Scaramuzza. 2017. The Zurich urban micro aerial vehicle dataset.The International Journal of Robotics Research36, 3 (2017), 269–273

  21. [21]

    Kepan Nan, Rui Xie, Penghao Zhou, et al . 2024. OpenVid-1M: A Large-Scale High-Quality Dataset for Text-to-video Generation. arXiv:2407.02371 [cs.CV] https://arxiv.org/abs/2407.02371

  22. [22]

    OpenAI, :, Aaron Hurst, Adam Lerer, et al . 2024. GPT-4o System Card. arXiv:2410.21276 [cs.CL] https://arxiv.org/abs/2410.21276

  23. [23]

    Abby O’Neill, Abdul Rehman, Abhiram Maddukuri, Abhishek Gupta, Abhishek Padalkar, Abraham Lee, Acorn Pooley, Agrim Gupta, Ajay Mandlekar, Ajinkya Jain, et al. 2024. Open x-embodiment: Robotic learning datasets and rt-x models: Open x-embodiment collaboration 0. (2024), 6892–6903

  24. [24]

    William Peebles and Saining Xie. 2023. Scalable diffusion models with transform- ers. InProceedings of the IEEE/CVF International Conference on Computer Vision. 4195–4205

  25. [25]

    Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. 2021. Learning transferable visual models from natural language supervision. 8748–8763 pages

  26. [26]

    Johannes L Schonberger and Jan-Michael Frahm. 2016. Structure-from-motion revisited. InProceedings of the IEEE conference on computer vision and pattern recognition. 4104–4113

  27. [27]

    Shang Hong Sim, Clarence Lee, Alvin Tan, and Cheston Tan. 2024. Evaluating the generation of spatial relations in text and image generative models.arXiv preprint arXiv:2411.07664(2024). https://arxiv.org/abs/2411.07664

  28. [28]

    Tomás Soucek and Jakub Lokoc. 2024. Transnet v2: An effective deep network architecture for fast shot transition detection. 11218–11221 pages

  29. [29]

    Pei Sun, Henrik Kretzschmar, Xerxes Dotiwalla, Aurelien Chouard, Vijaysai Patnaik, Paul Tsui, James Guo, Yin Zhou, Yuning Chai, Benjamin Caine, et al

  30. [30]

    2446–2454 pages

    Scalability in perception for autonomous driving: Waymo open dataset. 2446–2454 pages

  31. [31]

    Zachary Teed and Jia Deng. 2021. Droid-slam: Deep visual slam for monocular, stereo, and rgb-d cameras. 16558–16569 pages

  32. [32]

    Yi Wang, Yinan He, Yizhuo Li, Kunchang Li, Jiashuo Yu, Xin Ma, Xinhao Li, Guo Chen, Xinyuan Chen, Yaohui Wang, et al. 2023. Internvid: A large-scale video-text dataset for multimodal understanding and generation.arXiv preprint arXiv:2307.06942(2023). https://arxiv.org/abs/2307.06942

  33. [33]

    Zirui Wang, Shangzhe Wu, Weidi Xie, Min Chen, and Victor Adrian Prisacariu

  34. [34]

    NeRF–: Neural radiance fields without known camera parameters. (2021)

  35. [35]

    Zhouxia Wang, Ziyang Yuan, Xintao Wang, Yaowei Li, Tianshui Chen, Menghan Xia, Ping Luo, and Ying Shan. 2024. Motionctrl: A unified and flexible motion controller for video generation. InACM SIGGRAPH 2024 Conference Papers. 1–11

  36. [36]

    Pengfei Zhu, Longyin Wen, Xiao Bian, Haibin Ling, and Qinghua Hu. 2018. Vision meets drones: A challenge.arXiv preprint arXiv:1804.07437(2018). https: //arxiv.org/abs/1804.07437