pith. machine review for the scientific record. sign in

arxiv: 2605.12437 · v1 · submitted 2026-05-12 · 💻 cs.CV

Recognition: 2 theorem links

· Lean Theorem

3D Gaussian Splatting for Efficient Retrospective Dynamic Scene Novel View Synthesis with a Standardized Benchmark

Suryansh Kumar, Yunxiao Zhang

Pith reviewed 2026-05-13 06:26 UTC · model grok-4.3

classification 💻 cs.CV
keywords 3D Gaussian Splattingnovel view synthesisdynamic scenesmulti-view captureretrospective renderingbenchmark datasetsynchronized cameras
0
0 comments X

The pith

Calibrated synchronized multi-view cameras let standard 3D Gaussian Splatting perform efficient retrospective dynamic scene novel view synthesis without any temporal deformation constraint.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper argues that when multiple cameras are calibrated and perfectly time-synchronized, each instant of a dynamic scene is already rigidly constrained by geometry, so explicit motion models or deformation fields are unnecessary. The authors initialize a point cloud from structure-from-motion at the first frame and simply optimize and carry forward the 3D Gaussians to later frames. This produces novel views at any later time without adding temporal coupling. They also supply a Blender-based generator that creates reproducible, training-ready dynamic multi-view datasets with consistent coordinate conventions. The result is a simpler pipeline for applications such as sports replay that rely on recorded multi-camera footage.

Core claim

By initializing the SfM-derived point cloud at the start time and propagating optimized Gaussians over time, efficient retrospective NVS can be achieved without imposing a temporal deformation constraint in a synchronized multi-view setting.

What carries the argument

Forward propagation of 3D Gaussians that are optimized independently at each time step, starting from an initial SfM point cloud, with no added temporal deformation or coupling term.

If this is right

  • Dynamic-scene NVS pipelines become lighter because no separate motion network or deformation field is required.
  • The same 3DGS optimization routine used for static scenes can be reused directly for retrospective dynamic capture.
  • Benchmarking of NeRF and 3DGS methods becomes reproducible once the Blender dataset generator supplies identical camera rigs and coordinate conventions.
  • Applications that record events with fixed multi-camera rigs gain a practical route to high-quality novel-view replay without complex temporal modeling.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The approach may generalize to any capture rig that can enforce sub-frame synchronization and accurate extrinsic calibration.
  • Removing the temporal term reduces the risk of over-smoothing fast motion, which could be tested by comparing rendering sharpness on high-speed actions.
  • The dataset framework could be extended to generate ground-truth depth and segmentation maps, enabling direct quantitative evaluation of geometry quality beyond image metrics.

Load-bearing premise

Calibrated and synchronized multi-view cameras already supply enough spatial consistency that explicit temporal coupling is unnecessary.

What would settle it

A controlled experiment in which the same dynamic scene is captured once with perfect synchronization and once with small time offsets between cameras; if the unsynchronized version shows clear degradation while the synchronized version remains accurate, the claim is supported, and the reverse would falsify it.

Figures

Figures reproduced from arXiv: 2605.12437 by Suryansh Kumar, Yunxiao Zhang.

Figure 1
Figure 1. Figure 1: Contributions: (I) a) The multi-body rigidity constraint holds in a dynamic scene under synchronized multi-view camera setup for each time instance t, b) shows an example NVS results. (II) Proposed Blender API for generating multi-view dynamic scene dataset. In this work, we revisit dynamic scene modeling in a synchronized multi-view setup from the geometric perspec￾tive. We propose a warm-start 3DGS frame… view at source ↗
Figure 2
Figure 2. Figure 2: For complex highly dynamic scene using constraint pro [PITH_FULL_IMAGE:figures/full_fig_p004_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Showing different camera configurations (in black frustum) as well as file format than can be generated using the proposed API. [PITH_FULL_IMAGE:figures/full_fig_p006_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Qualitative comparison with recent TACV [ [PITH_FULL_IMAGE:figures/full_fig_p007_4.png] view at source ↗
read the original abstract

Retrospective novel view synthesis (NVS) of dynamic scenes is fundamental to applications such as sports. Recent dynamic 3D Gaussian Splatting (3DGS) approaches introduce temporally coupled formulations to enforce motion coherence across time. In this paper, we argue that, in a synchronized multi-view (MV) setting typical of sports, the dynamic scene at each time step is already strongly geometrically constrained. We posit that the availability of calibrated, synchronized viewpoints provides sufficient spatial consistency, and therefore, explicit temporal coupling, or complex multi-body constraints seems unnecessary for retrospective NVS. To this end, we propose an approach tailored for synchronized MV dynamic scene. By initializing the SfM-derived point cloud at the start time and propagating optimized Gaussians over time, we show that efficient retrospective NVS can be achieved without imposing a temporal deformation constraint. Complementing our methodological contribution, we introduce a Dynamic MV dataset framework built on Blender for reproducible NeRF and 3DGS research. The framework generates high-quality, synchronized camera rigs and exports training-ready datasets in standard formats, eliminating inconsistencies in coordinate conventions and data pipelines. Using the framework, we construct a dynamic benchmark suite and evaluate representative NeRF and 3DGS approaches under controlled conditions. Together, we show that, under a synchronized MV setup, efficient retrospective dynamic scene NVS can be achieved using 3DGS. At the same time, the dataset-generation framework enables reproducible and principled benchmarking of dynamic NVS methods.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

1 major / 0 minor

Summary. The paper claims that in synchronized multi-view captures of dynamic scenes, per-timestep geometric constraints from calibrated viewpoints suffice for 3D Gaussian Splatting-based retrospective novel view synthesis, rendering explicit temporal deformation fields or multi-body constraints unnecessary. The method initializes an SfM point cloud at the first timestep and propagates optimized Gaussians forward in time solely to warm-start independent per-frame optimizations (no cross-time losses or shared parameters). It also contributes a Blender-based Dynamic MV dataset framework that generates synchronized camera rigs and exports training-ready data in standard formats, then uses this to build a benchmark suite evaluating representative NeRF and 3DGS baselines under controlled conditions.

Significance. If the empirical results hold, the work would be significant for simplifying dynamic NVS pipelines in controlled MV settings such as sports capture, by showing that standard 3DGS per-frame optimization can suffice without added temporal machinery. The dataset-generation framework is a clear strength, directly addressing reproducibility issues in dynamic NeRF/3DGS research through standardized exports and controlled conditions. These elements together support more efficient and falsifiable benchmarking.

major comments (1)
  1. [Abstract and evaluation section] Abstract and evaluation section: the central claim that 'efficient retrospective NVS can be achieved without imposing a temporal deformation constraint' is presented as demonstrated via the benchmark, yet no quantitative metrics (PSNR, SSIM, LPIPS, runtime, or error analysis) or comparisons against temporally-coupled baselines are supplied. This is load-bearing because the argument that spatial consistency alone suffices rests entirely on empirical verification.

Simulated Author's Rebuttal

1 responses · 0 unresolved

We thank the referee for the constructive feedback and for recognizing the significance of the Dynamic MV dataset framework. We address the major comment point by point below.

read point-by-point responses
  1. Referee: [Abstract and evaluation section] Abstract and evaluation section: the central claim that 'efficient retrospective NVS can be achieved without imposing a temporal deformation constraint' is presented as demonstrated via the benchmark, yet no quantitative metrics (PSNR, SSIM, LPIPS, runtime, or error analysis) or comparisons against temporally-coupled baselines are supplied. This is load-bearing because the argument that spatial consistency alone suffices rests entirely on empirical verification.

    Authors: We agree that quantitative empirical verification is essential to support the central claim. The evaluation section of the manuscript reports results from the new benchmark suite, including representative NeRF and 3DGS baselines evaluated under controlled synchronized multi-view conditions. However, we acknowledge that the presentation of specific metrics (PSNR, SSIM, LPIPS), runtime measurements, error analysis, and explicit comparisons to temporally-coupled dynamic 3DGS methods (e.g., those using deformation fields) is not as detailed or prominent as needed. In the revised manuscript we will expand the evaluation section with dedicated quantitative tables, runtime analysis, and direct comparisons against temporally-coupled baselines to more rigorously demonstrate that per-frame 3DGS optimization suffices under strong spatial constraints from synchronized calibrated views. revision: yes

Circularity Check

0 steps flagged

No significant circularity detected

full rationale

The paper's derivation consists of a standard SfM initialization at t=0 followed by per-frame 3DGS optimization with Gaussian propagation used only as a warm-start. No equations reduce a claimed prediction to a fitted input by construction, no temporal deformation parameters are defined in terms of the target result, and no uniqueness theorems or ansatzes are imported via self-citation. The central claim (synchronized MV geometry suffices without explicit temporal coupling) is an empirical and architectural choice, not a self-referential definition. The new benchmark dataset is an independent contribution for evaluation and does not participate in the method's derivation chain. The pipeline remains self-contained against external 3DGS and SfM baselines.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 1 invented entities

Central claim rests on the domain assumption that spatial consistency from synchronized views suffices; introduces one new entity (the dataset framework) with no independent evidence beyond the paper itself.

axioms (1)
  • domain assumption Calibrated synchronized multi-view provides sufficient spatial consistency for dynamic scenes without needing temporal coupling
    Explicitly posited in the abstract as the basis for omitting temporal deformation constraints.
invented entities (1)
  • Dynamic MV dataset framework no independent evidence
    purpose: Generates high-quality synchronized camera rigs and exports training-ready datasets in standard formats for reproducible NeRF and 3DGS research
    New contribution introduced to address inconsistencies in coordinate conventions and data pipelines.

pith-pipeline@v0.9.0 · 5567 in / 1258 out tokens · 40424 ms · 2026-05-13T06:26:58.456808+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Lean theorems connected to this paper

Citations machine-checked in the Pith Canon. Every link opens the source theorem in the public Lean library.

Reference graph

Works this paper leans on

60 extracted references · 60 canonical work pages

  1. [1]

    Neural point-based graph- ics

    Kara-Ali Aliev, Artem Sevastopolsky, Maria Kolos, Dmitry Ulyanov, and Victor Lempitsky. Neural point-based graph- ics. InComputer Vision–ECCV 2020: 16th European Con- ference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXII 16, pages 696–712. Springer, 2020

  2. [2]

    Learning neural light fields with ray-space embedding

    Benjamin Attal, Jia-Bin Huang, Michael Zollh ¨ofer, Johannes Kopf, and Changil Kim. Learning neural light fields with ray-space embedding. InProceedings of the IEEE/CVF Con- ference on Computer Vision and Pattern Recognition, pages 19819–19829, 2022

  3. [3]

    Hyperreel: High-fidelity 6-dof video with ray- conditioned sampling

    Benjamin Attal, Jia-Bin Huang, Christian Richardt, Michael Zollhoefer, Johannes Kopf, Matthew O’Toole, and Changil Kim. Hyperreel: High-fidelity 6-dof video with ray- conditioned sampling. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16610–16620, 2023

  4. [4]

    4d visualization of dynamic events from unconstrained multi-view videos

    Aayush Bansal, Minh V o, Yaser Sheikh, Deva Ramanan, and Srinivasa Narasimhan. 4d visualization of dynamic events from unconstrained multi-view videos. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5366–5375, 2020

  5. [5]

    Mip-nerf: A multiscale representation for anti-aliasing neu- ral radiance fields

    Jonathan T Barron, Ben Mildenhall, Matthew Tancik, Peter Hedman, Ricardo Martin-Brualla, and Pratul P Srinivasan. Mip-nerf: A multiscale representation for anti-aliasing neu- ral radiance fields. InProceedings of the IEEE/CVF Inter- national Conference on Computer Vision, pages 5855–5864, 2021

  6. [6]

    Mip-nerf 360: Unbounded anti-aliased neural radiance fields

    Jonathan T Barron, Ben Mildenhall, Dor Verbin, Pratul P Srinivasan, and Peter Hedman. Mip-nerf 360: Unbounded anti-aliased neural radiance fields. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5470–5479, 2022

  7. [7]

    X-fields: Implicit neural view-, light- and time-image interpolation.ACM Transactions on Graph- ics (TOG), 39(6):1–15, 2020

    Mojtaba Bemana, Karol Myszkowski, Hans-Peter Seidel, and Tobias Ritschel. X-fields: Implicit neural view-, light- and time-image interpolation.ACM Transactions on Graph- ics (TOG), 39(6):1–15, 2020

  8. [8]

    Blender Foundation, Stichting Blender Foundation, Amsterdam, 2022

    Blender Online Community.Blender - a 3D modelling and rendering package. Blender Foundation, Stichting Blender Foundation, Amsterdam, 2022

  9. [9]

    Immersive light field video with a layered mesh representation.ACM Trans- actions on Graphics (TOG), 39(4):86–1, 2020

    Michael Broxton, John Flynn, Ryan Overbeck, Daniel Erick- son, Peter Hedman, Matthew Duvall, Jason Dourgarian, Jay Busch, Matt Whalen, and Paul Debevec. Immersive light field video with a layered mesh representation.ACM Trans- actions on Graphics (TOG), 39(4):86–1, 2020

  10. [10]

    Unstructured lumigraph ren- dering

    Chris Buehler, Michael Bosse, Leonard McMillan, Steven Gortler, and Michael Cohen. Unstructured lumigraph ren- dering. InProceedings of the 28th annual conference on Computer graphics and interactive techniques, pages 425– 432, 2001

  11. [11]

    Hexplane: A fast representa- tion for dynamic scenes

    Ang Cao and Justin Johnson. Hexplane: A fast representa- tion for dynamic scenes. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 130–141, 2023

  12. [12]

    Tensorf: Tensorial radiance fields

    Anpei Chen, Zexiang Xu, Andreas Geiger, Jingyi Yu, and Hao Su. Tensorf: Tensorial radiance fields. InEuropean con- ference on computer vision, pages 333–350. Springer, 2022

  13. [13]

    A neural rendering framework for free-viewpoint relighting

    Zhang Chen, Anpei Chen, Guli Zhang, Chengyuan Wang, Yu Ji, Kiriakos N Kutulakos, and Jingyi Yu. A neural rendering framework for free-viewpoint relighting. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5599–5610, 2020

  14. [14]

    3d gaussian splatting as new era: A survey.IEEE Transactions on Visualization and Computer Graphics, 2024

    Ben Fei, Jingyi Xu, Rui Zhang, Qingyuan Zhou, Weidong Yang, and Ying He. 3d gaussian splatting as new era: A survey.IEEE Transactions on Visualization and Computer Graphics, 2024

  15. [15]

    Signet: Ef- ficient neural representation for light fields

    Brandon Yushan Feng and Amitabh Varshney. Signet: Ef- ficient neural representation for light fields. InProceedings of the IEEE/CVF International Conference on Computer Vi- sion, pages 14224–14233, 2021

  16. [16]

    K-planes: Explicit radiance fields in space, time, and appearance

    Sara Fridovich-Keil, Giacomo Meanti, Frederik Rahbæk Warburg, Benjamin Recht, and Angjoo Kanazawa. K-planes: Explicit radiance fields in space, time, and appearance. In Proceedings of the IEEE/CVF Conference on Computer Vi- sion and Pattern Recognition, pages 12479–12488, 2023

  17. [17]

    Queen: Quantized efficient encoding of dynamic gaussians for streaming free-viewpoint videos.Advances in Neural Information Processing Systems, 37:43435–43467, 2024

    Sharath Girish, Tianye Li, Amrita Mazumdar, Abhinav Shri- vastava, Shalini De Mello, et al. Queen: Quantized efficient encoding of dynamic gaussians for streaming free-viewpoint videos.Advances in Neural Information Processing Systems, 37:43435–43467, 2024

  18. [18]

    Neural implicit dense semantic slam

    Yasaman Haghighi, Suryansh Kumar, Jean-Philippe Thiran, and Luc Van Gool. Neural implicit dense semantic slam. arXiv preprint arXiv:2304.14560, 2023

  19. [19]

    Cambridge university press, 2003

    Richard Hartley.Multiple view geometry in computer vision. Cambridge university press, 2003

  20. [20]

    Ro- bustifying the multi-scale representation of neural radiance fields

    Nishant Jain, Suryansh Kumar, and Luc Van Gool. Ro- bustifying the multi-scale representation of neural radiance fields. In33rd British Machine Vision Conference Proceed- ings, page 578. BMV A Press, 2022

  21. [21]

    Enhanced stable view synthesis

    Nishant Jain, Suryansh Kumar, and Luc Van Gool. Enhanced stable view synthesis. InProceedings of the IEEE/CVF Con- ference on Computer Vision and Pattern Recognition, pages 13208–13217, 2023

  22. [22]

    Learning robust multi-scale representation for neural radiance fields from unposed images.International Journal of Computer Vision, 132(4):1310–1335, 2024

    Nishant Jain, Suryansh Kumar, and Luc Van Gool. Learning robust multi-scale representation for neural radiance fields from unposed images.International Journal of Computer Vision, 132(4):1310–1335, 2024

  23. [23]

    Panoptic studio: A massively multiview system for social motion capture

    Hanbyul Joo, Hao Liu, Lei Tan, Lin Gui, Bart Nabbe, Iain Matthews, Takeo Kanade, Shohei Nobuhara, and Yaser Sheikh. Panoptic studio: A massively multiview system for social motion capture. InProceedings of the IEEE inter- national conference on computer vision, pages 3334–3342, 2015

  24. [24]

    Uncertainty-aware deep multi-view photometric stereo

    Berk Kaya, Suryansh Kumar, Carlos Oliveira, Vittorio Fer- rari, and Luc Van Gool. Uncertainty-aware deep multi-view photometric stereo. InProceedings of the IEEE/CVF con- ference on computer vision and pattern recognition, pages 12601–12611, 2022

  25. [25]

    Neural radiance fields approach to deep multi-view photometric stereo

    Berk Kaya, Suryansh Kumar, Francesco Sarno, Vittorio Fer- rari, and Luc Van Gool. Neural radiance fields approach to deep multi-view photometric stereo. InProceedings of the IEEE/CVF winter conference on applications of computer vision, pages 1965–1977, 2022

  26. [26]

    3d gaussian splatting for real-time 9 radiance field rendering.ACM Trans

    Bernhard Kerbl, Georgios Kopanas, Thomas Leimk ¨uhler, and George Drettakis. 3d gaussian splatting for real-time 9 radiance field rendering.ACM Trans. Graph., 42(4):139–1, 2023

  27. [27]

    3d gaussian splat- ting as markov chain monte carlo.Advances in Neural Infor- mation Processing Systems, 37:80965–80986, 2024

    Shakiba Kheradmand, Daniel Rebain, Gopal Sharma, Wei- wei Sun, Yang-Che Tseng, Hossam Isack, Abhishek Kar, Andrea Tagliasacchi, and Kwang Moo Yi. 3d gaussian splat- ting as markov chain monte carlo.Advances in Neural Infor- mation Processing Systems, 37:80965–80986, 2024

  28. [28]

    Streaming radiance fields for 3d video synthesis

    Lingzhi Li, Zhen Shen, zhongshu wang, Li Shen, and Ping Tan. Streaming radiance fields for 3d video synthesis. In Advances in Neural Information Processing Systems, 2022

  29. [29]

    Neural 3d video synthesis from multi-view video

    Tianye Li, Mira Slavcheva, Michael Zollhoefer, Simon Green, Christoph Lassner, Changil Kim, Tanner Schmidt, Steven Lovegrove, Michael Goesele, Richard Newcombe, et al. Neural 3d video synthesis from multi-view video. In Proceedings of the IEEE/CVF Conference on Computer Vi- sion and Pattern Recognition, pages 5521–5531, 2022

  30. [30]

    Neural scene flow fields for space-time view synthesis of dy- namic scenes

    Zhengqi Li, Simon Niklaus, Noah Snavely, and Oliver Wang. Neural scene flow fields for space-time view synthesis of dy- namic scenes. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6498– 6508, 2021

  31. [31]

    Neulf: Efficient novel view synthesis with neural 4d light field

    Zhong Li, Liangchen Song, Celong Liu, Junsong Yuan, and Yi Xu. Neulf: Efficient novel view synthesis with neural 4d light field. InEGSR (ST), pages 59–69, 2022

  32. [32]

    Spacetime gaus- sian feature splatting for real-time dynamic view synthesis

    Zhan Li, Zhang Chen, Zhong Li, and Yi Xu. Spacetime gaus- sian feature splatting for real-time dynamic view synthesis. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8508–8520, 2024

  33. [33]

    Gaussian-flow: 4d reconstruction with dynamic 3d gaus- sian particle

    Youtian Lin, Zuozhuo Dai, Siyu Zhu, and Yao Yao. Gaussian-flow: 4d reconstruction with dynamic 3d gaus- sian particle. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 21136– 21145, 2024

  34. [34]

    Neural vol- umes: learning dynamic renderable volumes from images

    Stephen Lombardi, Tomas Simon, Jason Saragih, Gabriel Schwartz, Andreas Lehrmann, and Yaser Sheikh. Neural vol- umes: learning dynamic renderable volumes from images. ACM Transactions on Graphics (TOG), 38(4):1–14, 2019

  35. [35]

    Mix- ture of volumetric primitives for efficient neural rendering

    Stephen Lombardi, Tomas Simon, Gabriel Schwartz, Michael Zollhoefer, Yaser Sheikh, and Jason Saragih. Mix- ture of volumetric primitives for efficient neural rendering. ACM Transactions on Graphics (ToG), 40(4):1–13, 2021

  36. [36]

    Dynamic 3d gaussians: Tracking by per- sistent dynamic view synthesis

    Jonathon Luiten, Georgios Kopanas, Bastian Leibe, and Deva Ramanan. Dynamic 3d gaussians: Tracking by per- sistent dynamic view synthesis. In2024 International Con- ference on 3D Vision (3DV), pages 800–809. IEEE, 2024

  37. [37]

    Nerf: Representing scenes as neural radiance fields for view syn- thesis.Communications of the ACM, 65(1):99–106, 2021

    Ben Mildenhall, Pratul P Srinivasan, Matthew Tancik, Jonathan T Barron, Ravi Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view syn- thesis.Communications of the ACM, 65(1):99–106, 2021

  38. [38]

    Instant neural graphics primitives with a mul- tiresolution hash encoding.ACM transactions on graphics (TOG), 41(4):1–15, 2022

    Thomas M ¨uller, Alex Evans, Christoph Schied, and Alexan- der Keller. Instant neural graphics primitives with a mul- tiresolution hash encoding.ACM transactions on graphics (TOG), 41(4):1–15, 2022

  39. [39]

    Donerf: Towards real- time rendering of compact neural radiance fields using depth oracle networks

    Thomas Neff, Pascal Stadlbauer, Mathias Parger, Andreas Kurz, Joerg H Mueller, Chakravarty R Alla Chaitanya, Anton Kaplanyan, and Markus Steinberger. Donerf: Towards real- time rendering of compact neural radiance fields using depth oracle networks. InComputer Graphics Forum, pages 45–

  40. [40]

    Wiley Online Library, 2021

  41. [41]

    Nerfies: Deformable neural radiance fields

    Keunhong Park, Utkarsh Sinha, Jonathan T Barron, Sofien Bouaziz, Dan B Goldman, Steven M Seitz, and Ricardo Martin-Brualla. Nerfies: Deformable neural radiance fields. InProceedings of the IEEE/CVF International Conference on Computer Vision, pages 5865–5874, 2021

  42. [42]

    Hypernerf: A higher- dimensional representation for topologically varying neural radiance fields.ACM Transactions on Graphics (TOG), 40: 1 – 12, 2021

    Keunhong Park, Utkarsh Sinha, Peter Hedman, Jonathan T Barron, Sofien Bouaziz, Dan B Goldman, Ricardo Martin- Brualla, and Steven M Seitz. Hypernerf: A higher- dimensional representation for topologically varying neural radiance fields.ACM Transactions on Graphics (TOG), 40: 1 – 12, 2021

  43. [43]

    D-nerf: Neural radiance fields for dynamic scenes

    Albert Pumarola, Enric Corona, Gerard Pons-Moll, and Francesc Moreno-Noguer. D-nerf: Neural radiance fields for dynamic scenes. InProceedings of the IEEE/CVF Con- ference on Computer Vision and Pattern Recognition, pages 10318–10327, 2021

  44. [44]

    Structure- from-motion revisited

    Johannes L Schonberger and Jan-Michael Frahm. Structure- from-motion revisited. InProceedings of the IEEE con- ference on computer vision and pattern recognition, pages 4104–4113, 2016

  45. [45]

    View morphing

    Steven M Seitz and Charles R Dyer. View morphing. InPro- ceedings of the 23rd annual conference on Computer graph- ics and interactive techniques, pages 21–30, 1996

  46. [46]

    Tensor4d: Efficient neural 4d decomposition for high-fidelity dynamic reconstruction and rendering

    Ruizhi Shao, Zerong Zheng, Hanzhang Tu, Boning Liu, Hongwen Zhang, and Yebin Liu. Tensor4d: Efficient neural 4d decomposition for high-fidelity dynamic reconstruction and rendering. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16632– 16642, 2023

  47. [47]

    Light field neural rendering

    Mohammed Suhail, Carlos Esteves, Leonid Sigal, and Ameesh Makadia. Light field neural rendering. InProceed- ings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8269–8279, 2022

  48. [48]

    Nerfstudio: A modular framework for neural radiance field development

    Matthew Tancik, Ethan Weber, Evonne Ng, Ruilong Li, Brent Yi, Justin Kerr, Terrance Wang, Alexander Kristof- fersen, Jake Austin, Kamyar Salahi, Abhik Ahuja, David McAllister, and Angjoo Kanazawa. Nerfstudio: A modular framework for neural radiance field development. InACM SIGGRAPH 2023 Conference Proceedings, 2023

  49. [49]

    De- ferred neural rendering: Image synthesis using neural tex- tures.Acm Transactions on Graphics (TOG), 38(4):1–12, 2019

    Justus Thies, Michael Zollh ¨ofer, and Matthias Nießner. De- ferred neural rendering: Image synthesis using neural tex- tures.Acm Transactions on Graphics (TOG), 38(4):1–12, 2019

  50. [50]

    R2l: Distilling neural radiance field to neural light field for efficient novel view synthesis

    Huan Wang, Jian Ren, Zeng Huang, Kyle Olszewski, Men- glei Chai, Yun Fu, and Sergey Tulyakov. R2l: Distilling neural radiance field to neural light field for efficient novel view synthesis. InEuropean Conference on Computer Vi- sion, pages 612–629. Springer, 2022

  51. [51]

    Freetimegs: Free gaussian primitives at anytime any- where for dynamic scene reconstruction

    Yifan Wang, Peishan Yang, Zhen Xu, Jiaming Sun, Zhan- hua Zhang, Yong Chen, Hujun Bao, Sida Peng, and Xiaowei Zhou. Freetimegs: Free gaussian primitives at anytime any- where for dynamic scene reconstruction. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 21750–21760, 2025

  52. [52]

    10 4d gaussian splatting for real-time dynamic scene rendering

    Guanjun Wu, Taoran Yi, Jiemin Fang, Lingxi Xie, Xiaopeng Zhang, Wei Wei, Wenyu Liu, Qi Tian, and Xinggang Wang. 10 4d gaussian splatting for real-time dynamic scene rendering. InProceedings of the IEEE/CVF conference on computer vi- sion and pattern recognition, pages 20310–20320, 2024

  53. [53]

    Neural radiance fields for the real world: A survey.arXiv preprint arXiv:2501.13104, 2025

    Wenhui Xiao, Remi Chierchia, Rodrigo Santa Cruz, Xuesong Li, David Ahmedt-Aristizabal, Olivier Salvado, Clinton Fookes, and Leo Lebrat. Neural radiance fields for the real world: A survey.arXiv preprint arXiv:2501.13104, 2025

  54. [54]

    Neural fields in visual computing and beyond

    Yiheng Xie, Towaki Takikawa, Shunsuke Saito, Or Litany, Shiqin Yan, Numair Khan, Federico Tombari, James Tomp- kin, Vincent Sitzmann, and Srinath Sridhar. Neural fields in visual computing and beyond. InComputer graphics forum, pages 641–676. Wiley Online Library, 2022

  55. [55]

    Deformable 3d gaussians for high- fidelity monocular dynamic scene reconstruction

    Ziyi Yang, Xinyu Gao, Wen Zhou, Shaohui Jiao, Yuqing Zhang, and Xiaogang Jin. Deformable 3d gaussians for high- fidelity monocular dynamic scene reconstruction. InPro- ceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 20331–20341, 2024

  56. [56]

    Real- time photorealistic dynamic scene representation and render- ing with 4d gaussian splatting

    Zeyu Yang, Hongye Yang, Zijie Pan, and Li Zhang. Real- time photorealistic dynamic scene representation and render- ing with 4d gaussian splatting. InInternational Conference on Learning Representations (ICLR), 2024

  57. [57]

    Time- archival camera virtualization for sports and visual perfor- mances.Computer Vision and Image Understanding, 267: 104714, 2026

    Yunxiao Zhang, William Stone, and Suryansh Kumar. Time- archival camera virtualization for sports and visual perfor- mances.Computer Vision and Image Understanding, 267: 104714, 2026

  58. [58]

    Time- archival camera virtualization for sports and visual perfor- mances.arXiv preprint arXiv:2602.15181, 2026

    Yunxiao Zhang, William Stone, and Suryansh Kumar. Time- archival camera virtualization for sports and visual perfor- mances.arXiv preprint arXiv:2602.15181, 2026. 11 3D Gaussian Splatting for Efficient Retrospective Dynamic Scene Novel View Synthesis with a Standardized Benchmark Supplementary Material

  59. [59]

    More Detail on Blender API While Neural Radiance Fields (NeRF) and 3D Gaussian Splatting (3DGS) have achieved remarkable progress in novel view synthesis , preparing multi-view datasets across different methods remains a bottleneck due to varying for- mat requirements and coordinate system conventions To al- leviate the extensive manual setup and debuggin...

  60. [60]

    dis- tant warm-start

    Additional Ablation Studies Here, we provide additional analyses supporting the two central claims of the paper: (i) our time-archival 3DGS formulation preserves the practical efficiency of Gaussian splatting for retrospective rendering, and (ii) warm-start training primarily exploits local temporal smoothness rather than any implicit temporal coupling in...