← Back to papers

Paper deep dive

ChopGrad: Pixel-Wise Losses for Latent Video Diffusion via Truncated Backpropagation

Dmitriy Rivkin, Parker Ewen, Lili Gao, Julian Ost, Stefanie Walz, Rasika Kangutkar, Mario Bijelic, Felix Heide

Year: 2026Venue: arXiv preprintArea: cs.CVType: PreprintEmbeddings: 89

Intelligence

Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 93%

Last extracted: 3/22/2026, 5:58:21 AM

Summary

ChopGrad is a truncated backpropagation scheme for video decoding in latent video diffusion models. It addresses the prohibitive memory costs of training with pixel-wise losses by limiting gradient computation to local frame windows, while maintaining global temporal consistency. This approach enables efficient fine-tuning for tasks like video super-resolution, inpainting, and controlled driving video generation.

Entities (6)

ChopGrad · method · 100%Truncated Backpropagation · algorithm · 98%Causal Caching · technique · 95%Latent Video Diffusion Models · model-architecture · 95%CogVideoX · model · 90%Wan 2.1 · model · 90%

Relation Signals (4)

ChopGrad enables Pixel-wise Losses

confidence 95% · ChopGrad unlocks pixel-wise losses for high resolution, long-duration video diffusion models.

ChopGrad reduces Memory Consumption

confidence 95% · ChopGrad reduces training memory from scaling linearly with the number of video frames to constant memory.

ChopGrad appliedto Wan 2.1

confidence 90% · We analyze the proposed method by first confirming that temporal locality holds in the popular WAN 2.1 video decoder.

Causal Caching introduces Recurrent Structure

confidence 90% · Notably, this approach introduces a recurrent structure into the autoencoder.

Cypher Suggestions (2)

Find all models that utilize ChopGrad for training optimization. · confidence 90% · unvalidated

MATCH (m:Model)-[:OPTIMIZED_BY]->(c:Method {name: 'ChopGrad'}) RETURN m.name

Identify tasks supported by ChopGrad. · confidence 85% · unvalidated

MATCH (c:Method {name: 'ChopGrad'})-[:SUPPORTS_TASK]->(t:Task) RETURN t.name

Abstract

Abstract:Recent video diffusion models achieve high-quality generation through recurrent frame processing where each frame generation depends on previous frames. However, this recurrent mechanism means that training such models in the pixel domain incurs prohibitive memory costs, as activations accumulate across the entire video sequence. This fundamental limitation also makes fine-tuning these models with pixel-wise losses computationally intractable for long or high-resolution videos. This paper introduces ChopGrad, a truncated backpropagation scheme for video decoding, limiting gradient computation to local frame windows while maintaining global consistency. We provide a theoretical analysis of this approximation and show that it enables efficient fine-tuning with frame-wise losses. ChopGrad reduces training memory from scaling linearly with the number of video frames (full backpropagation) to constant memory, and compares favorably to existing state-of-the-art video diffusion models across a suite of conditional video generation tasks with pixel-wise losses, including video super-resolution, video inpainting, video enhancement of neural-rendered scenes, and controlled driving video generation.

Tags

ai-safety (imported, 100%)cscv (suggested, 92%)preprint (suggested, 88%)

Links

Your browser cannot display the PDF inline. Open PDF directly →

Full Text

88,846 characters extracted from source content.

Expand or collapse full text

ChopGrad: Pixel-Wise Losses for Latent Video Diffusion via Truncated Backpropagation Dmitriy Rivkin 1 , Parker Ewen 1 , Lili Gao 1 , Julian Ost 1,2 , Stefanie Walz 1 , Rasika Kangutkar 1 , Mario Bijelic 1,2 , and Felix Heide 1,2 1 Torc Robotics 2 Princeton University Abstract. Recent video diffusion models achieve high-quality genera- tion through recurrent frame processing where each frame generation depends on previous frames. However, this recurrent mechanism means that training such models in the pixel domain incurs prohibitive mem- ory costs, as activations accumulate across the entire video sequence. This fundamental limitation also makes fine-tuning these models with pixel-wise losses computationally intractable for long or high-resolution videos. This paper introduces ChopGrad, a truncated backpropagation scheme for video decoding, limiting gradient computation to local frame windows while maintaining global consistency. We provide a theoretical analysis of this approximation and show that it enables efficient fine- tuning with frame-wise losses. ChopGrad reduces training memory from scaling linearly with the number of video frames (full backpropagation) to constant memory, and compares favorably to existing state-of-the- art video diffusion models across a suite of conditional video generation tasks with pixel-wise losses, including video super-resolution, video in- painting, video enhancement of neural-rendered scenes, and controlled driving video generation. 1 Introduction Recent methods in latent video diffusion are capable of generating high-resolution videos over long time horizons [22, 51, 57, 66]. Similar to latent image diffu- sion models, latent video diffusion models rely on pre-trained autoencoders to compress videos into latent embeddings and then learn over these embed- dings [4,5,19]. An enabling factor for recent video diffusion results is the use of temporal compression, where the autoencoder not only compresses video frames along spatial dimensions, but also along the temporal dimension [37,76,84]. Temporal compression groups multiple image frames into a single latent frame group. To incentivize temporal consistency between these frame groups causal caching has been introduced [63,70,74]. This technique appends embed- dings from previous frame group encodings onto the beginning of subsequent frame groups at each layer of the video encoder and decoder. Notably, this ap- proach introduces a recurrent structure into the autoencoder, where the depen- dency graph of video latents requires gradients to be propagated through all previous frame embeddings. arXiv:2603.17812v1 [cs.CV] 18 Mar 2026 2D. Rivkin et al. Fig. 1: ChopGrad Method. ChopGrad unlocks pixel-wise losses for high resolution, long-duration video diffusion models. It leverages truncated backpropagation to elimi- nate recursive activation accumulation in video autoencoders with causal caching. Solid arrows indicate the flow of information in the decoder forward pass, dashed ones in- dicate the backward flow of gradients with ChopGrad. Adding ChopGrad to training procedures is easy and produces state of the art performance in a variety of applications that benefit from pixel-wise losses, such as video super-resolution, video inpainting, video enhancement of neural rendered scenes, and controlled driving video generation. At the same time, most successful latent video diffusion models are trained within the latent space [2, 22, 51, 70], meaning gradients are not propagated through the encoder or decoder during latent video diffusion training. As such, existing methods make pixel-wise losses intractable for long-duration videos as the gradients of these losses require the recurrent accumulation of activations through the decoder. These pixel-level perceptual losses are used extensively in finetuning image diffusion models and video models with short-duration, low- resolution videos in applications such as single-step model distillation [73], en- hancement of neural rendered scenes [11,61], image translation [41], video super- resolution [12], and controlled driving video generation [34, 55]. In work such as [34, 41, 55, 61], the decoder itself is finetuned, making support for pixel-wise losses a strict requirement for training these types of models. To enable pixel-wise losses for high-resolution, long duration video diffusion, this work introduces ChopGrad, a truncated backpropagation scheme for video decoding (Fig. 1). Truncated backpropagation prevents activation accumulation over the full unrolled network by limiting the number of previous frames the gra- dients can propagate through. To validate this, we define latent temporal locality to demonstrate that the effect of prior video frames in the gradient error drops off at an exponential rate. We show that the proposed method enables efficient training using pixel-wise losses, such as the LPIPS [78] loss, across a variety of tasks and multiple video diffusion models. We evaluate our method on several ap- plications, including video super-resolution, video inpainting, video enhancement of neural rendered scenes, and controlled driving video generation, outperform- ing existing latent video diffusion adaptation methods in terms of quantitative frame-wise and video performance metrics. These results are achieved with mod- est computational resources (training times of approximately 3 to 4 hours on 4 to 8 A100 GPUs). The contributions of this paper are: – A mathematical derivation and error analysis of truncated backpropagation for causal video autoencoders, ChopGrad3 Fig. 2: ChopGrad Model Architecture. Given the processed video frame latents, the video decoder iteratively applies causal caching at each layer, producing pixel out- puts. Caching is performed by taking a subset of the layer outputs and appending these to the beginning of the layer inputs for the next frame group. While substan- tially reducing memory use at inference time compared to full 3D convolution over all frame groups, during training this process introduces recursive activation accumulation in the decoder, making backpropagation prohibitively expensive for high-resolution or long videos when using pixel-wise losses. Using truncated backpropagation, we only al- low gradients to accumulate through a fixed number (D trunc ) of previous frame groups. – A memory-efficient, practical approach for implementing pixel-wise losses for fine-tuning latent video diffusion models that generalizes across multiple diffusion models, – Validation of the method across several tasks requiring pixel-wise losses, including video super-resolution, video inpainting, video enhancement of neural-rendered scenes, and controlled driving video generation, comparing favorably to existing baselines in all experiments. 2 Related Work Latent video diffusion has experienced rapid advancement in recent years thanks in part to novel video auto-encoding methods [5,8,37,66]. In particular, temporal compression and causal caching have demonstrated significant improvements in video quality and temporal consistency. Latent video diffusion models extend latent image diffusion methods to model temporally coherent video sequences by operating in a compressed latent space rather than pixel space [4, 25, 66, 75]. Operating in a latent space [3, 31, 47, 57] reduces per-frame dimensionality and enables tractable scaling to longer and higher-resolution clips while preserving perceptual fidelity [22, 24]. Early video diffusion formulations applied standard image-based diffusion techniques directly to short clips, jointly denoising fixed-length frame blocks and introducing con- ditioning strategies to extend temporal length [2,13,25,47]. One of the most prevalent architectural advancements powering latent video diffusion is the use of temporal compression [9,14,18,19,80,84] and causal caching to preserve latent integrity and temporal consistency when processing long se- quences [2, 17, 74]. Causal caching has been used to maintain reconstruction fidelity and avoid temporal flicker while dramatically reducing memory and latency during encoding/decoding [32, 63]. Unfortunately, this causal caching mechanism for video encoding introduces a recurrent structure into the encoders 4D. Rivkin et al. and decoders used by latent video diffusion models, resulting in prohibitive mem- ory consumption due to activation accumulation during training when pixel-wise losses are used. A similar problem was encountered in early natural language processing with recurrent neural networks [42–44], where truncated backpropagation through time was used to mitigate this issue [1, 60]. To the best of our knowledge, this paradigm has not been investigated or applied for image or video models. Diffusion models often require long inference times, as the model must be run many times to generate an output. Single and few-step distillation [6,26,36,39, 45,52,69,73] has been used to reduce the number of steps required. Single-step distillation has also been used to adapt diffusion models to image-to-image trans- lation tasks like changing weather or generating images from sketches [10,30,41]. In applications where input/output pairs are readily available (such as super- resolution [12,21,53,58] or 3D gaussian splatting post-processing [16,34,55,61]), pretrained diffusion models [12, 16, 50, 55] or their one-step distilled counter- parts [34, 61] have been finetuned for single-step inference on the given task. Many of these single-step distillation and finetuning approaches rely on pixel- wise perceptual losses, albeit at low resolution and video duration in the case of video models due to memory constraints. As such, these single-step diffusion applications can derive the most benefit from ChopGrad. 2.1 Preliminaries Latent video diffusion models work by first mapping from the high-dimensional pixel space to a lower-dimensional latent space, down-sampling both the spatial and temporal dimensions via a pre-trained 3D VAE video encoder [3,70]. Once encoded, the video embeddings are then processed by the network backbone, often a transformer, which learns the temporal evolution of the video embed- dings. Finally, the output embeddings are re-projected into pixel space via the pre-trained 3D VAE decoder. The structure of such 3D VAE networks groups a set of frames into a single latent embedding. To retain temporal consistency these networks use what is called causal logic padding [63] or causal caching [32], where the trailing N outputs from the previous frame group are concatenated to the beginning of the subsequent frame group at each layer of the encoder and decoder [70,74]. This results in a recurrent structure, where the gradients of pixel-wise losses on later frames propagate through all previous frame groups. When training 3D VAEs, computational resources are dedicated solely to the VAE, and approaches such as sequence parallelism can be used to miti- gate these issues, as described in [70]. In addition, 3D VAEs are also able to be trained at lower resolutions/durations with results generalizing to higher resolu- tion/duration videos with no additional fine-tuning [70]. However, when training or fine-tuning latent video diffusion model transformers or U-nets, the majority of the memory budget is consumed by these backbones, prohibiting the alloca- tion of significant memory resources to decoder backpropagation. The backbones must also be trained at high resolution/duration if they are to perform well for ChopGrad5 high-resolution/duration inference, further compounding these memory require- ments, especially as adding pixel-wise losses also requires the decoders to perform inference at high resolution/duration, even if their own parameters are frozen. 3 ChopGrad In order to enable training of video diffusion models on long, high-resolution videos with pixel-wise losses while maintaining modest memory requirements we present ChopGrad, a novel method for backpropagating through the video de- coder. Sections 3.1 and 3.2 report that popular pre-trained video autoencoders with causal caching demonstrate temporal locality, where frame groups only af- fect other frame groups in close temporal proximity. Motivated by this insight, ChopGrad applies truncated backpropagation through time to the decoder cache to increase computational efficiency with minimal degradation in performance. With truncated backpropagation, gradients of each frame group are only able to accumulate to a portion of prior frame groups set by the truncation dis- tance. This breaks the recursive loop present in popular video autoencoders and enables pixel-wise losses for long, high-resolution videos. In Section 3.3 we quan- tify temporal locality and truncation gradient error in the Wan2.1 decoder and transformer. Implementation details are provided in the Supplemental Materials. 3.1 Causal Caching in Temporal VAEs The temporal VAE architecture with causal masking is first formalized. Let X = x 1 , x 2 ,..., x T denote a video sequence of T frames, where each frame x t ∈R H×W×C has height H, width W, and C channels. The 3D VAE encoder groups consecutive frames into non-overlapping seg- ments. For a frame group of size G, the i-th frame group contains frames X i =x iG , x iG+1 ,..., x iG+G−1 for i = 0, 1, 2,...,⌈T/G⌉. Let z i,m ∈R d m ×T ′ ×W ′ ×H ′ be the video latent embedding of frame group i at encoder layer m, where H ′ ,W ′ are the down-sampled spatial dimensions, T ′ is the down-sampled temporal dimension, and d m is the latent dimension for layer m. The causal caching mechanism ensures that the decoder (D) for frame group i receives context from the previous group. Specifically, let z c i−1,m denote the causal cache of size N of decoded features from group i−1 for the decoder layer m. The decoder then reconstructs the frames and constructs the cache z i,m+1 , z c i,m =D m (Concat(z c i−1,m , z i,m )).(1) The causal structure creates a recurrent dependency where the pixel-wise loss L pix i for group i depends on all previous groups through the concatenated context z c i−1 at each decoder layer. 6D. Rivkin et al. 3.2 Truncated Backpropagation and Locality Truncated backpropagation leverages temporal locality to enable efficient train- ing while preserving the essential temporal dependencies. The following analysis focuses on causal caching within the decoder network. Let z i ∈R d denote the unrolled latent, where the layer indices m are omitted for notational convenience. Let D(i,j) be a distance metric such that D(i,j) = 0 if and only if i and j refer to latents belonging to the same frame group. This index-based distance formalism allows us to reason about temporal proximity and the influence of one latent on another. Let J i,j = ∂z i /∂z j ∈R d×d denote the Jacobian of latent i with respect to latent j. The scalar influence measure is then defined as L i←j :=∥J i,j ∥,(2) for a chosen matrix norm. This quantity captures the effect of latent j on latent i and is a vector-norm on a vector space. Temporal locality is defined as the existence of constants C,α > 0 such that the influence measure decays exponentially with distance L i←j ≤ C· exp(−αD(i,j)).(3) Intuitively, this means that a latent only meaningfully affects nearby latents in time. Using the chain rule, the gradient of the overall loss L with respect to a latent z i decomposes as ∂L ∂z j = X i ∂L ∂z i ∂z i ∂z j = X i ∂L ∂z i J i,j .(4) Taking the norm of both sides and applying the triangle inequality, ∂L ∂z j ≤ X i ∂L ∂z i L i←j ,(5) which shows that the loss gradient at z i is dominated by contributions from latents in close temporal proximity assuming temporal locality holds. Our key insight is that the temporal locality enables effective truncated backpropagation in the 3D VAE decoder. When we truncate gradients to only flow through a limited number of previous frame groups, the exponential decay in the influence measure ensures that the approximation error is bounded. Specifically, for truncated backpropagation at temporal distance D trunc , the error in gradient computation is bounded by ∂L ∂z j − ∂L trunc ∂z j ≤ C· exp(−αD trunc ) X i ∂L ∂z i , (6) where L trunc denotes the loss computed with truncated backpropagation. ChopGrad7 0510152025 Temporal Distance D(i, j) 0.00 0.01 0.02 Locality Measure L i ← j Influence Measure Samples Mean Influence Measure Fit: L = exp(−0.68x − 4.27) Fig. 3: Temporal Locality. Influence measure samples (2) as a function of tem- poral distance between decoder inputs (i.e. latent embeddings) and outputs (i.e. pixels) alongside the mean and line of best fit. As temporal distance increases, the in- fluence between embeddings decreases ex- ponentially, resulting in minimal gradient contributions (5). 012345 Truncation Distance 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 Normalized MAE (MAE / mean|g*|) Normalized MAE Cosine Similarity 0.0 0.2 0.4 0.6 0.8 1.0 Cosine Similarity Fig. 4: Impact of Truncation Dis- tance on Backbone Model Param- eter Gradients. Normalized MAE and cosine distance (computed by flattening all model parameters) are shown. Though error is significant at small truncation dis- tances, the cosine similarity remains high across all distances, implying that the er- rors are primarily of magnitude, not di- rection. A truncation distance D trunc ≥ 1 α log( C ε ) can therefore be chosen to satisfy a desired error tolerance ε. In practice, the network still learns effectively with a small truncation distance as shown in Sections 3.3 and 4. The integration of causal caching with truncated backpropagation creates a hybrid approach: the network backbone can still attend to all video latent embeddings for global temporal understanding, while the 3D VAE decoder op- erates with limited temporal context, reducing computational complexity. This design preserves essential temporal dependencies while making large-scale video diffusion model training using pixel-wise losses computationally tractable. 3.3 Analysis Temporal Locality. We analyze the proposed method by first confirming that temporal locality holds in the popular WAN 2.1 video decoder [51]. The locality measure (3) is averaged across several videos, each with 97 frames and down- sampled to a resolution of 64×128 to prevent prohibitive memory requirements. Fig. 3 reports the mean of the influence measure (2) as a function of temporal distance, where a distance of 0 indicates pixel i is in the frame group of latent j. Notably, the locality measure decays at an exponential rate, meaning the influ- ence of pixels on frame groups significantly decreases as the temporal distance increases. This property is demonstrated implicitly for other 3D VAEs by the results presented in Section 4. Decoder Input Gradient Error. We likewise present the gradient error (6) be- tween the full and truncated backpropagation algorithms as a function of trun- cation distance. Gradients are computed by backpropagating pixel-wise losses to each decoder input latent considering varying truncation distances. Reported 8D. Rivkin et al. 5101520 Truncation Distance 2% 4% 6% 8% 10% Relative Differences Relative Diffs Norm Diffs 10 −7 10 −6 10 −5 Norm Differences Fig. 5: Truncation Induced Gradient Error. Mean gradient error (6) between the truncated and full backpropagation algorithms as a function of truncation dis- tance. 05101520 Truncation Distance 1 2 3 4 5 6 7 8 Time (seconds) Computation Time Memory Usage 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0 Memory (GB) Fig. 6: Resource Utilization. Compu- tational time and memory requirements as a function of truncation distance. results are the absolute and relative difference between the gradients for the truncated distance and the full backpropagation scheme. Differences are mea- sured using the Frobenius matrix norm and these, along with relative differences, are presented in Fig. 5. From this plot we see that, even for low truncation dis- tances, gradients approach those of full backpropagation, confirming that trun- cated backpropagation can be applied with minimal degradation in temporal consistency as the decoder network only considers small temporal neighborhoods. Effect on Backbone Model Parameters. Next, we evaluate the effect of gradient truncation on the backbone model parameters during training by computing the average gradient of the parameters of the public Wan 2.1 1.3B transformer checkpoint over the entire training set of the DL3DV-benchmark dataset (see Section 4.2), around 100 videos. We perform this computation over a range of truncation distances and compare to the gradients of the full backwards pass, with results presented in Fig. 4. Reported is the normalized mean absolute error (MAE) and cosine similarity, computed by flattening all model parameters into a single vector. The error is large for small truncation distances, indicating that the errors introduced by truncation are not averaged out over the dataset, and are propagated to model parameters. However, the high cosine similarity indicates that the error is primarily one of magnitude, not direction, and since gradient magnitudes are scaled by optimizers, the impact on training is negligible. This is confirmed by the results in Table 2, where increasing truncation distance only modestly improves performance. Runtime and Memory. Fig. 6 confirms that the proposed approach scales lin- early with respect to truncation distance in terms of both computational time and memory. We reiterate that memory use is constant with respect to video length. To further save on memory, gradients are truncated spatially as well as temporally, such that gradients are computed over spatial chunks of the video separately. This spatial locality is illustrated in Fig. 7 and has been explored and leveraged by existing state-of-the-art video diffusion models [51,70]. ChopGrad9 Fig. 7: Spatial Locality in 3D VAEs. The video frame on the left is decoded from the original latents, while on the right a section of latents is zeroed. The red line indicates the boundary between original and zeroed latents. The upper portion of the frame is entirely unaffected by the corruption of the bottom. 4 Applications We validate the efficacy of ChopGrad in four applications across multiple dif- fusion models: video super-resolution (Sec. 4.1), novel view synthesis (Sec. 4.2), video inpainting (Sec. 4.3), and controlled driving video generation (Sec. 4.4). 4.1 Video Super-Resolution We first show that adding ChopGrad to a state-of-the-art video super-resolution method yields significant improvements in perceptual losses by finetuning DOVE [12] using ChopGrad. DOVE finetunes CogVideoX [70], a DiT (Diffusion Trans- former) model, for super-resolution. DOVE uses pixel-wise losses, including MSE and DISTS [15], but is forced to encode and decode each video frame separately during loss computation due to memory constraints, reducing inter-frame con- sistency and requiring the addition of a frame consistency loss to attempt to compensate for this. In contrast, for ChopGrad, we start with the publicly avail- able DOVE checkpoint and perform full finetuning on the HQ-VSR dataset [12] for 500 steps using video lengths of 24 frames, omitting interframe consistency losses. We use frame-wise DISTS loss with a weight of 0.1 and pixel-wise MSE with a weight of 1. All other settings are consistent with the original DOVE Stage-2 implementation, except that in DOVE 80% of the batches are images, not videos, while we train on videos only. For the DOVE baseline, the publicly available DOVE checkpoint is used. As we found additional fine-tuning using the original DOVE method to result in equivalent performance, the results for the original model are presented. Quantitative results for video super-resolution are presented in Table 1. The addition of the proposed truncated backpropagation scheme improves perfor- mance across the majority of datasets and metrics, and the improvements are more pronounced for perceptual metrics (LPIPS and DISTS). Selected frames from processed videos are shown in Fig. 8, where ChopGrad synthesizes fine- grained details such as fur, hair, and clouds better than the baseline approach. 10D. Rivkin et al. Table 1: Quantitative Comparison for Video Super-Resolution. The first, sec- ond, and third best results are highlighted with dark green, light green, and yellow, re- spectively. ChopGrad outperforms all baselines in the majority of metrics and datasets, and achieves competitive performance otherwise. DatasetMetrics RealESRGANResShiftRealBasicVSRUpscale-A-VideoMGLD-VSRVEnhancerSTARDOVEChopGrad [56][77][7][82][67][20][65][12](Ours) UDM10 PSNR (↑)24.0423.6524.1321.7224.2321.3223.4726.4826.70 SSIM (↑)0.71070.60160.68010.59130.69570.68110.68040.78270.7753 LPIPS (↓)0.38770.55370.39080.41160.32720.43440.42420.26960.2346 DISTS (↓)0.21840.28980.20670.22300.16770.23100.21560.14920.1143 SPMCS PSNR (↑)21.2221.6822.1718.8122.3918.5821.2423.1123.67 SSIM (↑)0.56130.51530.56380.41130.58960.48500.54410.62100.6274 LPIPS (↓)0.37210.44670.36620.44680.32630.53580.52570.28880.2647 DISTS (↓)0.22200.26970.21640.24520.19600.26690.28720.17130.1448 YouHQ40 PSNR (↑)22.8223.3222.3919.6223.1719.7822.6424.3024.58 SSIM (↑)0.63370.62730.58950.48240.61940.59110.63230.67400.6760 LPIPS (↓)0.35710.42110.40910.42680.36080.47420.46000.29970.2581 DISTS (↓)0.17900.21590.19330.20120.16850.21400.22870.14770.1079 RealVSR PSNR (↑)20.8520.8122.1220.2922.0215.7517.4322.3222.43 SSIM (↑)0.71050.62770.71630.59450.67740.40020.52150.73010.7193 LPIPS (↓)0.20160.23120.18700.26710.21820.37840.29430.18510.1934 DISTS (↓)0.12790.14350.09830.14250.11690.16880.15990.09780.0944 MVSR4x PSNR (↑)22.4721.5821.8020.4222.7720.5022.4222.4222.55 SSIM (↑)0.74120.64730.70450.61170.74180.71170.74210.75230.7550 LPIPS (↓)0.45340.59450.42350.47170.35680.44710.43110.34760.3212 DISTS (↓)0.30210.33510.24980.26730.22450.28000.27140.23630.2071 Fig. 8: Video Super-Resolution Comparison. Shown from left to right: high- resolution, low-resolution input, DOVE [12], and the proposed approach, ChopGrad. ChopGrad synthesizes fine textures better and reduces motion blur, especially in re- gions with high-frequency details like fur, hair, cloth, and clouds. LPIPS scores for each frame are shown in the bottom right-hand corner, where a lower score indicates better perceptual quality. The associated videos can be found in the supplementary materials. 4.2 Artifact Removal in Novel View Synthesis Next, we use ChopGrad for refining renders from imperfect neural rendering models [29,38], which has recently become an established task [11,61]. Renders of 3D Gaussian Splatting novel view synthesis methods [29] often contain artifacts such as “floaters” that a set of recent diffusion models mitigate. Specifically, MVSplat-360 [11] and Difix3D+ [61] are designed for this task. MVSplat-360 is trained to refine video sequences of 14 frames rendered from 3DGS models while Difix is trained to refine individual frames. As a result, MVSplat-360 operates at a lower resolution (448×256) with a small window of temporal consistency while Difix operates at a higher resolution (960× 544) but has no capacity to enforce temporal consistency. While MVSplat-360 and Difix both leverage pixel-wise losses, they are unable to scale to long and high-resolution videos. We generate a dataset using the DL3DV-Benchmark [33], a collection of 140 videos and camera trajectories. Gaussian splat models are generated using ChopGrad11 Table 2: Neural Novel View Synthesis Results. Top section: ChopGrad out- performs all baselines across all metrics except temporal flickering, where it achieves competitive performance with MVSplat-360 [11]. Interestingly, while increasing the truncation distance noticeably increases training time memory, the metric differences are minimal. Bottom section: Ablation Results for ChopGrad. ChopGrad * uses the same 1-step diffusion network, but is only trained using latent mean-squared error. ChopGrad † likewise uses latent mean-squared error for training but is trained twice as long. As such, both ablations do not propagate gradients through the video decoder. The performance of ChopGrad using various truncation distances is also presented. MethodFID (↓) PSNR (↑) SSIM (↑) LPIPS (↓) Dists (↓) VBench Overall VBench Temporal Inference Time Train Time Quality (↑)Flickering (↑)[s/frame][H] Difix [61]16.63717.2130.5610.4070.1220.7660.8980.372.0 MVSplat-360 [11]38.20315.5020.4920.5320.2310.7430.9262.89- ChopGrad11.20919.2370.6100.3420.1130.7830.9211.114.0 ChopGrad * 48.52519.501 0.5880.4400.2440.7530.9331.112.3 ChopGrad † 48.17319.401 0.5860.4390.2380.7510.9321.114.5 ChopGrad D trunc = 011.775 19.2310.6050.3450.1150.7820.9201.113.5 ChopGrad D trunc = 1 11.209 19.2370.6100.3420.1130.7830.9211.114.0 ChopGrad D trunc = 211.74219.3080.6090.3430.1150.7820.9221.114.5 Fig. 9: ChopGrad vs Baselines for Neural Novel View Synthesis. Ground truth video frames and 3D Gaussian Splat renders are shown on the left. Results for MVSplat-360 [11] and Difix [61] are pre- sented alongside ChopGrad. Fig. 10: Ablation Experiments for Neural Novel View Synthesis. Chop- Grad * and ChopGrad † are trained using only the MSE loss in the latent space. The D trunc cases show ChopGrad results at various truncation distances. every 50th frame of each video and rendered videos are constructed along entire camera trajectories. For ChopGrad, we initialize the video diffusion model from a pre-trained Wan 2.1 14B [51] model and fine-tune the transformer backbone for 10 epochs. Difix is fine-tuned for 10000 steps on the same data. As MVSplat- 360 is trained on the DL3DV dataset, no fine-tuning is applied. We found that using the MVSplat-360 refinement model on our rendered videos led to poor performance. Performance was significantly improved using the same number of sparse views for constructing the 3DGS model when using the views specified in the MVSplat-360 repository. As such, we opt to use these improved selections for computing MVSplat-360 metrics. Fig. 9 depicts ChopGrad alongside the baseline methods for several scenes from the DL3DV-Benchmark test set and Table 2 presents quantitative results. ChopGrad out-performs the baselines across all metrics except temporal flicker- ing where results are competitive with MVSplat-360. A user study, available in 12D. Rivkin et al. the supplemental material, also found that 95.6% of users preferred the videos generated by ChopGrad over those generated by MVSplat-360 or Difix. Notably, while MVSplat-360 requires 60K training iterations [11], ChopGrad requires a small number of fine-tuning iterations when starting with the WAN2.1 14B pre- trained model. This demonstrates that ChopGrad enables diffusion models to quickly generalize to unseen tasks by fine-tuning using pixel-space losses. To demonstrate that the performance gains are a result of pixel-wise losses enabled by ChopGrad and not simply a more powerful backbone, we report ablation experiments in Table 2 (bottom section) and a qualitative comparison in Fig. 10, where ChopGrad is trained using only MSE loss in the latent space and using various truncation distances. While training only on the video latents is faster, the perceptual quality is worse and blurring is prevalent, especially in regions with fine details. As discussed in Section 3.3, truncation distance has a minor impact on result quality. Videos of the DL3DV-Benchmark for ChopGrad and baselines can be found in the supplementary materials. 4.3 Video Inpainting We demonstrate that in video inpainting applications, ChopGrad allows for re- ducing inference time by 50× while remaining on-par in terms of quality. We eval- uate ChopGrad for video inpainting on three datasets: DL3DV-Benchmark [33], Waymo Open Dataset [48], and ROVI [62]. For DL3DV-Benchmark and Waymo, we mask a fixed central region covering half the height and width of each frame and use an uninformative prompt. With ROVI, we use the included object masks and text descriptions. For ChopGrad we finetune a Wan 2.1 14B model using latent MSE and pixel LPIPS losses for single-step inference using a truncation distance of 1. The baseline is VACE [28] 14B, a control adapter for Wan 2.1 14B which is trained for a variety of tasks, including inpainting. VACE inference is performed using the default 50 steps from the VACE repository [28]. For all datasets, we train both ChopGrad and VACE the same number of steps. More training details are available in the supplemental material. Quantitative results are reported in Table 3, qualitative results in Fig. 11. ChopGrad outperforms VACE on reconstruction-based metrics and maintains similar video quality metrics (VBench overall quality score within 1% across all datasets) while reducing inference time compute budget by 50×. FVD (Fréchet Video Distance) is higher for ChopGrad on ROVI but lower for the other two datasets, likely stemming from the overall more extreme masking in D3LDV and ROVI. Qualitatively, we observe that the ChopGrad model adheres better to the scene and introduces fewer novel structures compared to VACE, occasionally at the cost of visual quality. In the more extreme masking regime of DL3DV and Waymo, VACE is penalized less for novel structures (relative to ChopGrad), as the unmasked region is less informative about the region inside the mask, resulting in smaller relative improvements in reconstruction-based losses. ChopGrad13 Table 3: Video Inpainting Results. ChopGrad results are output in a single step, a 50× compute time improvement over VACE. VBench components are provided in the supplemental material. Dataset Method FID ↓ FVD ↓ PSNR ↑ SSIM ↑ LPIPS ↓ DISTS ↓ VBench Overall ↑ DL3DV VACE45.060 574.441 20.678 0.757 0.2360.0830.792 ChopGrad 40.948 583.581 21.699 0.765 0.221 0.0770.792 Waymo VACE34.856 440.651 23.229 0.804 0.2120.0790.836 ChopGrad 27.057 470.545 25.048 0.823 0.192 0.0710.835 ROVI VACE30.491 201.610 22.618 0.834 0.2230.1120.752 ChopGrad 27.547 188.546 25.200 0.859 0.199 0.0950.747 Fig. 11: Video Inpainting. We find that the recent VACE [28] tends to hallucinate (e.g., left, top panel), while ChopGrad stays closer to the input but can also produce implausible results. ChopGrad results are output in a single step, a 50× compute time improvement over VACE. Left:DL3DV, Middle: Waymo, Right: ROVI. 4.4 Controlled Driving Video Generation Visually realistic controlled driving video generation is essential for autonomous vehicle safety as it enables validation of vehicle behavior in rarely encountered scenarios. 3DGS [29] offers powerful scene reconstruction approaches, and recent neural driving simulators allow for manipulation of vehicles and reconstructed assets using scene graphs [35,40,81] of reconstructed splats to enable this kind of simulation. However, large manipulation of vehicles and assets in these sim- ulators [35, 81] leads to myriad visual artifacts (see Naive Insertion columns of Fig. 12 for examples). Post-processing videos rendered from such neural scenes with single-step diffusion is a promising approach for overcoming these issues, but existing methods such as [34,55] suffer from resolution / duration limitations. Following [34, 55] we create a dataset based on Waymo Open Dataset [48] where 3DGS models are constructed, then assets are extracted and reinserted, producing the desired artifacts and input/output pairs to train and test with. Full details of dataset construction are presented in the Supplemental Material. We demonstrate ChopGrad for controlled driving video generation on Mi- rage [55] with our own Wan2.1-based implementation (details in the Supple- mental Material), as we were unable to acquire the original implementation even after contacting the authors. We train our implementation on 9-frame clips at a resolution of 480x832. After training Mirage we performed inference and evalu- 14D. Rivkin et al. Table 4: Controlled Driving Video Generation Results. ChopGrad was pro- duced by initializing with Mirage followed by further finetuning Mirage’s Harmoniza- tion stage for 1000 steps at high resolution / duration using ChopGrad. MethodPSNR ↑ SSIM ↑ LPIPS ↓ DISTS ↓ FID ↓ FVD ↓ Mirage (hires inference) 27.30 0.8912 0.2067 0.0740 10.28 204.66 ChopGrad29.49 0.9031 0.1719 0.0561 5.86 154.49 Fig. 12: Controlled Driving Video Generation. Training with ChopGrad improves lighting, removes more artifacts, and produces better shadows. ation at high resolution / duration (720x1280, 97 frames) as up-scaling training resolution outputs yielded poorer results. Subsequently, we finetuned Mirage’s harmonization stage model using ChopGrad for 1000 steps at 720x1280 resolu- tion, 49 frame duration, and performed inference on 49 frame segments. Results are reported in Table 4 and Fig. 12. Quantitative metrics are improved across all tests, while inspection of the qualitative results shows that finetuning Mirage with ChopGrad improves lighting fixing, artifact removal, and shadow insertion. Notably, the parameters of the decoder itself are finetuned in Mirage, confirming that ChopGrad can be used for decoder, as well as transformer, training. 5 Conclusion We introduce ChopGrad, a truncated backpropagation approach that enables pixel-wise supervision at high resolutions and long durations in latent video diffusion models with causal caching. In architectures where the decoder is fine- tuned (e.g. [55]) this capability is required, while in others it leads to significantly improved results (bottom of Table 2). Applications of such architectures trained with pixel-wise losses are numerous, including but not limited to single-step model distillation [73], enhancement of neural rendered scenes [11, 61], image translation [41], video super-resolution [12], and controlled driving video gener- ation [34,55]. By analyzing latent temporal locality, we demonstrate that long-range gradi- ent dependencies in causal video autoencoders decay exponentially, allowing gra- dients to be truncated without compromising performance. This insight allows for efficient fine-tuning of high-resolution, long-duration video diffusion models using perceptual losses that were previously intractable due to recursive activa- tion accumulation. ChopGrad15 Supplementary Material The following document provides supplemental information in support of the findings in the main manuscript. Section A provides additional implementation information for the proposed ChopGrad architecture using the WAN 2.1 [51] and CogVideoX [70] video autoencoders. Next, Section B reports additional details regarding evaluation setups and baseline implementations for all applications, while Section C provides details about model architectures, training setups, and inference schemes. Next, Section D describes additional algorithmic opti- mizations for ChopGrad to minimize computation times when long truncation distances are used. Finally, Sections E and F report additional qualitative and qualitative results, respectively. A Implementation Details ChopGrad is formalized in Algorithm 1 and illustrated in Fig.2 of the primary document. The latent cache is first either initialized as empty or detached from the previous decoder pass (lines 2 - 3). Critically, the cache is detached prior to running the forward pass, so the gradients do not propagate backwards through the full video. The pixel-wise loss is computed using the decoded frames (line 4) and the gradients backpropagated to the latents and cache (lines 5-6). Trun- cated backpropagation is then run using the specified truncation distance (lines 7-9), and the compute graph for latent z i−D trunc is subsequently released. Cache gradients are zeroed after each backpropagation. Notably, maintaining the com- pute graph in memory requires storing activations, resulting in memory use that scales linearly with D trunc , as does compute time as shown in Fig.6. B Additional Evaluation and Baseline Details This section provides additional evaluation and baseline details. Sec. B.1 presents details for Video Super-Resolution, Sec. B.2 for Artifact Removal in Novel View Synthesis, Sec. B.3 for Video Inpainting, and Sec. B.4 for Controlled Driving Video Generation. B.1 Video Super-Resolution Video super-resolution is evaluated across the following datasets: UDM10 [72], SPMCS [49], YouHQ40 [83], RealVSR [68], and MVSR4x [54]. Metric evaluations are performed using the publicly available DOVE evaluation script and reference baseline metrics are taken from those reported by DOVE. Notably, the evaluation metrics computed using this script using the DOVE checkpoint match those originally reported. 16D. Rivkin et al. Algorithm 1 ChopGrad. Require: Video latents z i ⌈T/G⌉ i=1 , D trunc , L pix . Ensure: Gradients ∇ z i L for all latent frame groups 1: for i from 1 to ⌈T /G⌉ do 2:z c i−1 ← detach(z c i−1 ) 3: ˆ X i , z c i =D(Concat(z c i−1 , z i )) 4: L i = 1 T P t L pix i ( ˆ X i,t , X i,t ) 5: ∇ z i ← ∂L i ∂z i 6: ∇ z c i−1 ← ∂L i ∂z c i−1 7: for k = 1 to min(D trunc , i) do 8: ∇ z (i−k) + = ∂z c (i−k) ∂z (i−k) ∇ z c (i−k) 9: ∇ z c (i−k−1) ← ∂z c (i−k) ∂z c (i−k−1) ∇ z c (i−k) 10: end for 11: if i≥ D trunc then 12:Release compute graph for z (i−D trunc ) 13: end if 14: end for DOVE. The publicly available DOVE code is used for inference and evalua- tion. DOVE contains a 2-stage training scheme, where the first stage uses video sequences and the second stage uses a combination of video sequences and indi- vidual frames to train the network. Additional Stage 2 fine-tuning was performed for 500 iterations, but no per- formance improvement was observed, indicating that the public checkpoint was already converged and as such did not benefit from additional training. As such, the original DOVE checkpoint is used in the evaluations. B.2 Artifact Removal in Novel View Synthesis We evaluate our neural-rendering enhancements on the DL3DV-Benchmark [33], which we split into 95 training scenes, 40 testing scenes, and 5 validation scenes. Each video in the dataset is approximately 300 frames long and a 3DGS model is trained using GSplat [71] using only every 50th frame for each scene. Training of the 3DGS models follows the standard gsplat implementation, where the first 500 iterations do not modify the number of Gaussians, then the next 14.5k iterations are tailored for adaptive density control and the remaining iterations are for Gaussian parameter optimization. The 3DGS model is trained using 960× 544 resolution images for a total of 30k iterations, requiring approximately 5 minutes of training for each scene using an A100 GPU. In addition, we conduct an anonymized study for determining user preference for novel view synthesis video generation. Participants are presented with three side-by-side videos from ten scenes randomly selected from the test set. Each of ChopGrad17 05101520 Truncation Distance 1.0 1.5 2.0 2.5 3.0 3.5 4.0 4.5 5.0 Time (seconds) Computation Time Memory Usage 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0 Memory (GB) Fig. 13: Computational time and memory requirements as a function of truncation distance for the modified ChopGrad algorithm. The full video length is 24 frame groups. The dashed horizontal lines indicate time and memory requirements of the original backpropagation scheme. the 3 videos is generated by one of the baselines, MVSplat [11], Difix [61], and the proposed method, ChopGrad (D trunc = 1), and users are asked to mark the video which they preferred. The order of video appearance for each scene was randomized. A total of 34 users participated in the survey. The percentage of users who preferred ChopGrad is computed for each scene and then averaged across all 10 scenes, yielding an overall preference rate of 95.6% for the proposed method. Difix. Difix utilizes an image diffusion model backbone to process individual frames from input video sequences. This diffusion network is initialized from the SD-Turbo [46] checkpoint and is fine-tuned for neural render enhancement. For a fair comparison, additional fine-tuning was performed on the DL3DV dataset using the training settings provided in the paper. For each training iteration, video frames are randomly sampled from the dataset. Difix is fine- tuned for 10k iterations on 4 A100 GPUs, taking approximately 2 hours. MVSplat360. The MVSplat-360 baseline approach takes as input a sparse set of views and uses a pre-trained feed-forward network to estimate a 3DGS model. Next, a video sequence is generated along a camera trajectory and this video is refined using a fine-tuned video diffusion network. The video diffusion network is initialized using the publicly available Stable Video Diffusion (SVD) checkpoint [3]. The fine-tuned video diffusion checkpoint is provided by the MVSplat-360 authors and used in the experiments. The SVD model groups together 14 video frames during the diffusion process. Self attention is used to enable frame groups to attend to one another, but this resulted in prohibitive memory requirements when evaluating videos with 294 frames, even at low resolution. As such, this feature was disabled and frame groups were not able to attend to each other. 18D. Rivkin et al. As MVSplat-360 was trained on the DL3DV dataset, no additional fine- tuning was performed. B.3 Video Inpainting We evaluate inpainting on 3 datasets: D3LDV-benchmark [33], Waymo [48], and ROVI [62]. For D3LDV-Benchmark we use the same dataset and train/test split as was described in Sec. B.2. For Waymo, we use the dataset as described in Sec. B.4. For both of these cases we mask the ground truth videos as described in Section 4.2, with a fixed rectangular mask having half the height and half the width of the video. We also use a fixed prompt for both. We also include a new experiment (not included in the original paper) where vehicle bounding boxes are masked in Waymo. Specifically, we randomly select 50% of the labeled vehicles in the scene, and mask out the vehicle throughout the entire video. We refer to this setup as “Waymo-Bbox.” For ROVI we use the train/test split from the original dataset, as well as prompts provided by this dataset. VACE. For all experiments, we train the VACE baseline for the same number of steps as ChopGrad, with the same video duration, and resolution. All other settings are also the same, unless otherwise indicated in this section. VACE is trained using the standard latent MSE velocity objective (as Wan 2.1 and thus VACE are flow models), with the denoising timestep sampled using a timestep shift of 5, the same shift being used during sampling. The model is sampled over 50 steps following default settings in the VACE implementation 3 . Masks are provided to the model during training and inference. B.4 Controlled Driving Video Generation We evaluate on the Waymo Open Dataset [48], reconstructing 300 sequences, 230 for training and 70 for evaluation, using SplatAD [23], a dynamic 3D Gaussian Splatting-based method. SplatAD [23] decomposes each scene into static back- ground and dynamic actors represented as Gaussian primitives, which allows us to remove selected actors and replace them with generated 3D vehicles at the original pose. Our vehicle generation pipeline consists of vehicle extraction fol- lowed by vehicle alignment. We extract object-centric image patches from the curated Waymo object vehicle set using camera detections and LIDAR track IDs to assemble multi-view crops. Using instance and segmentation masks, we remove background pixels, and the resulting patches are used as input to the TRELLIS image→3D pipeline [64] to produce 3D Gaussian representations of the vehicles. Since TRELLIS produces reconstructions in an unanchored coordinate frame, the generated vehicles can be arbitrarily rotated. To ensure consistent forward-facing poses, we render each vehicle from angles of 0 °, 90°, 180°, 270° and estimate its yaw with Orient-Anything [59]. vehicles with inconsistent cyclic orientation patterns or orientation confidence below a threshold are discarded, and the original 3DGS scene models are used. 3 https://github.com/ali-vilab/VACE ChopGrad19 Mirage. Mirage code is not publicly available, and we were not able to gain access by emailing the authors, so we re-implement it based on the description of the method provided in the paper. We choose to use Wan 2.1 14B instead of CogVideoX as this is a more modern and capable diffusion model, but uses a similar VAE architecture. For fair comparison, for ChopGrad we keep the same model architecture, losses, etc. and only modify the training duration and resolution, since use of ChopGrad allows us to greatly increase these. The main architectural modifications to the diffusion model proposed by Mi- rage are the addition of skip connections to the decoder, and the addition of several LoRAs. The skip connections are extracted from the encoder in “2D” mode, i.e., where each frame is encoded separately and no temporal compression is used. They are extracted from the output of the layer immediately preceding the first spatiotemporal downsample step, and are fused into the decoder im- mediately following the final spatio-temporal upsample step. To perform fusion, we concatenate the skip connections with decoder features along the channel dimension, then perform convolution with a 3× 3 kernel to compute the fused features. We use LoRA rank and alpha values of 16 in the VAE, while for the transformer we set rank to 128 and alpha to 64. We replicate Mirage’s two training stages – Reconstruction and Harmoniza- tion. We train each for 10k steps, with separate LoRAs for each stage. The Re- construction phase trains the fusion blocks and a decoder reconstruction LoRA, while Harmonization trains a transformer LoRA and another decoder LoRA and keeps the fusion blocks frozen. For both stages learning rate is 2× 10 −4 , batch size is 8, clip length is 9 frames and clip resolution is 480× 832. We updated the resolution slightly from the original paper to match Wan 2.1’s training res- olution. We trained the reconstruction phase with LPIPS and MSE losses, with the LPIPS component scaled by a factor of 0.1. The harmonization phase was trained with a fixed timestep (200). To be better aligned with Wan 2.1’s velocity prediction training scheme, we trained the model to predict the difference between output and input. A combination of LPIPS and Gram losses was used, with Gram loss being scaled by 0.1. The AdamW optimizer was used with a weight decay of 0.01, betas of 0.9 and 0.99. All trainings were performed on a node with 8 80GB A100 GPUs. Though the original paper trained on H200s with a smaller base model, we were able to fit our implementation on A100s by fully sharding the model and optimizer parameters with FSDP 4 , and using transformer and decoder activation checkpointing. C Architecture, Training, and Inference Details This section provides additional details on architectures, training, and inference. Sec. C.1 presents details for Video Super-Resolution, Sec. C.2 for Artifact Re- moval in Novel View Synthesis, Sec. C.3 for Video Inpainting, and Sec. C.4 for Controlled Driving Video Generation. 4 https://docs.pytorch.org/docs/stable/fsdp.html 20D. Rivkin et al. C.1 Video Super-Resolution Network Architecture. For video super-resolution, the DOVE [12] checkpoint is used to initialize ChopGrad. DOVE uses the CogVideoX [70] autoencoder. Similar to the WAN 2.1 [51] video autoencoder, the CogVideoX autoencoder compresses multiple video frames into frame groups. This temporal compression differs from the WAN 2.1 autoencoder however, in that the number of decoded frames changes depending on how many frame groups are used. When an odd number of frame groups are passed to the decoder, the first frame group is decoded into a single frame and the remaining frame groups are decoded into 4 frames. When an even number of frame groups are passed to the decoder, all frame groups are decoded into 4 frames. This behavior necessitates that 2 frame groups must be decoded together for each decoding step. Although this behavior could be addressed through minor modifications to the implementation, no changes were made to preserve compatibility with the original network. Training. The default DOVE configuration performs training at a resolution of 640× 320. Given this low resolution, no spatial chunking was used. Fine-tuning is conducted using videos with lengths of 24 frames. The second stage training implementation from DOVE is adopted with original hyperparameters and ini- tialization is performed from the provided checkpoint. Fine-tuning is performed for 500 iterations on 4 A100 GPUs, requiring approximately 8 hours. In contrast to the original DOVE Stage 2 procedure, only videos are used, no images are trained on. C.2 Artifact Removal in Novel View Synthesis Network Architecture. ChopGrad and ablations are initialized using the pre- trained Wan 2.1 14B [51] diffusion transformer model. This model is then fine- tuned using the latent embeddings of the 3DGS renders as inputs and the ground- truth images as targets. Notably, WAN is trained to output velocity v = ˆ z−z, where ˆ z is the latent embedding of the rendered video andz is the latent embed- ding of the ground truth video. In order to better align with the original training objective of Wan 2.1 [51], we leverage the same training scheme for fine-tuning. A fixed text caption is used to condition the refinement and the diffusion timestep is fixed to 200. No modifications to the video encoder have been made and the pre-trained WAN network has not been pre-distilled for single-step inference. The WAN 2.1 video autoencoder has a temporal compression factor of 4, meaning there are 4 video frames per frame group latent. Notably, the network pads the first video frame with 3 empty frames, meaning the total length of the video processed by WAN 2.1 is 4N + 1, where N is the number of frame groups. This temporal compression factor corresponds to G = 4 from Section 3.4. ChopGrad decodes latents to pixels using a spatial chunk size of H/2×W/2, resulting in 4 chunks. This preserves the aspect ratio of the video and enables parallel processing for each chunk. ChopGrad21 Training. ChopGrad is trained using both a latent MSE loss and a pixel-space LPIPS [78] loss with VGG features. An LPIPS weight of 100 was used while the latent MSE weight was set to 1 for all experiments, including ablations. A scene is randomly chosen from the dataset and a random 81-frame sequence from this scene is used for each training iteration. Videos have a resolution of 832 × 480. Notably, larger resolution videos may be used for training (i.e. 1280×720), but lower resolution videos were used in experiments for fair baseline comparisons. Training is conducted for approximately 3-4 hours on 8 A100 GPUs. Py- Torch’s Fully Sharded Data Parallel architecture [79] is leveraged to shard the model parameters and optimizer states of the WAN diffusion transformer. To minimize memory, 8-way sequence parallelism is used. The AdamW optimizer is used with a learning rate of 1e −5 , a weight decay of 0.1, betas of 0.9 and 0.99, and a batch size of 1. ChopGrad and all ablations, ex- cluding ChopGrad † , are trained for 880 training iterations. ChopGrad † is trained for 1760 training iterations. Training times are presented in Table 2. Inference. Inference times in Table 1 are measured by processing the entire video and dividing by the number of total video frames and finally multiplied by the number of GPUs to account for sequence parallelism. This ensures a fair comparison with the baseline methods which both utilize only a single GPU. These inference times also include pre-processing as well as the full network pass (i.e., video encoding, decoding, and transformer forward pass). Inference is performed on the first 297 frames of the video as this corresponds to the temporal compression of the WAN 2.1 video autoencoder using a total of N = 75 frame groups. The evaluation is conducted on the first 294 frames of the video to maintain compatibility and a fair comparison with the baseline methods as MVSplat-360 is only able to process multiples of 14 frames. C.3 Video Inpainting Network Architecture. We use the same network setup as described in Sup- plemental Section 4.2. ChopGrad is initialized using the pretrained Wan 2.1 14B [51] diffusion transformer model. This model is then fine-tuned using the latent embeddings of the masked videos as inputs and the ground-truth videos as targets. Notably, WAN is trained to output velocity v = ˆ z −z, where ˆ z is the latent embedding of the rendered video andz is the latent embedding of the ground truth video. In order to better align with the original training objective of Wan 2.1 [51], we leverage the same training scheme for fine-tuning. A fixed text caption (except for the ROVI case) is used to condition the refinement and the diffusion timestep is fixed to 200. No modifications to the video encoder have been made and the pre-trained WAN network has not been pre-distilled for single-step inference. The WAN 2.1 video autoencoder has a temporal compression factor of 4, meaning there are 4 video frames per frame group latent. Notably, the network 22D. Rivkin et al. pads the first video frame with 3 empty frames, meaning the total length of the video processed by WAN 2.1 is 4N + 1, where N is the number of frame groups.This temporal compression factor corresponds to G = 4 from Section 3.4. ChopGrad decodes latents to pixels using a spatial chunk size of H/2× W/2, resulting in 4 chunks. This preserves the aspect ratio of the video and enables parallel processing for each chunk. Training. Training is done using the same settings as described in Supplemental Section 4.2, except for differences in duration and number of training steps, which vary across the datasets. For DL3DV-benchmark we trained for 5 epochs at 49 frames, for Waymo (and Waymo-Bbox) 5 epochs at 49 frames, for ROVI 10k steps at 29 frames. The train times were 1.5, 3, and 18 hours, respectively. Train time for Waymo-Bbox was the same as for Waymo. The increased train steps for ROVI reflect the fact that it is a much larger dataset than the other two (5172 videos in the training set). Inference. We perform inference at the same duration as training, evaluating on the first N frames of each video, where N is the training/inference duration. C.4 Controlled Driving Video Generation Network Architecture. For ChopGrad, we keep the same network as for Mirage (described in detail in Supplemental Section B.4). Training. We initialize with the Mirage harmonization checkpoint which had been trained for 10k steps. The validation loss plateaued around 5k steps so we are confident the model had converged. We then train it for a subsequent 1k steps using ChopGrad at a resolution of 720× 1280 and a duration of 49 frames. We use 16 spatial chunks (4h× 4w) and a truncation distance of 1. During this training batch size was set to 1 and the learning rate is maintained at 2× 10 −4 . The AdamW optimizer is used with a weight decay of 0.1, betas of 0.9 and 0.99. The training process took approximately 10 hours on 8 A100 GPUs. D Algorithmic Optimizations for Truncated Backpropagation The time complexity of ChopGrad as described in Algorithm 1 scales linearly with respect to D trunc . This complexity stems from the need to backpropagate over each frame group D trunc times. This is not the case with the standard backpropagation scheme, which only needs to make one backward pass, having accumulated gradients from all frame groups i + 1...N prior to computing the gradient for frame group i. This section examines in more detail the origin of this difference in time complexity and introduces a minor modification to Algorithm 1 that ensures, as the truncation distance approaches the full video length, the overall time complexity converges to that of the full backpropagation scheme. ChopGrad23 ChopGrad requires multiple backward passes over each frame group because in order to compute the gradient for frame group i, gradients for frame groups i + 1...i + D trunc must be available. Furthermore, the compute graph for frame group i cannot be released from memory until all D trunc future gradients have backpropagated through it. In order to release frame group i as soon as possi- ble, i.e. once the algorithm reaches frame group i + D trunc , it is imperative to backpropagate all the way from i+D trunc to i as soon i+D trunc is decoded and the loss computed. This necessitates performing a backward pass through all intermediate frame groups as well. Since backpropagation is performed through D trunc frame groups each time the compute graph from frame group i is released, the compute must scale linearly with D trunc . There is a time-memory tradeoff present – graphs could be released less often, for example every s steps instead of every single step, reducing the time complexity by a factor of s but increasing memory consumption accordingly, since D trunc + s frames worth of activations need to be stored in memory. In Algorithm 1, backpropagation is performed all the way back through the previous D trunc steps as soon as a new frame group is decoded. If this is delayed so that the backpropagation only occurs when it is time to evict the compute graph for frame group i− D trunc , a performance improvement can be gained in regimes where D trunc is close to the full video length, T. This is because the total number of backpropagation steps through individual frame groups (and as such the time complexity) becomes equal to D trunc ∗ n evict , where n evict is the number of cache evictions, and n evict = T−D trunc . Empirical complexity results for this modification are shown in Figure 13. Note that a low resolution video (128x64) is used to make computation of the vanilla backpropagation method and ChopGrad with high truncation distances tractable. In Figure 13 ChopGrad has slightly worse time and memory performance that vanilla backpropagation at D trunc = T as some overhead is introduced by fragmenting the compute graph and maintaining detached versions of the cache. It is not recommended to use ChopGrad at truncation distances greater than 2, as the results from Section 4 demonstrate that causal VAEs exhibit strong spatial locality, and terms from faraway latents do not contribute significantly to the gradient. In regimes with small values of D trunc , the modified and un- modified algorithms exhibit practically equivalent performance characteristics. This section is included for completeness, but should not have much impact on real-world uses of ChopGrad. E Additional Qualitative Results Additional qualitative results for Video Super-Resolution, Artifact Removal in Novel View Synthesis, and Controlled Driving Video Generation experiments are presented in Figures 14, 15, and 20 respectively. Qualitative results for inpaint- ing are presented in four separate figures based on dataset: DL3DV in Fig. 16, Waymo in Fig. 17, Waymo-Bbox in Fig. 18, and ROVI in Fig. 19. These addi- tional results report a number of the benefits of training at increased resolution / 24D. Rivkin et al. duration with ChopGrad. In the case of Video Super-Resolution (Fig. 14), Chop- Grad enabled back-propagating through the decoder for entire videos (rather than individual frames), allowing the transformer to properly account for tem- poral compression and resulting in improved visual quality, especially for fine details such as fur, cloth, and clouds. In the case of Artifact Removal in Novel View Synthesis (Fig. 15), ChopGrad trained models have access to many more views of the scene simultaneously thanks to their ability to handle long videos, leading to strongly enhanced artifact removal capabilities. In Video Inpainting (Figs. 16, 17, 18, and 19), ChopGrad is used to produce a significantly faster model (single-step vs 50 steps for VACE) while also reducing hallucinations. Fi- nally, Fig. 20 shows that for Controlled Driving Video Generation, tuning with ChopGrad (as compared to training at low resolution/duration and perform- ing high resolution/duration inference) leads to a stronger model that is able to make larger changes to input images, improving lighting and shadows and removing more Gaussian Splat artifacts. Collectively, these results illustrate a variety of ways in which state-of-the-art methods may be improved further by using truncated backpropagation to enable pixel-wise perceptual losses. F Additional Quantitative Results Supplemental Table 6 presents individual VBench [27] component scores used to compile the VBench Overall Quality metric reported for the Video Inpainting application in main document Table 3. Higher is better for all scores. In addition to a 50× reduction in compute time, ChopGrad delivers modest improvements in motion smoothness and temporal flickering across all datasets, while VACE consistently has slightly higher imaging quality and subject consistency, with the remaining scores having mixed results across datasets. This pattern is consistent with our observation that VACE is more prone to hallucination – conforming less closely to the input video allows it to generate slightly higher quality images (despite being included in VBench, aesthetic quality is an image-based metric) and more consistent subjects. The ChopGrad trained single-step model halluci- nates less, retains higher reconstruction metrics (as evidenced in main document Table 3), and its improved motion smoothness and temporal flickering can be attributed to it staying closer to the original video, as real videos often have smooth motion and less flicker than generated videos. Supplemental Table 5 presents the same metrics as main document Table 3 for the new Waymo-Bbox task. Results are consistent with other tasks, with ChopGrad training resulting in improved reconstruction metrics and similar video quality metrics, while achieving a 50× inference time reduction. ChopGrad25 Fig. 14: Additional Video Super-Resolution Comparison. Shown from left to right: high-resolution, low-resolution input, DOVE [12], and the proposed approach, Chop- Grad. ChopGrad synthesizes fine textures better and reduces motion blur, especially in regions with high-frequency details like fur, hair, cloth, and clouds. 26D. Rivkin et al. Fig. 15: Additional Qualitative Results for Artifact Removal in Novel View Synthesis on the DL3DV-Benchmark Dataset [33]. Ground truth video frames and 3DGS model renders are shown on the left. Results for MVSplat-360 [11] and Difix [61] are presented alongside the ChopGrad with a truncation distance of 1 and 2. ChopGrad corrects sig- nificantly more artifacts than other methods (e.g., fourth row from the top) with fewer hallucinations (e.g., 5th row from the bottom), and maintains temporal consistency over the entire video sequence. ChopGrad27 VACE ChopGrad Ground Truth Fig. 16: Additional Video Inpainting Comparison on D3LDV Dataset. Shown from left to right: VACE, ChopGrad, Ground Truth. Training with ChopGrad reduces hallucinations despite 50× lower in- ference budget. 28D. Rivkin et al. VACE ChopGrad Ground Truth Fig. 17: Additional Video Inpainting Comparison on Waymo Dataset. Shown from left to right: VACE, ChopGrad, Ground Truth. Training with ChopGrad reduces hallucinations despite 50× lower in- ference budget. ChopGrad29 VACE ChopGrad Ground TruthMask Fig. 18: Additional Video Inpainting Comparison on Waymo-Bbox Task. In this task, 50% of the vehicles are randomly selected for masking. Shown from left to right: VACE, ChopGrad, Ground Truth. Training with ChopGrad reduces hallucinations despite 50× lower inference budget. 30D. Rivkin et al. VACE ChopGrad Ground TruthMask Fig. 19: Additional Video Inpainting Comparison on ROVI Dataset. Shown from left to right: VACE, ChopGrad, Ground Truth. Training with ChopGrad reduces halluci- nations despite 50× lower inference budget. ChopGrad31 Fig. 20: Additional Controlled Driving Video Generation Comparison. Shown from left to right: Naive Insertion, Mirage [55], ChopGrad, and Ground Truth. Rows A-F demonstrate that training with ChopGrad increases realism by improving lighting and shadows, and removing more Gaussian Splat artifacts. Rows G and H, which have been cropped, demonstrate that training with ChopGrad enables the model to make stronger collections in the presence of very poor vehicle model quality – note that these very poor vehicle models are relatively rare in the dataset. 32D. Rivkin et al. Table 5: Waymo-Bbox Video Inpainting Results. Quantitative comparison be- tween VACE and ChopGrad on the Waymo-Bbox setting. Results are consistent with other tasks, with ChopGrad training resulting in improved reconstruction metrics and similar video quality metrics, while achieving a 50× inference time reduction. Method FID FVD PSNR SSIM LPIPS DISTS VBench Overall VACE 12.084 253.710 27.661 0.873 0.154 0.0610.834 ChopGrad 12.358 252.599 28.838 0.875 0.146 0.0570.835 Table 6: Breakdown of VBench Scores for Video Inpainting. Full VBench com- ponent scores corresponding to the VBench Overall values reported in main document Table 3. In addition to a 50× reduction in compute time, ChopGrad delivers modest improvements in motion smoothness and temporal flickering across all datasets, while VACE consistently has slightly higher imaging quality and subject consistency, with the remaining scores having mixed results across datasets. DatasetMethod Aesthetic Quality Background Consistency Dynamic Degree Imaging Quality Motion Smoothness Subject Consistency Temporal Flickering VBench Overall Quality DL3DV VACE0.5310.9170.950 0.7310.9570.9120.9120.792 ChopGrad 0.5190.9120.950 0.7290.9610.9030.9190.792 Waymo VACE0.5120.9560.882 0.7160.9870.9530.9690.836 ChopGrad 0.5170.9600.894 0.6950.9880.9450.9720.835 Waymo-Bbox VACE0.5190.9570.859 0.7040.9870.9500.9700.834 ChopGrad 0.5250.9620.847 0.6960.9880.9550.9720.835 ROVI VACE0.4720.9260.816 0.5380.9590.8850.9410.752 ChopGrad 0.4570.9240.749 0.5210.9640.8840.9470.747 ChopGrad33 References 1. Aicher, C., Foti, N.J., Fox, E.B.: Adaptively truncating backpropagation through time to control gradient bias. In: Uncertainty in Artificial Intelligence. p. 799–808. PMLR (2020) 4 2. An, J., Zhang, S., Yang, H., Gupta, S., Huang, J.B., Luo, J., Yin, X.: Latent-Shift: Latent diffusion with temporal shift for efficient text-to-video generation. arXiv preprint arXiv:2304.08477 (2023) 2, 3 3. Blattmann, A., Dockhorn, T., Kulal, S., Mendelevitch, D., Kilian, M., Lorenz, D., Levi, Y., English, Z., Voleti, V., Letts, A., et al.: Stable video diffusion: Scaling latent video diffusion models to large datasets. arXiv preprint arXiv:2311.15127 (2023) 3, 4, 17 4. Blattmann, A., Rombach, R., Ling, H., Dockhorn, T., Kim, S.W., Fidler, S., Kreis, K.: Align Your Latents: High-resolution video synthesis with latent diffusion mod- els. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. p. 22563–22575 (2023) 1, 3 5. Ceylan, D., Huang, C.H.P., Mitra, N.J.: Pix2Video: Video editing using image diffusion. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. p. 23206–23217 (2023) 1, 3 6. Chadebec, C., Tasar, O., Benaroche, E., Aubin, B.: Flash diffusion: Accelerating any conditional diffusion model for few steps image generation. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 39, p. 15686–15695 (2025) 4 7. Chan, K.C., Zhou, S., Xu, X., Loy, C.C.: Investigating tradeoffs in real-world video super-resolution. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. p. 5962–5971 (2022) 10 8. Chen, H., Zhang, Y., Cun, X., Xia, M., Wang, X., Weng, C., Shan, Y.: VideoCrafter2: Overcoming data limitations for high-quality video diffusion mod- els. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. p. 7310–7320 (2024) 3 9. Chen, L., Li, Z., Lin, B., Zhu, B., Wang, Q., Yuan, S., Zhou, X., Cheng, X., Yuan, L.: OD-VAE: An omni-dimensional video compressor for improving latent video diffusion model. arXiv preprint arXiv:2409.01199 (2024) 3 10. Chen, S., Ye, T., Lin, Y., Jin, Y., Yang, Y., Chen, H., Lai, J., Fei, S., Xing, Z., Tsung, F., et al.: Genhaze: Pioneering controllable one-step realistic haze gen- eration for real-world dehazing. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. p. 9194–9205 (2025) 4 11. Chen, Y., Zheng, C., Xu, H., Zhuang, B., Vedaldi, A., Cham, T.J., Cai, J.: MVS- plat360: Feed-forward 360 scene synthesis from sparse views. Advances in Neural Information Processing Systems 37, 107064–107086 (2024) 2, 10, 11, 12, 14, 17, 26 12. Chen, Z., Zou, Z., Zhang, K., Su, X., Yuan, X., Guo, Y., Zhang, Y.: DOVE: Efficient one-step diffusion model for real-world video super-resolution. In: The Thirty-ninth Annual Conference on Neural Information Processing Systems (2025) 2, 4, 9, 10, 14, 20, 25 13. Danier, D., Zhang, F., Bull, D.: LDMVFI: Video frame interpolation with latent diffusion models. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 38, p. 1472–1480 (2024) 3 14. D’Avino, D., Cozzolino, D., Poggi, G., Verdoliva, L.: Autoencoder with recurrent neural networks for video forgery detection. arXiv preprint arXiv:1708.08754 (2017) 3 34D. Rivkin et al. 15. Ding, K., Ma, K., Wang, S., Simoncelli, E.P.: Image quality assessment: Unifying structure and texture similarity. CoRR abs/2004.07728 (2020), https://arxiv. org/abs/2004.07728 9 16. Dong, Y., Zhang, Q., Jiang, M., Wu, Z., Fan, Q., Feng, Y., Zhang, H., Bao, H., Zhang, G.: One-shot refiner: Boosting feed-forward novel view synthesis via one- step diffusion. arXiv preprint arXiv:2601.14161 (2026) 4 17. Gao, K., Shi, J., Zhang, H., Wang, C., Xiao, J., Chen, L.: Ca2-VDM: Efficient autoregressive video diffusion model with causal generation and cache sharing. arXiv preprint arXiv:2411.16375 (2024) 3 18. Golinski, A., Pourreza, R., Yang, Y., Sautiere, G., Cohen, T.S.: Feedback recurrent autoencoder for video compression. In: Proceedings of the Asian Conference on Computer Vision (2020) 3 19. HaCohen, Y., Chiprut, N., Brazowski, B., Shalem, D., Moshe, D., Richardson, E., Levin, E., Shiran, G., Zabari, N., Gordon, O., et al.: LTX-Video: Realtime video latent diffusion. arXiv preprint arXiv:2501.00103 (2024) 1, 3 20. He, J., Xue, T., Liu, D., Lin, X., Gao, P., Lin, D., Qiao, Y., Ouyang, W., Liu, Z.: VEnhancer: Generative space-time enhancement for video generation. arXiv preprint arXiv:2407.07667 (2024) 10 21. He, X., Tang, H., Tu, Z., Zhang, J., Cheng, K., Chen, H., Guo, Y., Zhu, M., Wang, N., Gao, X., et al.: One step diffusion-based super-resolution with time- aware distillation. arXiv preprint arXiv:2408.07476 (2024) 4 22. He, Y., Yang, T., Zhang, Y., Shan, Y., Chen, Q.: Latent video diffusion models for high-fidelity long video generation. arXiv preprint arXiv:2211.13221 (2022) 1, 2, 3 23. Hess, G., Lindström, C., Fatemi, M., Petersson, C., Svensson, L.: Splatad: Real- time lidar and camera rendering with 3d gaussian splatting for autonomous driving. In: Proceedings of the Computer Vision and Pattern Recognition Conference. p. 11982–11992 (2025) 18 24. Ho, J., Chan, W., Saharia, C., Whang, J., Gao, R., Gritsenko, A., Kingma, D.P., Poole, B., Norouzi, M., Fleet, D.J., et al.: Imagen video: High definition video generation with diffusion models. arXiv preprint arXiv:2210.02303 (2022) 3 25. Ho, J., Salimans, T., Gritsenko, A., Chan, W., Norouzi, M., Fleet, D.J.: Video diffusion models. Advances in neural information processing systems 35, 8633– 8646 (2022) 3 26. Huang, X., Li, Z., He, G., Zhou, M., Shechtman, E.: Self forcing: Bridging the train-test gap in autoregressive video diffusion. arXiv preprint arXiv:2506.08009 (2025) 4 27. Huang, Z., He, Y., Yu, J., Zhang, F., Si, C., Jiang, Y., Zhang, Y., Wu, T., Jin, Q., Chanpaisit, N., Wang, Y., Chen, X., Wang, L., Lin, D., Qiao, Y., Liu, Z.: VBench: Comprehensive benchmark suite for video generative models. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2024) 24 28. Jiang, Z., Han, Z., Mao, C., Zhang, J., Pan, Y., Liu, Y.: Vace: All-in-one video creation and editing. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. p. 17191–17202 (2025) 12, 13 29. Kerbl, B., Kopanas, G., Leimkühler, T., Drettakis, G.: 3d gaussian splatting for real-time radiance field rendering. ACM Trans. Graph. 42(4) (2023) 10, 13 30. Lee, S., Kim, K., Ye, J.C.: Single-step bidirectional unpaired image translation us- ing implicit bridge consistency distillation. arXiv preprint arXiv:2503.15056 (2025) 4 ChopGrad35 31. Li, X., Zhang, Y., Ye, X.: DrivingDiffusion: Layout-guided multi-view driving sce- narios video generation with latent diffusion model. In: European Conference on Computer Vision. p. 469–485. Springer (2024) 3 32. Li, Z., Lin, B., Ye, Y., Chen, L., Cheng, X., Yuan, S., Yuan, L.: WF-VAE: En- hancing video VAE by wavelet-driven energy flow for latent video diffusion model. In: Proceedings of the Computer Vision and Pattern Recognition Conference. p. 17778–17788 (2025) 3, 4 33. Ling, L., Sheng, Y., Tu, Z., Zhao, W., Xin, C., Wan, K., Yu, L., Guo, Q., Yu, Z., Lu, Y., et al.: DL3DV-10k: A large-scale scene dataset for deep learning-based 3D vision. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2024) 10, 12, 16, 18, 26 34. Ljungbergh, W., Taveira, B., Zheng, W., Tonderski, A., Peng, C., Kahl, F., Pe- tersson, C., Felsberg, M., Keutzer, K., Tomizuka, M., et al.: R3d2: Realistic 3d asset insertion via diffusion for autonomous driving simulation. arXiv preprint arXiv:2506.07826 (2025) 2, 4, 13, 14 35. Ljungbergh, W., Tonderski, A., Johnander, J., Caesar, H., Åström, K., Felsberg, M., Petersson, C.: Neuroncap: Photorealistic closed-loop safety testing for au- tonomous driving. In: European Conference on Computer Vision. p. 161–177. Springer (2024) 13 36. Mao, X., Jiang, Z., Wang, F.Y., Zhang, J., Chen, H., Chi, M., Wang, Y., Luo, W.: Osv: One step is enough for high-quality image to video generation. In: Proceedings of the Computer Vision and Pattern Recognition Conference. p. 12585–12594 (2025) 4 37. Melnik, A., Ljubljanac, M., Lu, C., Yan, Q., Ren, W., Ritter, H.: Video diffusion models: A survey. arXiv preprint arXiv:2405.03150 (2024) 1, 3 38. Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: Nerf: Representing scenes as neural radiance fields for view synthesis. Commu- nications of the ACM 65(1) (2021) 10 39. Noroozi, M., Hadji, I., Martinez, B., Bulat, A., Tzimiropoulos, G.: You only need one step: Fast super-resolution with stable diffusion via scale distillation. In: Eu- ropean Conference on Computer Vision. p. 145–161. Springer (2024) 4 40. Ost, J., Mannan, F., Thuerey, N., Knodt, J., Heide, F.: Neural scene graphs for dynamic scenes. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. p. 2856–2865 (2021) 13 41. Parmar, G., Park, T., Narasimhan, S., Zhu, J.Y.: One-step image translation with text-to-image models. arXiv preprint arXiv:2403.12036 (2024) 2, 4, 14 42. Pascanu, R., Mikolov, T., Bengio, Y.: On the difficulty of training recurrent neural networks. In: International conference on machine learning. p. 1310–1318. Pmlr (2013) 4 43. Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning internal representations by error propagation. Tech. rep., Institute of Cognitive Science (1985) 4 44. Salehinejad, H., Sankar, S., Barfett, J., Colak, E., Valaee, S.: Recent advances in recurrent neural networks. arXiv preprint arXiv:1801.01078 (2017) 4 45. Sauer, A., Boesel, F., Dockhorn, T., Blattmann, A., Esser, P., Rombach, R.: Fast high-resolution image synthesis with latent adversarial diffusion distillation. In: SIGGRAPH Asia 2024 Conference Papers. p. 1–11 (2024) 4 46. Sauer, A., Lorenz, D., Blattmann, A., Rombach, R.: Adversarial diffusion distilla- tion. In: European Conference on Computer Vision. Springer (2024) 17 47. Singer, U., Polyak, A., Hayes, T., Yin, X., An, J., Zhang, S., Hu, Q., Yang, H., Ashual, O., Gafni, O., et al.: Make-A-Video: Text-to-video generation without text-video data. arXiv preprint arXiv:2209.14792 (2022) 3 36D. Rivkin et al. 48. Sun, P., Kretzschmar, H., Dotiwalla, X., Chouard, A., Patnaik, V., Tsui, P., Guo, J., Zhou, Y., Chai, Y., Caine, B., et al.: Scalability in perception for autonomous driving: Waymo open dataset. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. p. 2446–2454 (2020) 12, 13, 18 49. Tao, X., Gao, H., Liao, R., Wang, J., Jia, J.: Detail-revealing deep video super- resolution. In: The IEEE International Conference on Computer Vision (ICCV) (Oct 2017) 15 50. Teng, S., Gao, G., Danier, D., Jiang, Y., Zhang, F., Davis, T., Liu, Z., Bull, D.: Gfix: Perceptually enhanced gaussian splatting video compression. arXiv preprint arXiv:2511.06953 (2025) 4 51. Wan, T., Wang, A., Ai, B., Wen, B., Mao, C., Xie, C.W., Chen, D., Yu, F., Zhao, H., Yang, J., et al.: WAN: Open and advanced large-scale video generative models. arXiv preprint arXiv:2503.20314 (2025) 1, 2, 7, 8, 11, 15, 20, 21 52. Wang, H., Liu, F., Chi, J., Duan, Y.: Videoscene: Distilling video diffusion model to generate 3d scenes in one step. In: 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). p. 16475–16485. IEEE (2025) 4 53. Wang, J., Lin, S., Lin, Z., Ren, Y., Wei, M., Yue, Z., Zhou, S., Chen, H., Zhao, Y., Yang, C., et al.: Seedvr2: One-step video restoration via diffusion adversarial post-training. arXiv preprint arXiv:2506.05301 (2025) 4 54. Wang, R., Liu, X., Zhang, Z., Wu, X., Feng, C.M., Zhang, L., Zuo, W.: Benchmark dataset and effective inter-frame alignment for real-world video super-resolution. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (2023) 15 55. Wang, S., Sun, H., Wang, B., Ye, H., Yu, X.: Mirage: One-step video diffusion for photorealistic and coherent asset editing in driving scenes. arXiv preprint arXiv:2512.24227 (2025) 2, 4, 13, 14, 31 56. Wang, X., Xie, L., Dong, C., Shan, Y.: Real-ESRGAN: Training real-world blind super-resolution with pure synthetic data. In: Proceedings of the IEEE/CVF in- ternational conference on computer vision. p. 1905–1914 (2021) 10 57. Wang, Y., Chen, X., Ma, X., Zhou, S., Huang, Z., Wang, Y., Yang, C., He, Y., Yu, J., Yang, P., et al.: LAVIE: High-quality video generation with cascaded latent dif- fusion models. International Journal of Computer Vision 133(5), 3059–3078 (2025) 1, 3 58. Wang, Y., Yang, W., Chen, X., Wang, Y., Guo, L., Chau, L.P., Liu, Z., Qiao, Y., Kot, A.C., Wen, B.: Sinsr: diffusion-based image super-resolution in a single step. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. p. 25796–25805 (2024) 4 59. Wang, Z., Zhang, Z., Pang, T., Du, C., Zhao, H., Zhao, Z.: Orient anything: Learn- ing robust object orientation estimation from rendering 3d models. arXiv preprint arXiv:2412.18605 (2024) 18 60. Williams, R.J., Zipser, D.: Gradient-based learning algorithms for recurrent net- works and their computational complexity. In: Backpropagation, p. 433–486. Psy- chology Press (2013) 4 61. Wu, J.Z., Zhang, Y., Turki, H., Ren, X., Gao, J., Shou, M.Z., Fidler, S., Gojcic, Z., Ling, H.: Difix3D+: Improving 3D reconstructions with single-step diffusion mod- els. In: Proceedings of the Computer Vision and Pattern Recognition Conference (2025) 2, 4, 10, 11, 14, 17, 26 62. Wu, J., Li, X., Si, C., Zhou, S., Yang, J., Zhang, J., Li, Y., Chen, K., Tong, Y., Liu, Z., et al.: Towards language-driven video inpainting via multimodal large language models. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. p. 12501–12511 (2024) 12, 18 ChopGrad37 63. Wu, P., Zhu, K., Liu, Y., Zhao, L., Zhai, W., Cao, Y., Zha, Z.J.: Improved video VAE for latent video diffusion model. In: Proceedings of the Computer Vision and Pattern Recognition Conference. p. 18124–18133 (2025) 1, 3, 4 64. Xiang, J., Lv, Z., Xu, S., Deng, Y., Wang, R., Zhang, B., Chen, D., Tong, X., Yang, J.: Structured 3d latents for scalable and versatile 3d generation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. p. 21469–21480 (2025) 18 65. Xie, R., Liu, Y., Zhou, P., Zhao, C., Zhou, J., Zhang, K., Zhang, Z., Yang, J., Yang, Z., Tai, Y.: STAR: Spatial-temporal augmentation with text-to-video models for real-world video super-resolution. arXiv preprint arXiv:2501.02976 (2025) 10 66. Xing, Z., Feng, Q., Chen, H., Dai, Q., Hu, H., Xu, H., Wu, Z., Jiang, Y.G.: A survey on video diffusion models. ACM Computing Surveys 57(2), 1–42 (2024) 1, 3 67. Yang, X., He, C., Ma, J., Zhang, L.: Motion-Guided latent diffusion for temporally consistent real-world video super-resolution. In: European conference on computer vision. p. 224–242. Springer (2024) 10 68. YANG, X., Xiang, W., Zeng, H., Zhang, L.: Real-world video super-resolution: A benchmark dataset and a decomposition based learning scheme. ICCV (2021) 15 69. Yang, Y., Huang, H., Peng, X., Hu, X., Luo, D., Zhang, J., Wang, C., Wu, Y.: Towards one-step causal video generation via adversarial self-distillation. arXiv preprint arXiv:2511.01419 (2025) 4 70. Yang, Z., Teng, J., Zheng, W., Ding, M., Huang, S., Xu, J., Yang, Y., Hong, W., Zhang, X., Feng, G., et al.: CogVideoX: Text-to-video diffusion models with an expert transformer. arXiv preprint arXiv:2408.06072 (2024) 1, 2, 4, 8, 9, 15, 20 71. Ye, V., Li, R., Kerr, J., Turkulainen, M., Yi, B., Pan, Z., Seiskari, O., Ye, J., Hu, J., Tancik, M., et al.: gsplat: An open-source library for gaussian splatting. Journal of Machine Learning Research 26(34) (2025) 16 72. Yi, P., Wang, Z., Jiang, K., Shao, Z., Ma, J.: Multi-temporal ultra dense memory network for video super-resolution. IEEE Transactions on Circuits and Systems for Video Technology 30(8) (2019) 15 73. Yin, T., Gharbi, M., Park, T., Zhang, R., Shechtman, E., Durand, F., Freeman, B.: Improved distribution matching distillation for fast image synthesis. Advances in neural information processing systems 37, 47455–47487 (2024) 2, 4, 14 74. Yu, L., Lezama, J., Gundavarapu, N.B., Versari, L., Sohn, K., Minnen, D., Cheng, Y., Birodkar, V., Gupta, A., Gu, X., et al.: Language model beats diffusion– Tokenizer is key to visual generation. arXiv preprint arXiv:2310.05737 (2023) 1, 3, 4 75. Yu, S., Sohn, K., Kim, S., Shin, J.: Video probabilistic diffusion models in projected latent space. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. p. 18456–18466 (2023) 3 76. Yu, W., Xing, J., Yuan, L., Hu, W., Li, X., Huang, Z., Gao, X., Wong, T.T., Shan, Y., Tian, Y.: ViewCrafter: Taming video diffusion models for high-fidelity novel view synthesis. arXiv preprint arXiv:2409.02048 (2024) 1 77. Yue, Z., Wang, J., Loy, C.C.: ResShift: Efficient diffusion model for image super- resolution by residual shifting. Advances in Neural Information Processing Systems 36, 13294–13307 (2023) 10 78. Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE conference on computer vision and pattern recognition. p. 586–595 (2018) 2, 21 38D. Rivkin et al. 79. Zhao, Y., Gu, A., Varma, R., Luo, L., Huang, C.C., Xu, M., Wright, L., Shojanazeri, H., Ott, M., Shleifer, S., et al.: PyTorch FSDP: Experiences on scaling fully sharded data parallel. arXiv preprint arXiv:2304.11277 (2023) 21 80. Zheng, Z., Peng, X., Yang, T., Shen, C., Li, S., Liu, H., Zhou, Y., Li, T., You, Y.: Open-Sora: Democratizing efficient video production for all. arXiv preprint arXiv:2412.20404 (2024) 3 81. Zhou, H., Lin, L., Wang, J., Lu, Y., Bai, D., Liu, B., Wang, Y., Geiger, A., Liao, Y.: Hugsim: A real-time, photo-realistic and closed-loop simulator for autonomous driving. IEEE Transactions on Pattern Analysis and Machine Intelligence (2025) 13 82. Zhou, S., Yang, P., Wang, J., Luo, Y., Loy, C.C.: Upscale-A-Video: Temporal- consistent diffusion model for real-world video super-resolution. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. p. 2535–2545 (2024) 10 83. Zhou, S., Yang, P., Wang, J., Luo, Y., Loy, C.C.: Upscale-A-Video: Temporal- consistent diffusion model for real-world video super-resolution. In: CVPR (2024) 15 84. Zhou, Y., Wang, Q., Cai, Y., Yang, H.: Allegro: Open the black box of commercial- level video generation model. arXiv preprint arXiv:2410.15458 (2024) 1, 3