← Back to papers

Paper deep dive

TRACE: Trajectory Recovery with State Propagation Diffusion for Urban Mobility

Jinming Wang, Hai Wang, Hongkai Wen, Geyong Min, Man Luo

Year: 2026Venue: arXiv preprintArea: cs.LGType: PreprintEmbeddings: 49

Intelligence

Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 97%

Last extracted: 3/23/2026, 12:05:20 PM

Summary

TRACE is a novel diffusion-based trajectory recovery framework designed to reconstruct dense, continuous GPS trajectories from sparse, irregular inputs. It introduces the State Propagation Diffusion Model (SPDM), which utilizes a memory mechanism to retain and leverage intermediate denoising features across steps, significantly improving recovery accuracy for complex spatio-temporal patterns.

Entities (5)

Jinming Wang · researcher · 100%SPDM · model-component · 100%TRACE · model · 100%GPS Trajectory Recovery · task · 95%MCGRU · neural-network-architecture · 95%

Relation Signals (3)

TRACE incorporates SPDM

confidence 100% · At the core of TRACE, we propose a State Propagation Diffusion Model (SPDM)

TRACE performs GPS Trajectory Recovery

confidence 100% · TRACE: Trajectory Recovery with State Propagation Diffusion for Urban Mobility

SPDM utilizes MCGRU

confidence 95% · the state propagation network is responsible for updating the cumulative knowledge... we implement a novel Multi-scale Convolutional Gated Recurrent Unit (MCGRU)

Cypher Suggestions (2)

Identify researchers associated with the TRACE paper · confidence 95% · unvalidated

MATCH (r:Researcher)-[:AUTHORED]->(p:Paper {title: 'TRACE: Trajectory Recovery with State Propagation Diffusion for Urban Mobility'}) RETURN r

Find all components of the TRACE model · confidence 90% · unvalidated

MATCH (m:Model {name: 'TRACE'})-[:INCORPORATES]->(c) RETURN m, c

Abstract

Abstract:High-quality GPS trajectories are essential for location-based web services and smart city applications, including navigation, ride-sharing and delivery. However, due to low sampling rates and limited infrastructure coverage during data collection, real-world trajectories are often sparse and feature unevenly distributed location points. Recovering these trajectories into dense and continuous forms is essential but challenging, given their complex and irregular spatio-temporal patterns. In this paper, we introduce a novel diffusion model for trajectory recovery named TRACE, which reconstruct dense and continuous trajectories from sparse and incomplete inputs. At the core of TRACE, we propose a State Propagation Diffusion Model (SPDM), which integrates a novel memory mechanism, so that during the denoising process, TRACE can retain and leverage intermediate results from previous steps to effectively reconstruct those hard-to-recover trajectory segments. Extensive experiments on multiple real-world datasets show that TRACE outperforms the state-of-the-art, offering $>$26\% accuracy improvement without significant inference overhead. Our work strengthens the foundation for mobile and web-connected location services, advancing the quality and fairness of data-driven urban applications. Code is available at: this https URL

Tags

ai-safety (imported, 100%)cslg (suggested, 92%)preprint (suggested, 88%)

Links

Your browser cannot display the PDF inline. Open PDF directly →

Full Text

48,241 characters extracted from source content.

Expand or collapse full text

TRACE: Trajectory Recovery with State Propagation Diffusion for Urban Mobility Jinming Wang University of Exeter Department of Computer Science Exeter, United Kingdom jw1294@exeter.ac.uk Hai Wang Jingdong Logistics Beijing, China hai@seu.edu.cn Hongkai Wen University of Warwick Department of Computer Science Coventry, United Kingdom hongkai.wen@warwick.ac.uk Geyong Min University of Exeter Department of Computer Science Exeter, United Kingdom G.Min@exeter.ac.uk Man Luo* University of Exeter Department of Computer Science Exeter, United Kingdom M.Luo@exeter.ac.uk Abstract High-quality GPS trajectories are essential for location-based web services and smart city applications, including navigation, ride- sharing and delivery. However, due to low sampling rates and limited infrastructure coverage during data collection, real-world trajectories are often sparse and feature unevenly distributed loca- tion points. Recovering these trajectories into dense and continuous forms is essential but challenging, given their complex and irreg- ular spatio-temporal patterns. In this paper, we introduce a novel diffusion model for TRAjectory rECovery named TRACE, which reconstruct dense and continuous trajectories from sparse and in- complete inputs. At the core of TRACE, we propose a State Propa- gation Diffusion Model (SPDM), which integrates a novel memory mechanism, so that during the denoising process, TRACE can retain and leverage intermediate results from previous steps to effectively reconstruct those hard-to-recover trajectory segments. Extensive experiments on multiple real-world datasets show that TRACE out- performs the state-of-the-art, offering>26% accuracy improvement without significant inference overhead. Our work strengthens the foundation for mobile and web-connected location services, ad- vancing the quality and fairness of data-driven urban applications. Code is available at: https://github.com/JinmingWang/TRACE CCS Concepts • Information systems→Incomplete data; Location based ser- vices; Temporal data. Keywords Diffusion Models, Trajectory Recovery, Location-based Services, Urban Mobility *Corresponding author. This work is licensed under a Creative Commons Attribution 4.0 International License. W ’26, Dubai, United Arab Emirates. © 2026 Copyright held by the owner/author(s). ACM ISBN 979-8-4007-2307-0/2026/04 https://doi.org/10.1145/3774904.3792461 ACM Reference Format: Jinming Wang, Hai Wang, Hongkai Wen, Geyong Min, and Man Luo*. 2026. TRACE: Trajectory Recovery with State Propagation Diffusion for Urban Mobility. In Proceedings of the ACM Web Conference 2026 (W ’26), April 13–17, 2026, Dubai, United Arab Emirates. ACM, New York, NY, USA, 9 pages. https://doi.org/10.1145/3774904.3792461 1 Introduction The proliferation of GPS-enabled devices has woven a digital fabric over our urban landscapes, generating vast streams of trajectory data. Location-based web services and smart city applications rely on high-quality trajectories - for navigation, ride-hailing, last-mile delivery, and traffic management. The promise of these technologies - optimized routes, efficient traffic management, and equitable ser- vice access - hinges on the availability of continuous, high-quality trajectory data. In the wild, however, trajectories recorded by mo- bile and edge devices are often sparse and irregular: sampling rates may drop to save energy, infrastructure coverage is uneven across regions, and device storage constraints limit logging. As a result, real-world trajectories are frequently sampled at low rates, yielding spatially skewed location points with large, uneven time gaps. As shown in Fig. 1, such fragmented paths can misrepresent actual movement, degrading downstream task accuracy and fairness. To bridge this gap, trajectory recovery serves as an essential pre-processing step, aiming to reconstruct dense and uniform tra- jectories from sparse and incomplete observations. This task is naturally conditional generation: given i) a sparse observation and i) a set of query timestamps, infer the missing locations, so as to ob- tain a complete and dense trajectory. While various methods have been developed, they face significant limitations. Map-matching techniques [15,16] are constrained by the availability of detailed digital road networks, which is not always accessible everywhere, while sequence models like RNNs [6,24] often suffer from error accumulation over long, unobserved segments. Transformer-based architectures [23,27] show promise by capturing long-range depen- dencies, but still struggle when spatio-temporal distribution of input locations is highly skewed. Generative models like VAEs [2,10,19] and GANs [5,9] have framed recovery as a conditional genera- tion task, thus improving fidelity of the generated trajectories, but VAEs can over-smooth or suffer posterior collapse, GANs risk mode arXiv:2603.19474v1 [cs.LG] 19 Mar 2026 W ’26, April 13–17, 2026, Dubai, United Arab Emirates.Jinming Wang, Hai Wang, Hongkai Wen, Geyong Min, and Man Luo Logistics & Delivery Traffic Managing Route Planning Data-Driven Applications Sparse And Irregular Trajectory Dense And Uniform Trajectory Figure 1: The trajectories collected from mobile edge devices are usually irregular and sparse, which fail to precisely de- scribe the behavior of moving agents, thus cannot be used in data-driven applications. collapse and unstable - and neither explicitly aggregates evidence across long, irregular gaps. Recently, diffusion models [7,17] have emerged as a dominant force in generative modeling, demonstrating remarkable success in related tasks including trajectory generation [26,28,29] and recovery of time-series [12]. Applying them naively to trajectory recovery, however, over- looks domain-specific irregularities: unlike typical time-series data, observations (i.e. of agent locations) arrive at uneven, often long time gaps, are spatially constrained by road topology and agent kinematics, and the query timestamps to be filled rarely align with the observed ones. In standard diffusion, each denoising step treats the input almost in isolation: a UNet-style denoising network [18] repeatedly re-encodes the same sparse evidence and static condi- tions (timestamps, masks, coarse priors) without carrying forward the intermediate features, i.e. what was already inferred. Early steps - dominated by low SNR noise - tend to wash out weak cues from distant observations; later steps then lack a consolidated memory to resolve long, ambiguous gaps, leading to over-smoothed trajec- tory segments, temporal jitter, or hallucinated detours — especially under fast sampling schedules. To overcome these challenges, we introduce TRACE, a novel diffusion-based trajectory recovery framework specifically engi- neered for the irregular spatio-temporal structure of trajectory data. At the core of TRACE is the State Propagation Diffusion Model (SPDM), which fundamentally redesigns the denoising processes to be memoryful rather than step-wise isolated. Given sparse observa- tions and a set of query timestamps, TRACE first builds a unified conditioning context that fuses observation/query masks, time gaps, and lightweight priors (e.g., linear interpolation for easy segments; optional context such as day-of-week or user/route IDs). The pro- posed SPDM then performs conditional denoising while carrying a compact, multi-scale hidden state across diffusion steps, i.e., its stateful denoising network acts as a memory mechanism, allowing multi-scale spatio-temporal features to be retained, accumulated, and reused across successive denoising steps. This propagated state summarizes geometry and motion cues discovered early and feeds them forward, turning a sequence of isolated denoising steps into a coherent, memoryful refinement process. Concretely, in TRACE a UNet backbone processes the conditioning context, while a step- aware state propagation module updates the hidden state to con- centrate capacity on long, uncertain gaps and avoid re-processing those stable regions. This not only avoids redundant computation but also provides richer, evolving context, i.e., knowledge from previous denoising steps can now serve as auxiliary conditions for subsequent steps, empowering the model to reconstruct challeng- ing, irregular segments with far greater accuracy. The technical contributions of this paper are summarized as follows: • To the best of our knowledge, we are the first to formalize trajectory recovery as conditional generation under severe spatial/temporal irregularity, and propose a novel diffusion- based framework, TRACE, that is explicitly designed to handle the sparse and irregular nature of real-world trajectory data. This provides a more reliable data foundation for a broad spec- trum of downstream applications, including location-based web services. • We introduce the State Propagation Diffusion Model (SPDM), a novel architecture that augments the standard diffusion models with a step-aware state propagation mechanism. SPDM carries compact, multi-scale states across the denoising process - by reusing this intermediate knowledge as conditioning, SPDM avoids redundant processing but concentrates capacity on the hard-to-recover segments while skipping the already stable ones. •We evaluate TRACE across multiple real-world datasets and sparsity regimes. Extensive experiments show that TRACE consistently outperforms strong baselines by significant mar- gins. Ablations confirm that the proposed SPDM mechanism is the key, achieving up to a 26.65% improvement in recovery accuracy over the standard diffusion model, establishing a new state-of-the-art without significant inference overhead. 2 Preliminaries 2.1 Problem Formulation Timestamps: Let푆=푠 1 ,푠 2 , ...푠 푁 denote a set of푁timestamps where푠 푖 ∈R + represents the exact time of a GPS recording. Times- tamps are ordered (푠 1 < 푠 2 < ...< 푠 푁 ) but are usually irregularly spaced, reflecting real-world constraints such as energy-saving sampling or signal loss. Trajectory: Given timestamps푆, a corresponding trajectory휏 푆 = 푝 푠 1 ,푝 푠 2 , ...푝 푠 푁 is a sequence of GPS coordinates, where each point 푝 푠 푖 records the longitude and latitude of the moving agent at time 푠 푖 be a location point recording longitude and latitude. Sample Interval: For consecutive timestamps푠 푖 and푠 푖+1 , the sam- ple intervalΔ푠 푖 = 푠 푖+1 −푠 푖 quantifies temporal sparsity. Small and constant sample intervals computed from푆and휏 푆 indicate a dense and uniform trajectory, otherwise the trajectory is sparse and ir- regular. Query: The query푄=푞 1 ,푞 2 , ...푞 푀 specifies푀timestamps where the corresponding GPS coordinates푝 푞 1 ,푝 푞 2 , ...푝 푞 푀 are to be in- ferred. Note that these coordinates form a complementary trajec- tory denoted as휏 푄 . Non-trivial trajectory recovery occurs when 푄∩ 푆=∅, requiring inference of locations at unobserved times. Trajectory Recovery: Given a trajectory휏with large or irregular sample intervals, its corresponding푆, a query푄, and other types TRACE: Trajectory Recovery with State Propagation Diffusion for Urban MobilityWWW ’26, April 13–17, 2026, Dubai, United Arab Emirates. (a) Merge Timestamps (e) Apply Denoising with SPDM Recovery Mask 필∈ℝ 푳×ퟏ Sparse Traj 흉 푺 Noise Points 흉 푻 푸 흉 푻 푺∪푸 ∈ℝ 푳×ퟐ Interpolated Traj 흉 푳 ∈ℝ 푳×ퟐ 푠 0 푠 3 푠 5 푞 1 푞 2 푞 4 Observed Timestamps 푺 Querying Timestamps 푸 01 0 101 푠 0 푞 1 푞 2 푠 3 푞 4 푠 5 (b) Fill Gaps with Noise 푺∪푸∈ℝ 푳×ퟏ (c) Include Prior Knowledge (d) Multi-modal Contexts 푪 ퟏ 푪 ퟐ 푪 풌 푬 ퟏ ∈ℝ 푳×푫 ퟏ 푬 ퟐ ∈ℝ 푳×푫 ퟐ 푬 풌 ∈ℝ 푳×푫 풌 Embedder 1 Embedder k Embedder 2 ഥ 풉 푻 =ퟎ 푨 푻 ∈ℝ 푳×(ퟔ+푫 ퟏ +⋯+푫 풌 ) State Propagation Pipeline ഥ 풉 풕−ퟏ ←푴푪푮푹푼( ഥ 풉 풕 ,풉 풕−ퟏ ,풕) Denoising Step 푻−ퟏ Denoising Step 푻 Denoising Step 1 푨 푻−ퟏ 푨 ퟎ ഥ 풉 푻−ퟏ ഥ 풉 ퟏ Dense Traj 흉 ퟎ 푺∪푸 ∈ℝ 푳×ퟐ 풉 푻−ퟏ 풉 푻−ퟐ Concat 푺∪푸 흉 푻 푺∪푸 필 흉 푳 푬 ퟏ 푬 ퟐ 푬 풌 ⋮ ⋮ ⋮ ⋮ ⋯ Figure 2: TRACE Framework. (a) Merge querying timestamps into observed timestamps in chronological order. (b) Initialize missing points with Gaussian noise. (c) Prior knowledge are included to explicitly tell the model where to recover. (d) The embedding of multi-modal contextual information. (e) The denoising process with SPDM that generates the dense trajectory. of auxiliary contextsCsuch as date, user ID and waybills. The objective of the trajectory recovery task is to find a function that produces prediction ˆ 휏 푄 and minimizes the error between ˆ 휏 푄 and the ground-truth 휏 푄 . 2.2 Diffusion Models Denoising diffusion probabilistic models (DDPM) [7] is constructed on the foundation of a forward Markov chain, which progressively corrupts clean data by adding Gaussian noise at each discrete step. The model trains a denoising neural network to accurately predict the noise introduced at each step of the forward process. Once trained, the network can be used to generate clean data such as trajectory points, starting from pure Gaussian noise. Given a clean data sample x 0 drawn from the training distribution x 0 ∼ 푝 data ( x). The forward Markov process is defined by a noise schedule훽 1 ,훽 2 , . . .,훽 푇 , specifying the variance of Gaussian noise added at each time step. At step푡, the noisy data x 푡 is obtained through a conditional distribution: 푞(x 푡 |x 푡−1 )=N(x 푡 ; √︁ 1− 훽 푡 x 푡−1 ,훽 푡 I) 푥 푡 = √ 훼 푡 푥 푡−1 + √︁ 훽 푡 휀 푡−1:푡 (1) where훼 푡 =1− 훽 푡 and휀 푡 :푡+1 ∼ N(0,1)is single-step noise added. Alternatively, x 푡 can be expressed directly in terms of the clean data initial x 0 as: 푞(x 푡 |x 0 )=N(x 푡 ; √ ̄ 훼 푡 x 0 ,(1− ̄ 훼 푡 )I) 푥 푡 = √ ̄ 훼 푡 푥 0 + √ 1− ̄ 훼 푡 휀 0:푡 (2) where ̄ 훼 푡 = Î 푖= 1 푡 훼 푖 , the notation휀 0:푡 ∼N(0,1)is the multi-step noise, and x 1 ,x 2 , . . .,x 푇 represent the intermediate states. As푇be- comes sufficiently large, the forward process effectively transforms the data distribution푝 data (x 0 )into a standard Gaussian distribution. To enable the reverse process, the model learns a denoising network parameterized by휽, which approximates the reverse tran- sition distribution푝 휽 (x 푡−1 |x 푡 )by minimizing the divergence from the true posterior푞(x 푡−1 |x 푡 ,x 0 ). This is achieved by training a noise predictor휺 0:푡,휃 (x 푡 ,푡), which estimates the noise휺 0:푡 introduced at each step. During inference, the generation process involves iterative sam- pling of the learned reverse transitions푝 휽 , guided by the denoising network휺 휃 . Additionally, the framework can incorporate condi- tional signals to improve generation quality. For instance, when conditioning on auxiliary information c, the model can be extended to train a conditional denoising network휀 0:푡,휃 (x 푡 ,푡,c)using an anal- ogous loss function, thus enhancing the fidelity and control of the generated data. There exist other formulations of diffusion processes that at- tempt to improve sampling efficiency through fewer generation steps. Such models include DDIM [20], score-based models [21], or PNDM [11]. However, as long as the model still involves a multi- step denoising procedure, the redundancy persists, and they can be used jointly with our SPDM. 3 Methodology 3.1 TRACE Overview Trajectory recovery requires the integration of heterogeneous data, including trajectories, timestamps, contextual metadata (e.g., date and trip duration), and other user-specific features, which must be processed to a unified representation compatible with diffusion frameworks. These input data — serving as the condition for tra- jectory recovery — are illustrated in Fig. 2(a-d), and the details will be discussed in Sec. 3.2. The aggregated representation is then fed into the denoising process, shown in Fig. 2(e), which iteratively reconstructs the querying locations by removing noise from an initial Gaussian noise. Crucially, TRACE replaces the conventional W ’26, April 13–17, 2026, Dubai, United Arab Emirates.Jinming Wang, Hai Wang, Hongkai Wen, Geyong Min, and Man Luo denoising pipeline with SPDM to address the spatio-temporal irreg- ularity problem. It introduces a state propagation mechanism that propagates multi-scale spatio-temporal features across denoising steps. The architecture of SPDM, including the specially designed neural networks used in each denoising step and the state propa- gation pipeline, will be explained in Sect. 3.3. Lastly, the training procedure for TRACE is very different from the training of tra- ditional diffusion models. This difference stems from the unique inter-step dependency introduced by the state propagation pipeline, and Sec. 3.4 will present the details of new training algorithm. 3.2 Condition Aggregation Stage To enable robust trajectory recovery under spatio-temporal irregu- larities, we integrate various prior knowledge to guide the recovery process. The sparse trajectory휏 푆 and associated timestamps푆are the primary observation that the recovery process can rely on, while the query timestamps푄specifies target timestamps with unknown coordinates휏 푄 , which is initialized as Gaussian noise. The query timestamps can be manually or procedurally selected, ensuring highly flexible and customizable recovery pattern. To synchronize observation and query, we merge푆and푄into a chronologically or- dered sequence denoted푆∪푄of length퐿. Similarly, we construct an assembled trajectory휏 푆∪푄 following the same chronological order. To explicitly indicate which coordinates in휏 푆∪푄 should undergo recovery, we label the missing point indices with a binary mask M ∈ 0,1 퐿 , whereM 푘 =1 indicates that the푘-th timestamp in 푆∪푄comes from푄and the푘-th coordinate in휏 푆∪푄 comes from휏 푄 . Moreover, we realize that simple linear interpolation is sufficient to recover some trajectory segments, e.g., straight paths, smooth and dense trajectory segments. Therefore, we introduce a linearly interpolated trajectory푙푒푟푝(휏)over푆 ∪ 푄to augment the input. Concretely, it fills the unknown coordinates with linearly interpo- lated values inferred from nearby points, providing the model with a geometrically plausible prior. Industrial applications usually provide a variety of heteroge- neous contexts퐶 1 ,퐶 2 , ...퐶 퐾 associated with trajectories, e.g. week- day, date, user ID, waybill details. Since the embedding part is not the focus of our method, we assume that one module is specially designed for each context, where each module consists of several simple operations such as linear or convolutional layers. In the end, each context퐶 푖 is converted to a푑 푖 dimensional sequential feature 퐸 푖 ∈R 퐿×푑 푖 aligned with the trajectories. At the end of this stage, the final aggregated feature tensor 퐴 ∈R 퐿×퐷 is constructed as: 퐴← 퐶표푛푐푎푡푒푛푎푡푒(휏 푆∪푄 ,푆∪푄,M,푙푒푟푝(휏),퐸 1 , ...퐸 퐾 )(3) where퐷is the final dimensionality of the features. To conclude, this formulation ensures sparse and irregular query regions, enriched with various prior knowledge and contextual information, enabling adaptive recovery of complex trajectory segments. 3.3 Denoising With SPDM Traditional diffusion models iteratively denoise the input noise푥 푇 over푇steps, transforming it into clean data푥 0 . In the proposed framework, our input is denoted as퐴 푇 , which will be recovered to 퐴 0 through intermediate steps퐴 푇−1 ,퐴 푇−2 ...퐴 1 . In fact, although ഥ 풉 풕 ഥ 풉 풕−ퟏ MCGRU UNet 풇 풕−ퟏ;풊 푨 풕 ො 휺 ퟎ:풕 Res 1D Concat 풕 ത 풇 풕;풊 Split Resize Conv input output CGRU 풕 Linear Linear ത 풇 풕;풊 ത 풇 풕−ퟏ;풊 UnetBasic Block CGRU With t Figure 3: The architecture of the modified UNet (upper, red) and the proposed MCGRU (lower, green). 퐴 푡 is the input for step푡, the denoising operation only applies to휏 푄 푡 , which is a portion inside퐴 푡 corresponding to the noisy coordinates, while the other portion of퐴 푡 is fixed throughout the entire denoising process. A novel formulation for the denoising step is essential to enable effective state propagation in SPDM. We begin by defining the states: letℎ 푡 be the single-step hidden state that encapsulates the local features extracted at step푡, and let ̄ ℎ 푡 be the multi-step hidden state containing all useful knowledge from previous steps[푡+1,푇]. To enhance feature richness and reusability, we define both single- and multi-step hidden states as multi-scale sequential features, i.e.ℎ 푡 = 푓 푡 ;1 , 푓 푡 ;2 , ...푓 푡 ;푏 and ̄ ℎ 푡 = ̄ 푓 푡 ;1 , ̄ 푓 푡 ;2 , ... ̄ 푓 푡 ;푏 , where푏is the number of sequential features in the state. Under this formulation, for each denoising step푡, the denoising network takes퐴 푡 , ̄ ℎ 푡 and 푡as inputs and producesℎ 푡−1 along with the predicted noise ˆ 휀 0:푡 . Subsequently, the state propagation network accepts ̄ ℎ 푡 ,ℎ 푡−1 and푡 to output the updated state ̄ ℎ 푡−1 . Based on this formulation, our neural network architecture is implemented accordingly. The denoising network in TRACE is built upon a UNet architecture, which inherently supports multi-scale feature extraction and aligns well with our state definitions. To accommodate state propagation, the UNet is modified as illustrated in the upper part of Fig. 3. Specifically, the UNet comprises exactly 푏basic blocks — matching the number of features in the states. Building on top of traditional UNet blocks, each modified block푖 also processes additional input ̄ 푓 푡 ;푖 ∈ ̄ ℎ 푡 to generate the correspond- ing output feature푓 푡−1;푖 . Finally, all output features of the푏UNet blocks are assembled to ℎ 푡−1 . After we getℎ 푡−1 generated by the denoising UNet, the state propagation network is responsible for updating the cumulative knowledge and producing ̄ ℎ 푡−1 . Concretely, we implement a novel Multi-scale Convolutional Gated Recurrent Unit (MCGRU) with diffusion step awareness, shown in the lower part of Fig. 3. MCGRU contains푏CGRU blocks in parallel, mirroring the UNet’s hierarchy. Building on top of traditional CGRU, the modified version also takes into account the diffusion time푡, which enhances its compatibility with the multi-step denoising pipeline. TRACE: Trajectory Recovery with State Propagation Diffusion for Urban MobilityWWW ’26, April 13–17, 2026, Dubai, United Arab Emirates. 3.4 Training of TRACE In traditional diffusion model training for trajectory recovery, a dense trajectory with corresponding timestamps is provided, which are then divided into the observation part (푆,휏 푆 ) and the query part (푄,휏 푄 0 ). With a randomly selected time step푡 ∈ [1,푇], the forward diffusion process is applied to generate휏 푄 푡 by adding noise 휀 0:푡 ∈ N(0,1). Then, the denoising network is trained to predict 휀 0:푡 . However, during TRACE training, the SPDM denoising step 푡requires the state ̄ ℎ 푡 from previous iterations. This disables ran- dom step sampling and necessitates a sequential training paradigm, where each data sample iterates through all steps from푡= 푇to 푡= 1, so that ̄ ℎ 푡 is always available. The above solution makes the training of the denoising network possible, yet we also have to train the state propagation network, i.e., each state ̄ ℎ 푡 must contain useful knowledge for the subsequent steps. This requirement imposes two sub-constraints: i) the denois- ing network at step푡should output appropriateℎ 푡−1 and i) the state propagation network should effectively aggregate or filter knowledge to yield beneficial ̄ ℎ 푡−1 . Intuitively, the ideal strategy is to jointly train the entire denoising process, with the overall objective being: min 휃 푇 ∑︁ 푡=1 푀푆퐸( ̄ 휀 0:푡 ,휀 0:푡 )(4) where휃represents the learnable parameters of the models. How- ever, incorporating all tens to hundreds of denoising steps in each training iteration is computationally impractical. Instead, we parti- tion the denoising process into smaller segments comprising several consecutive steps, e.g. including two steps in one training iteration: min 휃 (푀푆퐸( ˆ 휀 0:푡 ,휀 0:푡 )+ 푀푆퐸( ˆ 휀 0:푡−1 ,휀 0:푡−1 ))(5) Another challenge in training TRACE arises from the noise sam- pling process. The standard diffusion model training algorithm sam- ples휀 0:푡 ∼N(0,1)independently in each training iteration, because it does not involve the complex inter-step dependencies introduced by the state-propagation mechanism. In contrast, our algorithm it- erates푡from푇to 1 for each data sample, and one training iteration involves at least two denoising steps. Consequently, the training of TRACE cannot ignore the dependency among noises, i.e., multi-step noises휀 0:푡 and휀 0:푡−1 . To derive a set of dependent noise samples, we first sample a list of single-step noises휀 0:1 ,휀 1:2 , ...,휀 푇−1:푇 , where each휀 푡−1:푡 ∼ N(0,1)is always independent. Then, starting with 퐴 0 , we recursively apply single-step diffusion forward using Eq. 1 to obtain퐴 1 to퐴 푇 . Next, we apply following equation to compute the multi-step noise 휀 0:푡 for each 푡 : 휀 0:푡 = 퐴 푡 − √ ̄ 훼 푡 퐴 0 √ 1− ̄ 훼 푡 (6) This equation is a changed form of Eq. 2, it ensures that the depen- dency among the noises is respected within each training iteration, as well as during the entire life cycle of a data sample. Now, the gen- eral principles of training TRACE have finally been established, and a naive procedure of the proposed training algorithm is presented in Algorithm 1. Algorithm 1 Training One Sample (2-step) 0: Input: 푆, 푄, 휏 푆 , 휏 푄 0 , 푙푒푟푝(휏),M, 퐶 1 , 퐶 2 , · 퐶 퐾 . 1: Generate휀 푡−1:푡 |푡 ∈ [1,푇]. 2: Derive휀 0:푡 |푡 ∈ [1,푇] with Eq. 6. 3: Apply diffusion to 휏 푄 0 to get휏 푄 푡 |푡 ∈ [1,푇]. 4: ℎ 푇 ← 0 5: for 푡 ← 푇 to 1 do 6: 퐸 푖 ← 퐸푚푏푒푑 푖 (퐶 푖 ) for 푖 ∈ 1, ...퐾. 7:Obtain 퐴 푡−1 ,퐴 푡 using concatenation as in Eq. 3. 8: // Train the first step 9: ˆ 휀 0:푡 ,ℎ 푡−1 ← UNet(퐴 푡 , ̄ ℎ 푡 ,푡) 10: // Update hidden state 11: ̄ ℎ 푡−1 ← MCGRU( ̄ ℎ 푡 ,ℎ 푡−1 ,푡) 12: // Train the second step 13: ˆ 휀 0:푡−1 ,ℎ 푡−2 ← UNet(퐴 푡−1 , ̄ ℎ 푡−1 ,푡 − 1) 14:Compute loss with Eq. 5. and do optimization. 15: end for Batch Index Training Iteration Batch Index Batch Index Lifespan of a Data Sample During Training Load this t=T t=T-1 t=T-2 ... ... ... ... ... ... t=3 t=2 t=1 Load next Figure 4: Three ways of batch management. Top: Shared푡 among all samples in a batch. Middle: Different푡for each sample with little offset. Bottom: Uniformly distributed푡 over the range [1,푇]. Although the above design is sufficient for training, practical im- plementation demands additional techniques. As mentioned above, each data sample in the training batch undergoes푇training itera- tions with a synchronized countdown of푡across the entire batch (Fig. 4, top). Consequently, the model overfits to a narrow range of denoising steps and cannot generalize to the full range of푡. To ad- dress this, we implement a dynamic batch management technique that assigns a private푡to each sample in a batch, with an offset among them (Fig. 4, middle). We can further promote generalization across the entire range of푡by uniformly assigning푡 ∈ [1,푇]to each data sample, as illustrated in Fig. 4 bottom. In this scheme, the 푡of each sample is updated independently, and a new sample is loaded in-place once its 푡 reaches 0. 4 Experiments 4.1 Experimental Settings Datasets. We evaluate TRACE on three datasets. Two of them comprise dense and smooth taxi trajectories collected in Xi’an city and Chengdu city, China. To assess the model’s performance across W ’26, April 13–17, 2026, Dubai, United Arab Emirates.Jinming Wang, Hai Wang, Hongkai Wen, Geyong Min, and Man Luo MetricMethod Dataset Xi’anChengduLogistics MSE↓DeepMove5.29726.85730.060 (×10 −3 )AttnMove0.0680.2343.201 PriSTI0.0190.2922.206 DT + RP0.3266.9194.250 TRACE w/o SPDM0.0170.2022.025 TRACE0.0100.1590.449 NDTW↓DeepMove36.295169.47072.814 (×10 −3 ) AttnMove2.2575.45822.652 PriSTI1.1743.97512.656 DT + RP8.47328.77313.006 TRACE w/o SPDM1.1563.80911.062 TRACE1.1543.5975.520 MAE↓DeepMove31.695149.16994.151 (×10 −3 )AttnMove3.1197.25033.744 PriSTI1.5872.3972.233 DT + RP3.195112.21221.458 TRACE w/o SPDM1.3524.64621.796 TRACE1.3351.7252.121 Table 1: The overall comparisons of six models with three metrics on three dataset. The trajectory length is 512 points, with 50% GPS points erased and the models are asked to recover the erased points. 5.00e-04 5.05e-02 1.01e-01 0.000.040.080.120.16 Spatial Irregularity 1.02e-04 2.02e-04 3.02e-04 4.02e-04 MSE 5.00e-04 5.05e-02 1.01e-01 0.0000.0010.002 Temporal Irregularity 1.02e-04 2.02e-04 3.02e-04 4.02e-04 MSE DeepMove AttnMove DT + RP PriSTI TRACE w/o SP TRACE Figure 5: The trend of recovery errors as the trajectory gets more irregular in terms of spatial and temporal distributions. There are breaks in y-axis because the magnitude of MSE differs a lot for different methods. Tested on Xi’an with 70% points erased and trajectory length of 512 points. diverse mobility patterns and metadata, we also include a dataset from last-mile logistics, capturing courier movements in residential areas with irregular sampling intervals. For conditional contexts and metadata, we extract the trajectory starting date and time, the moving agent’s ID, and trip duration. In addition, the logistics data set includes information for package delivery and receiving. All GPS coordinates undergo z-score normalization, and the time stamp sequence of each trajectory is rescaled to the range [0, 1]. Metrics. Trajectory recovery focuses on the exact spatial similar- ity between the recovered coordinates and the ground truth ones. Therefore, we adopt three evaluation metrics: The Mean Squared Error (MSE), the Mean Absolute Error (MAE), and the Normalized Dynamic Time Warping (NDTW). Lower values are preferable for all metrics. Baselines. We evaluate TRACE against four baselines. i) Deep- Move [4]: An RNN-based trajectory forecasting model adapted for recovery via auto-regressive imputation. i) AttnMove [27]: A Transformer-based approach fusing intra-/inter-trajectory contexts via self- and cross-attention. i) PriSTI [12]: A diffusion model for time-series imputation, modified for trajectories by treating GPS co- ordinates as 2D signals. iv) DT+RP: Combines Diff Traj [28] (UNet- based trajectory generation) with RePaint [13] (image inpainting via diffusion), which is capable of adapting content generation dif- fusion models to recovery tasks. Additionally, we ablate TRACE by replacing SPDM with a standard diffusion pipeline (TRACE w/o SPDM), which aims to test the effectiveness of SPDM. 4.2 Results 4.2.1Overall Comparison. Table 1 summarizes the performance of TRACE against the baselines. Non-diffusion methods (DeepMove, AttnMove) exhibit significantly higher errors, underscoring the limitations of auto-regressive and attention-based architectures in handling sparse, irregular trajectories. In contrast, Diffusion-based baselines show marked improvements except for the approach combining Diff Traj with RePaint, which emphasizes the need to design and train diffusion models specific for recovery tasks. The ablation study (TRACE without SPDM) highlights the need for state propagation, while competitive, it underperforms TRACE in most test cases, highlighting the importance of SPDM. 4.2.2Addressing Spatio-Temporal Irregularity. We prove the effec- tiveness of SPDM on addressing the challenge of spatio-temporal irregularity inherent to trajectory data. Given a trajectory to be recovered휏 푆 , we compute two metrics to measure the spatial and temporal irregularities: i) the standard deviation of distances be- tween consecutive point pairs and i) the standard deviation of sample intervals. A large variance indicates that the trajectory is ir- regular and skewed; otherwise, the trajectory points are uniformly distributed. The results are shown in Fig. 5. As the trajectory gets increasing irregular, all methods tend to have worse performance. However, the proposed TRACE is apparently more robust to such irregularities. When comparing TRACE with and without SPDM, we can see that the SPDM achieves significant improvement when dealing with irregular trajectories, showcasing the effective and successful design of SPDM. 4.2.3Scalability. The length and sparsity of trajectories can vary greatly in industrial scenarios. In this experiment, we study the scalability of the proposed method with different trajectory lengths and sparsity levels. As illustrated in Fig. 6, we first select the three strongest methods: PriSTI, TRACE without SPDM, and TRACE. The left column presents three error metrics with trajectory length varying from 64 to 512, and the right column shows the slowly increasing errors as the trajectory gets more sparse. The sparsity is measured by the percentage of points removed from a dense trajectory. As we can see, the plots exhibit parallel trends, while TRACE consistently outperforms other methods. TRACE: Trajectory Recovery with State Propagation Diffusion for Urban MobilityWWW ’26, April 13–17, 2026, Dubai, United Arab Emirates. 100200300400500 Length −12.0 −11.4 −10.8 −10.2 −9.6 log MSE 100200300400500 Length −6.4 −6.2 −6.0 −5.8 −5.6 log NDTW 100200300400500 Length −10.8 −10.2 −9.6 −9.0 log JSD 0.30.40.50.60.7 Sparsity −12.0 −11.2 −10.4 −9.6 log MSE 0.30.40.50.60.7 Sparsity −6.6 −6.3 −6.0 −5.7 −5.4 log NDTW 0.30.40.50.60.7 Sparsity −11.4 −10.8 −10.2 −9.6 −9.0 log JSD TRACETRACE w/o SPDMPriSTI Figure 6: The comparison among sparsity levels and trajec- tory lengths. MethodSteps TimeMSENDTWMAE (s) ×10 −3 ×10 −3 ×10 −3 500119.1760.04322.013.068 DDIM5112.360.03492.8682.221 266.650.03512.8972.239 113.180.04903.0832.365 DDIM500129.570.03562.7502.502 With5113.020.02722.3321.847 State 267.030.02562.3201.822 Prop113.580.02602.4201.914 Table 2: The comparison of recovery time and quality with and without state propagation pipeline. Tested on Xi’an with 70% points erased and trajectory length of 512 points. 4.2.4 Efficiency. DDIM [20] is a variant in the diffusion model family that can speed up recovery by skipping some intermedi- ate denoising steps, trading result quality for faster recovery. We compare the recovery time and quality of TRACE using DDIM with or without state propagation to demonstrate the superior efficiency of TRACE, and the results are shown in Table 2. The denoising times are recorded with 100 samples in a batch on a NVIDIA GeForce RTX 3070 Ti Laptop GPU. For plain DDIM, the 11 denoising steps version is 119.176/3.18=37.48 faster than the full 500-step version, but the recovery quality is decreased by(0.0490−0.0432)/0.0490×100%=−11.84%. In contrast, build- ing a state propagation pipeline on top of the 11-step DDIM in- troduces only 12% more parameters and 0.4 seconds of running MethodParametersMACsFLOPs TRACE6.28M1.13G2.28G TRACE w/o SPDM6.17M0.99G2.01G DT + RP9.10M0.63G1.27G PriSTI7.32M3.24G6.51G AttnMove 8.15M3.94G7.92G DeepMove10.11M1.94G9.93G Table 3: The number of parameters, multiply-accumulate operations (MACs), and floating point operations (FLOPs) of all methods. Average SpeedMoving Distance (m/s)(km) Ground Truth11 Sparse Traj0.61800.5788 Recovered Traj0.82520.8480 Table 4: The moving speed and distance estimations com- puted from trajectories after normalization. time, and also improves the upper bound of recovery quality by (0.0349−0.0256)/0.0349×100%=26.65%. This indicates that the proposed SPDM can be jointly used with other diffusion sampling techniques that accelerate recovery, while achieving even higher recovery quality. We also provide a table describing the computational cost of all baselines, shown as Table 3. In particular, we present the number of parameters, multiply-accumulate operations (MACs), and floating point operations (FLOPs) of all baselines. We notice that TRACE is very parameter-efficient due to its efficient design. The SPDM only introduces a small number of parameters and computations to our diffusion model, this further proves its effectiveness. One thing to note is that the diffusion models require iteratively denoising, so the cost should be multiplied by the denoising steps, making them computationally costly compared to non-diffusion-based baselines. 4.2.5Effectiveness of Training Algorithms. We analyze the effective- ness of the proposed training algorithms and techniques introduced in Sect. 3.4. First, we want to study the effect of different number of denoising steps included in a single training iteration, as shown in Fig. 7a. The result indicates that the number of steps trained in each iteration do not have a great influence on the final performance, and 2-step is preferable since it requires the least training time. Second, we show the importance of dynamic batch management during training. As we can see in Fig. 7b, the uniformly distributed 푡for each data sample within a batch leads to the best convergence, while a shared푡for the entire batch results in the worst recovery quality. This study highlights the necessity of the proposed training algorithm. 4.3 Case Study A case study was conducted using courier trajectories in a logistics scenario. Analyzing the speed and distance of moving people or vehicles is crucial for many location-based data-driven Web appli- cations, such as navigation and last-mile delivery. The trajectory W ’26, April 13–17, 2026, Dubai, United Arab Emirates.Jinming Wang, Hai Wang, Hongkai Wen, Geyong Min, and Man Luo 0 0.4 0.8 1.2 1.6 2-step3-step4-step MSE (x 1e - 5) Xi'anLogisticsChengdu (a) Steps in each iteration 0 0.8 1.6 2.4 3.2 SharedConsecutiveUniform MSE (x 1e - 5) Xi'anLogisticsChengdu (b) Batch managing methods Figure 7: The comparisons among: (a) Number of denoising steps included in single training iteration. (b) Three ways of batch managing introduced in Fig 4. Tested on Xi’an with 70% points erased and trajectory length of 512 points. (a) Ground truth trajectory(b) Sparse trajectory (c) Recovered trajectory Figure 8: The dataset of: (a) Dense trajectories as ground truth. (b) Trajectories with the same sparsity as usual logistics scenario. (c) Recovered trajectories using TRACE. of the user is usually collected to compute the speed and the mov- ing distance. However, these trajectories are usually sparse due to constrained device-server communication rates, limited sampling rates, and unstable signal conditions. In addition, storing massive amounts of trajectories for large-scale web-service can be costly. All these limitations hinder accurate computation of user speed and moving distance. We collected very dense trajectories in a crowded apartment region. A subset was sampled from each dense trajectory according to a normal last-mile delivery sampling interval, resulting in only 30% of the points remaining. TRACE was trained to recover these sparse trajectories to dense ones. We used the moving distances and speed computed from the very dense trajectories as ground truth and then computed the estimation using the sparse and the recovered trajectories. For more direct comparison, we normalized the ground truth speed to 1 meter per second and the moving distance to 1 kilometer, and the results are shown in Table 4. We observed that the estimations using sparse trajectories captured 60% of the actual speed and distance, while the recovered trajectories increased this percentage to around 83%. An illustration of the trajectory before and after recovery is shown in Figure 8. 5 Related Work Trajectory Recovery. Numerous efforts have been made to in- crease the accuracy of trajectory recovery. Certain map-based meth- ods [3,8,14,16] utilize a predefined set of points of interest (POI) or a pre-collected road network as strong prior knowledge. In contrast, DHTR [25] performs free-space trajectory recovery using RNN, which eliminates the need for road network information. However, for very long and sparse trajectories, RNNs struggle to capture dependencies among points. TrajBERT [19] and AttnMove [27] are methods based on Transformers [23], which excel in capturing long-term dependencies within sequence. Diffusion-based Trajectory Generation/Recovery. Existing approaches have successfully applied diffusion models to generate or recover time series data, including CSDI [22] and SSSD [1]. In particular, PriSTI [12] is a diffusion-based method designed for time series re- covery, using spatio-temporal conditions and geographic factors for more accurate recovery. Moreover, another thread of existing work uses diffusion models for trajectory generation rather than recovery. For example, Diff Traj [28] employs UNet [18] to perform denoising steps with several simple contexts such as trajectory length, starting and ending points. More recently, some diffusion-based methods utilize road networks to guide the generation process, such works include Diff-RNTraj [26] and ControlTraj [29]. 6 Conclusion In conclusion, this paper presents TRACE, a diffusion-based frame- work for robust trajectory recovery. At its core, TRACE incorporates the State Propagation Diffusion Model (SPDM), which introduces a novel state propagation pipeline that enables knowledge shar- ing across denoising steps. This design effectively addresses the spatio-temporal irregularities of real-world trajectory data, leading to more accurate reconstruction of sparse segments while reducing redundant computation. Extensive experiments on multiple real- world datasets confirm the superior robustness and efficiency of TRACE over state-of-the-art methods. By providing precise and effi- cient trajectory recovery, our work enhances the data reliability for large-scale location-based web services and smart city applications. TRACE: Trajectory Recovery with State Propagation Diffusion for Urban MobilityWWW ’26, April 13–17, 2026, Dubai, United Arab Emirates. References [1]Juan Miguel Lopez Alcaraz and Nils Strodthoff. 2023.Diffusion-based Time Series Imputation and Forecasting with Structured State Space Models. arXiv:2208.09399 [cs.LG] [2]Xinyu Chen, Jiajie Xu, Rui Zhou, Wei Chen, Junhua Fang, and Chengfei Liu. 2021. TrajVAE: A Variational AutoEncoder model for trajectory generation. Neurocomputing 428 (2021), 332–339. doi:10.1016/j.neucom.2020.03.120 [3]Yuqi Chen, Hanyuan Zhang, Weiwei Sun, and Baihua Zheng. 2022. RNTrajRec: Road Network Enhanced Trajectory Recovery with Spatial-Temporal Trans- former. arXiv:2211.13234 [cs.LG] [4]Jie Feng, Yong Li, Chao Zhang, Funing Sun, Fanchao Meng, Ang Guo, and Depeng Jin. 2018. DeepMove: Predicting Human Mobility with Attentional Recurrent Networks. In Proceedings of the 2018 World Wide Web Conference (Lyon, France) (W ’18). International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, CHE, 1459–1468. doi:10.1145/3178876.3186058 [5]Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde- Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative Adversarial Networks. arXiv:1406.2661 [stat.ML] [6]Alex Graves. 2012. Supervised Sequence Labelling. Vol. 385. Springer, Berlin, Heidelberg, 5–13. doi:10.1007/978-3-642-24797-2_2 [7]Jonathan Ho, Ajay Jain, and Pieter Abbeel. 2020. Denoising Diffusion Probabilistic Models. CoRR abs/2006.11239 (2020). arXiv:2006.11239 https://arxiv.org/abs/ 2006.11239 [8]Fujin Hou, Xiaowei Lan, Jingrong Chen, Yuhuan Dong, Xucai Zhuang, and Jianqing Wu. 2021. A Sparse Taxi Trajectory Data Recovery and Calibrate Algorithm. In 2021 China Automation Congress (CAC). IEEE, 5414–5419. doi:10. 1109/CAC53003.2021.9728679 [9]Wenjun Jiang, Wayne Xin Zhao, Jingyuan Wang, and Jiawei Jiang. 2023. Con- tinuous Trajectory Generation Based on Two-Stage GAN. Proceedings of the AAAI Conference on Artificial Intelligence 37, 4 (Jun. 2023), 4374–4382. doi:10.1609/aaai.v37i4.25557 [10] Diederik P Kingma and Max Welling. 2022. Auto-Encoding Variational Bayes. arXiv:1312.6114 [stat.ML] [11] Luping Liu, Yi Ren, Zhijie Lin, and Zhou Zhao. 2022. Pseudo Numerical Methods for Diffusion Models on Manifolds. arXiv:2202.09778 [cs.CV] https://arxiv.org/ abs/2202.09778 [12]Mingzhe Liu, Han Huang, Hao Feng, Leilei Sun, Bowen Du, and Yanjie Fu. 2023. PriSTI: A Conditional Diffusion Framework for Spatiotemporal Imputation. In 2023 IEEE 39th International Conference on Data Engineering (ICDE). IEEE, 1–10. doi:10.1109/icde55515.2023.00150 [13] Andreas Lugmayr, Martin Danelljan, Andrés Romero, Fisher Yu, Radu Tim- ofte, and Luc Van Gool. 2022. RePaint: Inpainting using Denoising Diffu- sion Probabilistic Models. CoRR abs/2201.09865 (2022), –. arXiv:2201.09865 https://arxiv.org/abs/2201.09865 [14] Junwei Ma, Chao Yang, Shiwen Mao, Jian Zhang, Senthilkumar CG Periaswamy, and Justin Patton. 2022. Human Trajectory Completion with Transformers. In ICC 2022 - IEEE International Conference on Communications. IEEE, 3346–3351. doi:10.1109/ICC45855.2022.9838743 [15]Paul Newson and John Krumm. 2009. Hidden Markov map matching through noise and sparseness. In Proceedings of the 17th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems (Seattle, Washington) (GIS ’09). Association for Computing Machinery, New York, NY, USA, 336–343. doi:10.1145/1653771.1653818 [16]Huimin Ren, Sijie Ruan, Yanhua Li, Jie Bao, Chuishi Meng, Ruiyuan Li, and Yu Zheng. 2021. MTrajRec: Map-Constrained Trajectory Recovery via Seq2Seq Multi- task Learning. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining (Virtual Event, Singapore) (KDD ’21). Association for Computing Machinery, New York, NY, USA, 1410–1419. doi:10.1145/3447548. 3467238 [17] Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. 2021. High-Resolution Image Synthesis with Latent Diffusion Models. CoRR abs/2112.10752 (2021), 1–10. arXiv:2112.10752 https://arxiv.org/abs/2112. 10752 [18]Olaf Ronneberger, Philipp Fischer, and Thomas Brox. 2015. U-Net: Convolutional Networks for Biomedical Image Segmentation. CoRR abs/1505.04597 (2015), 1–10. arXiv:1505.04597 http://arxiv.org/abs/1505.04597 [19]Junjun Si, Jin Yang, Yang Xiang, Hanqiu Wang, Li Li, Rongqing Zhang, Bo Tu, and Xiangqun Chen. 2024. TrajBERT: BERT-Based Trajectory Recovery With Spatial-Temporal Refinement for Implicit Sparse Trajectories. IEEE Transactions on Mobile Computing 23, 5 (2024), 4849–4860. doi:10.1109/TMC.2023.3297115 [20] Jiaming Song, Chenlin Meng, and Stefano Ermon. 2020. Denoising Diffusion Implicit Models. CoRR abs/2010.02502 (2020), 1–10. arXiv:2010.02502 https: //arxiv.org/abs/2010.02502 [21]Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. 2020. Score-based generative modeling through stochastic differential equations. arXiv preprint arXiv:2011.13456 1 (2020), –. [22]Yusuke Tashiro, Jiaming Song, Yang Song, and Stefano Ermon. 2021. CSDI: Con- ditional Score-based Diffusion Models for Probabilistic Time Series Imputation. arXiv:2107.03502 [cs.LG] [23] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017.Attention Is All You Need. CoRR abs/1706.03762 (2017), 1–10. arXiv:1706.03762 http: //arxiv.org/abs/1706.03762 [24] Jingyuan Wang, Ning Wu, Xinxi Lu, Wayne Xin Zhao, and Kai Feng. 2021. Deep Trajectory Recovery with Fine-Grained Calibration using Kalman Filter. IEEE Transactions on Knowledge and Data Engineering 33, 3 (2021), 921–934. doi:10. 1109/TKDE.2019.2940950 [25]Jingyuan Wang, Ning Wu, Xinxi Lu, Wayne Xin Zhao, and Kai Feng. 2021. Deep Trajectory Recovery with Fine-Grained Calibration using Kalman Filter. IEEE Transactions on Knowledge and Data Engineering 33, 3 (2021), 921–934. doi:10. 1109/TKDE.2019.2940950 [26]Tonglong Wei, Youfang Lin, Shengnan Guo, Yan Lin, Yiheng Huang, Chenyang Xiang, Yuqing Bai, Menglu Ya, and Huaiyu Wan. 2024. Diff-RNTraj: A Structure- aware Diffusion Model for Road Network-constrained Trajectory Generation. arXiv:2402.07369 [cs.LG] https://arxiv.org/abs/2402.07369 [27] Tong Xia, Yong Li, Yunhan Qi, Jie Feng, Fengli Xu, Funing Sun, Diansheng Guo, and Depeng Jin. 2023. History-enhanced and Uncertainty-aware Trajectory Recovery via Attentive Neural Network. ACM Trans. Knowl. Discov. Data 18, 3, Article 53 (dec 2023), 22 pages. doi:10.1145/3615660 [28] Yuanshao Zhu, Yongchao Ye, Shiyao Zhang, Xiangyu Zhao, and James J. Q. Yu. 2023. Diff Traj: Generating GPS Trajectory with Diffusion Probabilistic Model. arXiv:2304.11582 [cs.LG] [29]Yuanshao Zhu, James Jianqiao Yu, Xiangyu Zhao, Qidong Liu, Yongchao Ye, Wei Chen, Zijian Zhang, Xuetao Wei, and Yuxuan Liang. 2024. ControlTraj: Controllable Trajectory Generation with Topology-Constrained Diffusion Model. arXiv:2404.15380 [cs.LG] https://arxiv.org/abs/2404.15380