Paper deep dive
Symbolic Graph Networks for Robust PDE Discovery from Noisy Sparse Data
Xingyu Chen, Junxiu An, Jun Guo, Yuqian Zhou
Intelligence
Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 94%
Last extracted: 3/26/2026, 1:34:14 AM
Summary
The paper introduces the Symbolic Graph Network (SGN), a framework for discovering partial differential equations (PDEs) from noisy and sparse observational data. SGN replaces traditional local differential operators with graph-based message passing to model spatial interactions, providing a non-local representation that is robust to noise. The learned latent features are then processed by a symbolic regression module to extract interpretable mathematical expressions, demonstrating improved performance on benchmark systems like the wave and Navier-Stokes equations.
Entities (6)
Relation Signals (3)
Symbolic Graph Network → incorporates → Symbolic Regression
confidence 95% · the learned latent features are further processed by a symbolic regression module
Symbolic Graph Network → utilizes → Graph Neural Networks
confidence 95% · SGN synergizes the non-local aggregation capabilities of Graph Neural Networks (GNNs)
Symbolic Graph Network → evaluateson → Wave equation
confidence 90% · We evaluate the proposed method on several benchmark systems, including the wave equation
Cypher Suggestions (2)
Find all methodologies used within the SGN framework · confidence 90% · unvalidated
MATCH (f:Framework {name: 'Symbolic Graph Network'})-[:UTILIZES|INCORPORATES]->(m:Methodology) RETURN m.nameList all physical systems evaluated by the framework · confidence 90% · unvalidated
MATCH (f:Framework {name: 'Symbolic Graph Network'})-[:EVALUATES_ON]->(s:PhysicalSystem) RETURN s.nameAbstract
Abstract:Data-driven discovery of partial differential equations (PDEs) offers a promising paradigm for uncovering governing physical laws from observational data. However, in practical scenarios, measurements are often contaminated by noise and limited by sparse sampling, which poses significant challenges to existing approaches based on numerical differentiation or integral formulations. In this work, we propose a Symbolic Graph Network (SGN) framework for PDE discovery under noisy and sparse conditions. Instead of relying on local differential approximations, SGN leverages graph message passing to model spatial interactions, providing a non-local representation that is less sensitive to high frequency noise. Based on this representation, the learned latent features are further processed by a symbolic regression module to extract interpretable mathematical expressions. We evaluate the proposed method on several benchmark systems, including the wave equation, convection-diffusion equation, and incompressible Navier-Stokes equations. Experimental results show that SGN can recover meaningful governing relations or solution forms under varying noise levels, and demonstrates improved robustness compared to baseline methods in sparse and noisy settings. These results suggest that combining graph-based representations with symbolic regression provides a viable direction for robust data-driven discovery of physical laws from imperfect observations. The code is available at this https URL
Tags
Links
- Source: https://arxiv.org/abs/2603.22380v1
- Canonical: https://arxiv.org/abs/2603.22380v1
Full Text
72,966 characters extracted from source content.
Expand or collapse full text
Symbolic Graph Networks for Robust PDE Discovery from Noisy Sparse Data Xingyu Chen Junxiu An Jun Guo Yuqian Zhou Abstract Data-driven discovery of partial differential equations (PDEs) offers a promising paradigm for uncovering governing physical laws from observational data. However, in practical scenarios, measurements are often contaminated by noise and limited by sparse sampling, which poses significant challenges to existing approaches based on numerical differentiation or integral formulations. In this work, we propose a Symbolic Graph Network (SGN) framework for PDE discovery under noisy and sparse conditions. Instead of relying on local differential approximations, SGN leverages graph message passing to model spatial interactions, providing a non-local representation that is less sensitive to high-frequency noise. Based on this representation, the learned latent features are further processed by a symbolic regression module to extract interpretable mathematical expressions. We evaluate the proposed method on several benchmark systems, including the wave equation, convection–diffusion equation, and incompressible Navier–Stokes equations. Experimental results show that SGN can recover meaningful governing relations or solution forms under varying noise levels, and demonstrates improved robustness compared to baseline methods in sparse and noisy settings. These results suggest that combining graph-based representations with symbolic regression provides a viable direction for robust data-driven discovery of physical laws from imperfect observations. The code is available at https://github.com/CXY0112/SGN keywords: PDE discovery, graph neural networks , symbolic regression, noise robustness, sparse data [inst1]organization=School of Software Engineering, Chengdu University of Information Technology, city=Chengdu, postcode=610225, country=China [inst2]organization=School of Statistics, Chengdu University of Information Technology, city=Chengdu, postcode=610103, country=China [inst3]organization=College of Applied Mathematics, Chengdu University of Information Technology, city=Chengdu, postcode=610225, country=China [inst4]Key Laboratory of Mathematical Meteorology, Chengdu University of Information Technology, city=Chengdu, postcode=610225, country=China 1 Introduction Partial differential equations (PDEs) play a central role in modeling complex physical systems [4, 5]. However, deriving governing equations from first principles can be challenging in many modern applications [29]. With the increasing availability of observational data, data-driven approaches for discovering PDEs have attracted growing attention [16, 1]. In this context, developing methods that can reliably identify interpretable and physically consistent equations from noisy and sparsely sampled data remains an important and challenging problem in computational science. Early sparse regression methods, such as SINDy [12, 19, 4] and PDE-FIND [32], rely on explicit numerical differentiation to construct candidate libraries, rendering them highly vulnerable to observational noise and sparse sampling [43, 39]. To address the limitations of rigid numerical derivatives, strong-form neural architectures like PDE-Net 2.0 [23] leverage learnable convolutional filters to approximate spatial derivatives. While more precise on structured meshes, these grid-dependent convolutions still fundamentally compute local derivatives, which inherently amplify high-frequency artifacts and struggle with non-uniform data [28, 13]. To systematically mitigate this noise amplification, integral-based weak-form methods (e.g., Weak-PDE-LEARN [35]) bypass direct differentiation by integrating PDEs against smooth test functions. However, this robust noise suppression comes at a cost: weak-form methods strictly demand dense data grids to satisfy numerical quadrature assumptions, severely limiting their applicability in data-scarce regimes [17]. Consequently, current discovery frameworks present a critical trade-off between noise robustness and data efficiency, frequently diverging or yielding physically inconsistent models when processing corrupted real-world measurements. Overcoming this methodological gap requires a novel architecture that transcends the rigid dichotomy of fragile local differential operators and data-hungry global integrals, enabling the robust extraction of interpretable physical laws directly from imperfect data. To bridge this methodological gap, we introduce a novel framework, the Symbolic Graph Network (SGN), which eschews grid-dependent local differential operators. SGN synergizes the non-local aggregation capabilities of Graph Neural Networks (GNNs) [7, 40] with the mathematical interpretability of symbolic regression. Crucially, the choice of a graph-based architecture is fundamentally driven by its unique alignment with physical realities. The graph message-passing paradigm structurally mirrors the fundamental laws of physics—where local interactions (edge messages) accumulate to drive global state evolution (node updates). By embedding this physics-aligned inductive bias, SGN transforms from a generic black-box approximator into a structured physical reasoning engine that naturally expresses the discrete spatial operators inherent in PDEs. The main contributions of this work are summarized as follows: 1. We introduce a graph-based formulation for representing spatial operators in data-driven PDE discovery. By leveraging message passing on spatial graphs, the proposed approach provides a mesh-free and non-local alternative to conventional numerical differentiation, which can improve robustness under noisy and sparsely sampled observations. 2. We develop the Symbolic Graph Network (SGN) framework, which incorporates graph neural networks to model local interactions between neighboring states. This design leverages the inductive bias of graph-based representations, where nearby nodes exert stronger mutual influence, to capture spatial dependencies in PDE systems. The learned representations are further combined with symbolic regression to identify interpretable governing relations from data. 3. We conduct systematic experiments on representative PDE systems, including the wave equation, convection–diffusion equation, and incompressible Navier–Stokes equations. The results demonstrate that the proposed method can recover meaningful physical relations and maintain stable performance under varying noise levels, particularly in sparse data regimes. Through these breakthroughs, this work not only demonstrates SGN’s exceptional capability in recovering accurate PDE structures and parameters but also extends the boundaries of PDE discovery to real-world physical scenarios constrained by degraded data quality. Ultimately, it offers a unified solution that combines the interpretability of symbolic methods with the robustness of deep learning. 2 Related Work The data-driven discovery of partial differential equations has evolved rapidly, shifting from black-box predictions to the extraction of interpretable physical laws. We categorize existing methodologies into several primary branches: symbolic regression, strong-form convolutional approaches, and integral-based weak-form frameworks. While these paradigms have achieved notable success, they inherently struggle with severe observational noise and coarse-grained grid dependencies. Analyzing these critical limitations provides the theoretical and practical motivation for our proposed method. 2.1 Symbolic Regression for Equation Discovery Symbolic regression (SR) has emerged as a powerful paradigm for discovering governing physical laws directly from observational data [9, 27, 38]. Unlike deep learning models that act as black-box interpolators, SR algorithms seek explicit, interpretable mathematical expressions. A prominent milestone in this domain is Sparse Identification of Nonlinear Dynamics (SINDy), which transforms equation discovery into a sparse linear regression problem over a predefined library of candidate functions [6]. More recently, genetic algorithm-based frameworks like PySR [10] have been developed to explore a much broader, flexible combinatorial space of equation trees without relying on rigid predefined libraries. However, the fundamental bottleneck of applying SR to spatiotemporal physical systems lies in its extreme sensitivity to the quality of input features, particularly the state derivatives (e.g., ux,uxx,utu_x,u_x,u_t). In real-world scenarios, these derivatives cannot be measured directly and must be estimated from data. When observations are corrupted by noise or sampled on sparse grids, traditional numerical differentiation methods amplify high-frequency errors exponentially[39, 36]. Consequently, SR engines exhibit a critical vulnerability to corrupted inputs, aggressively fitting high-frequency numerical artifacts rather than isolating the true physical operators. 2.2 Strong-Form Approaches: Differential Operator Approximation To address the challenge of derivative estimation, deep learning architectures have been introduced to discover PDEs in their strong form. A highly representative framework is PDE-Net 2.0 [24, 23], which ingeniously bridges convolutional neural networks (CNNs) [22] with numerical analysis. The essence of PDE-Net lies in constraining the learnable convolutional filters to mimic explicit finite difference operators via moment-matching conditions, enabling the simultaneous learning of differential operators and symbolic coefficients. While mathematically elegant on dense and pristine grids, this strong-form convolutional paradigm exhibits fatal defects on coarse-grained meshes with observational noise. Because CNN-based [14] difference operators are strictly local, their truncation errors grow prohibitively large when the spatial resolution is low [3]. Furthermore, attempting to capture high-order derivatives (e.g., Laplace operators) through localized convolutional stencils on noisy, coarse grids invariably leads to catastrophic noise amplification, gradient explosion, and the generation of physically inconsistent terms (such as negative diffusion), severely restricting their applicability in practical sparse-sensing environments [41, 25]. 2.3 Weak-Form Approaches: Integral-Based Noise Suppression To circumvent the catastrophic noise amplification inherent in strong-form differentiation, some recent literature has shifted towards integral-based, or weak-form, discovery frameworks. Notable examples include WSINDy [26] and Weak-PDE-LEARN [35]. The core algorithmic philosophy of these approaches relies on multiplying the governing PDE by a set of compactly supported, smooth test functions and integrating over local spatiotemporal domains. Through integration by parts, spatial and temporal derivatives are elegantly transferred from the noisy, discrete observational data onto the analytically differentiable test functions. For instance, Weak-PDE-LEARN typically employs neural networks to construct a continuous representation of the latent solution fields. It then evaluates these weak-form integrals across numerous local overlapping sub-domains to build a robust feature matrix. Finally, sparse regression is applied to this integral formulation rather than the point-wise differential form [42, 26]. This integration process effectively acts as a natural low-pass filter, granting weak-form approaches exceptional robustness against measurement noise. However, accurately evaluating high-dimensional spatiotemporal integrals via numerical quadrature inherently assumes the availability of dense collocation points. Consequently, these methods are fundamentally ”data-hungry” in the spatial domain [30]. When applied to highly coarse-grained grids, the quadrature approximations degrade rapidly, limiting their applicability in sparse-sensing environments. We provide a theoretical analysis of this vulnerability in Section 5, explicitly proving how SGN’s integral message-passing mechanism naturally compensates for such data sparsity. 3 Problem Formulation We aim to discover the governing physical laws of spatiotemporal systems from sparse and noisy observations. Generally, such systems are described by nonlinear PDEs: ut(x,t)=ℱ(u,ux,uxx,…;ξ),(x,t)∈Ω×[0,T],u_t(x,t)=F (u,u_x,u_x,…;ξ ), (x,t)∈ ×[0,T], (1) where ℱF is the governing nonlinear function parameterized by physical coefficients ξ. In realistic scenarios, the latent solution u(x,t)u(x,t) is sampled on a coarse spatiotemporal grid (,)(X,T) and corrupted by independent Gaussian measurement noise, yielding observations U~i,j=u(xi,tj)+ϵi,j U_i,j=u(x_i,t_j)+ _i,j, with ϵ∼(0,σ2)ε (0,σ^2). Our objective is to identify a parsimonious symbolic expression ℱ F for the system dynamics. We construct a flexible candidate library D integrating explicit spatiotemporal coordinates, state variables, and robust derivative estimates extracted by our SGN model Φθ _θ: =x,t∪U~∪Φθ(U~).D=\x,t\∪\ U\∪ _θ( U). (2) The discovery process is formulated as a sparse regression problem minimizing the discrepancy against a target Y (e.g., utu_t for PDEs, or u for analytical solutions): minθ,ξ‖−ℱ^(;ξ)‖22+λ‖ξ‖0. _θ,ξ \|Y- F(D;ξ) \|_2^2+λ\|ξ\|_0. (3) By design, incorporating explicit coordinates x,t\x,t\ into D enables the framework to potentially recover directly closed-form analytical solutions (e.g., u=f(x,t)u=f(x,t)) rather than being strictly confined to differential operator approximations. 4 Method As illustrated in Figure 1, the proposed Symbolic Graph Network framework is structured into two principal components: the Graph Neural Simulator and the Symbolic Inverse Analyzer. The Graph Neural Simulator constitutes the core of the framework, functioning as a robust mesh-free discretization engine. Rather than relying on rigid local differential approximations, this module leverages the message-passing mechanism of GNNs to natively represent continuous spatial interactions as non-local integral operators. To further ensure the stability of the extracted physical operators under extremely sparse and noisy observations, the simulator is augmented with auxiliary stabilization strategies, including geometric warm-starts and adaptive noise adversarial training. The Symbolic Inverse Analyzer, powered by the High-Performance Symbolic Regression (PySR) toolkit [11], subsequently serves as the discovery engine. It distills the robust derivative features extracted by the neural component into explicit, parsimonious mathematical expressions, thereby recovering the underlying governing equations. Figure 1: The overall framework of SGN. The architecture comprises two synergistic components: (1) the Graph Neural Simulator, which functions as a mesh-free discretization engine. It leverages graph message passing to extract robust non-local spatial operators from sparse and noisy observations, aided by auxiliary stabilization strategies; and (2) the Symbolic Inverse Analyzer, which distills these physics-aligned latent features into explicit, parsimonious governing equations. 4.1 The Graph Neural Simulator The Graph Neural Simulator serves as the dynamical modeling engine of the SGN framework. By leveraging the message-passing mechanism of GNNs, the Simulator functions as a structured reasoning engine that interprets the system’s evolution through a specific relational lens. According to the No Free Lunch theorem, effective learning from finite data requires strong inductive biases. In our framework, we embed a physics-aligned inductive bias by treating the continuous physical dynamics as discrete graph-based interactions. 4.1.1 Graph-Based Mesh-Free Discretization (Core Mechanism) From a computational physics perspective, the aggregation of local neighborhood states intrinsically mirrors the continuous spatial interactions that govern PDEs. While classical grid-based aggregations (e.g., finite differences) rigidly instantiate local differential operators that amplify noise, SGN parameterizes this aggregation via MLPs. This crucial upgrade transforms the rigid local approximations into robust, data-driven non-local integral operators. Consequently, the Simulator constructs a spatial graph =(,ℰ)G=(V,E) and employs message-passing GNNs [18] not merely as a denoising filter, but as a fundamental discrete expression of PDE interactions. This formulation effectively establishes topological dependencies among isolated nodes to take over the derivative refinement task. Specifically, we map the discrete grid points to graph nodes =1,…,NxV=\1,…,N_x\. Edges ℰE are established based on spatial proximity to capture local geometric connectivity. We adopt a standard message-passing paradigm parameterized by Multilayer Perceptrons (MLPs), where the node features hi(l)h_i^(l) at layer l are updated as follows: hi(l+1)=ϕ(l)(hi(l),⨁j∈(i)ψ(l)(hi(l),hj(l))),h_i^(l+1)=φ^(l) (h_i^(l), _j (i)ψ^(l) (h_i^(l),h_j^(l) ) ), (4) where ψ(l)ψ^(l) (the message function) and ϕ(l)φ^(l) (the update function) are small MLPs, and ⨁ denotes a permutation-invariant aggregation operator (e.g., summation or mean). This message-passing mechanism performs two crucial physics-aligned corrections. First, it acts as a noise dampener (pulling back): if node i experiences an artificial noise spike, the elongated edges connected to its stable neighbors j generate a strong restoring force, pulling the aberrant prediction back to the smooth physical surface. Second, it acts as a trend transmitter (pulling up): when a true low-frequency physical wave approaches, the elevation of neighbor j creates a pulling force that preemptively imparts the correct physical gradient to node i, avoiding lag. Beyond acting as a robust, non-local integral filter that naturally averages out high-frequency stochastic noise (theoretically analyzed in Section 5), this message-passing design crucially decouples the complex spatial convolution into two simpler, independent functions. Instead of performing symbolic regression on a monolithic “black-box” GNN, the subsequent Analyzer only needs to recover the mathematical forms of the message function ψ and the update function ϕφ individually. This divide-and-conquer strategy significantly reduces the search space complexity for the symbolic regression module. 4.1.2 Auxiliary Stabilization Strategies To ensure the graph simulator robustly focuses on extracting fundamental non-local physical operators, we incorporate two standard auxiliary techniques to stabilize the optimization process. Figure 2: Savitzky-Golay filtering. The dashed line exhibits the severe high-frequency jitter of raw noisy observations. In contrast, the solid curves demonstrate the smoothed physical trends recovered by increasing window sizes. This filtering effectively suppresses stochastic noise while preserving geometric structure, providing a stable, derivative-friendly initialization for the GNN simulator. Geometric Warm-Start via Savitzky-Golay (S-G) Filtering [33, 34]. Neural networks often struggle with the “cold-start” problem when estimating gradients directly from extremely noisy grids. To provide a stable initialization, we apply a non-parametric S-G polynomial filter to the raw observations. As intuitively illustrated in Figure 2, this operator performs local least-squares polynomial fitting to effectively suppress high-frequency stochastic noise while preserving the underlying geometric structure. This explicit geometric prior generates high-fidelity initial feature estimates, empowering the subsequent GNN to bypass basic noise artifacts and focus exclusively on resolving complex nonlinear physical interactions. To prevent the GNN from overfitting to specific numerical artifacts, we employ a dynamic noise injection strategy [2, 37]. By adaptively perturbing the input node features with synthetic Gaussian noise during training based on the estimated observational noise level, we force the message-passing mechanism to act as an active structural filter. Only those latent operator forms that remain stable under strong, randomized perturbations are retained as authentic physical laws, ensuring the model captures robust physical essence rather than transient data biases. 4.2 The Symbolic Inverse Analyzer The Graph Neural Simulator yields a curated dataset of input-output pairs encoding the underlying dynamics. To distill these implicit neural representations into explicit mathematical expressions, we employ PySR, a genetic algorithm-based symbolic regression engine. PySR searches a vast combinatorial space of binary equation trees, discovering complex functional forms without rigid structural priors. We exploit the structural inductive bias of the GNN, which decomposes complex many-body dynamics into a shared pairwise interaction function, ψθ(hi,hj) _θ(h_i,h_j). The symbolic analyzer thus only needs to model the fundamental interaction between two discrete nodes (sender j and receiver i). This decoupling inherently reduces the regression search space complexity from (Nneighbors)O(N_neighbors) to a constant (2)O(2), rendering the discovery process independent of grid connectivity. Guided by this architectural simplification, we execute a two-stage symbolic discovery protocol. 4.2.1 Stage I: Distilling Interaction Laws (The Message Function) The primary objective of the first stage is to decode the learned message function ψθ _θ, which dictates the localized physical interactions (e.g., spatial fluxes or discrete gradients) between adjacent nodes. Specifically, we extract a dataset of pairwise node states (hi,hj)(h_i,h_j) and their latent messages eije_ij from the trained simulator, and employ PySR to discover a symbolic mapping fmsgf_msg: eij=fmsg(hi,hj)≈ψθ(hi,hj).e_ij=f_msg(h_i,h_j)≈ _θ(h_i,h_j). (5) Motivated by the manifold hypothesis [15], we posit that the essential physical interactions dictating node-to-node influence inherently reside on a low-dimensional manifold. We restrict dmsgd_msg to an extremely low dimensionality (typically 11 or 22), the neural network is compelled to distill complex neighborhood dynamics into a minimal, noise-filtered physical representation. This architectural constraint drastically reduces the symbolic search space, enabling PySR to rapidly converge onto fundamental spatial operators. 4.2.2 Stage I: Distilling Evolution Laws (The Update Function) Once the interaction law fmsgf_msg is identified, the second objective is to uncover the time-evolution rule. We target the Update MLP ϕθ _θ, which takes the center node features hih_i and the aggregated messages ∑eΣ e as input. Since the complex spatial interactions have already been encapsulated into the scalar message e (from Stage I), the dimensionality of this regression task is minimal. We seek a symbolic function fupdf_upd such that: u(t+Δt)≈fupd(u(t),∑j∈(i)fmsg(hi,hj)).u(t+ t)≈ f_upd (u(t), _j (i)f_msg(h_i,h_j) ). (6) This step effectively recovers the temporal term of the PDE. The final interpretable PDE is then constructed by composing the symbolic expressions found in these two stages: PDE=fupd∘Agg∘fmsgPDE=f_upd f_msg. 5 Theoretical Analysis To rigorously justify the empirical robustness of the Graph Neural Simulator against severe observational noise, we provide a comprehensive analysis from the perspective of functional analysis. We demonstrate that the discrete message-passing paradigm is mathematically equivalent to a continuous spatial integral operator, which intrinsically functions as a low-pass filter to suppress high-frequency artifacts. In traditional PDE discovery, numerical differentiation (e.g., Finite Difference) is directly applied to noisy data. Mathematically, the derivative operator acts as a high-pass filter. Given a perturbation δ(x)=ϵsin(ωx)δ(x)=ε (ω x) with frequency ω, the spatial derivative ∂xδ(x)=ϵωcos(ωx) ∂ xδ(x)=εω (ω x) amplifies the noise linearly with ω. As ω→∞ω→∞ (typical for measurement jitter), the error diverges. Beyond noise amplification, local differential operators are fundamentally crippled by truncation errors in sparse regimes. To intuitively grasp this, consider the standard finite difference approximation derived via Taylor expansion: ∂u∂x=u(x+Δx)−u(x)Δx−(Δx2!∂2u∂x2+Δx23!∂3u∂x3+…). ∂ u∂ x= u(x+ x)-u(x) x- ( x2! ∂^2u∂ x^2+ x^23! ∂^3u∂ x^3+… ). (7) The numerical scheme forcefully truncates the infinite series in the parentheses, leaving a residual error bounded by (Δxp)O( x^p), where p is the order of accuracy. In ideal, ultra-dense grids (Δx→0 x→ 0), this discarded tail vanishes. However, in sparse sensing environments where Δx x is large, this truncation error morphs into a massive numerical artifact. The calculated derivative becomes heavily polluted by higher-order ”ghost” terms. When fed into a symbolic regression engine, this triggers severe error propagation, forcing the algorithm to aggressively overfit these numerical artifacts rather than isolating the true physical dynamics. In contrast, our framework abandons this local Taylor-expansion paradigm. By delegating the learning of spatial interactions to the GNN’s aggregation mechanism, SGN evaluates the macroscopic accumulation of information over a neighborhood. We formalize this process as follows. Let =(,ℰ)G=(V,E) be a spatial graph constructed on a uniform grid with spacing Δx x in a domain Ω⊂ℝd ^d. As the grid resolution increases (Δx→0 x→ 0), the discrete message aggregation with a normalized permutation-invariant operator (e.g., Mean) converges to a continuous Fredholm-type integral operator [20]. Consider the discrete aggregation at a target node i located at spatial coordinate x∈Ωx∈ : mi=1|(i)|∑j∈(i)ψθ(u(x),u(xj)).m_i= 1|N(i)| _j (i) _θ (u(x),u(x_j) ). (8) Assuming the neighborhood (i)N(i) is defined by a spatial radius R, the continuous limit of this summation over the local neighborhood BR(x)B_R(x) as Δx→0 x→ 0 can be expressed as: ℐψ[u](x)=1Vol(BR)∫BR(x)ψθ(u(x),u(y))y.I_ψ[u](x)= 1Vol(B_R) _B_R(x) _θ (u(x),u(y) )dy. (9) Here, the learned message function ψθ(⋅,⋅) _θ(·,·) implicitly parameterizes the kernel of the integral operator ℐψI_ψ. The equivalence to an integral operator provides a rigorous explanation for the Simulator’s noise immunity. Suppose the observed field is corrupted by high-frequency noise: u~(y)=utrue(y)+ϵeiωy u(y)=u_true(y)+ε e^iω y. Substituting this into our continuous aggregation operator and applying a first-order Taylor expansion to the neural kernel ψθ _θ with respect to the perturbation, the noise contribution is dominated by integrals of the form ϵ∫BR(x)∂ψθ∂ueiωyy.ε _B_R(x) ∂ _θ∂ ue^iω ydy. (10) Provided that the learned derivative of the kernel is L1L^1 integrable, the Riemann-Lebesgue Lemma [31] strictly dictates that its Fourier transform decays asymptotically as the frequency increases: lim|ω|→∞∫BR(x)ψθ(…)eiωyy=0. _|ω|→∞ _B_R(x) _θ(…)e^iω ydy=0. (11) Equation 11 mathematically guarantees that high-frequency observational noise (ω→∞ω→∞) is inherently annihilated by the integral nature of the graph aggregation process. While the S-G filter provides an explicit geometric prior for cold-start smoothing, the GNN’s message passing acts as a trainable, data-driven low-pass filter. It extracts the pristine low-frequency physical interactions (the latent message e) while discarding the numerical artifacts. More importantly, unlike explicit numerical differentiation or standard quadrature methods that are strictly bottlenecked by the (Δxp)O( x^p) truncation error on sparse grids, the neural integral operator adapts its learned kernel ψθ _θ to the available data distribution. Furthermore, this non-local formulation implicitly relaxes the strict Courant-Friedrichs-Lewy (CFL) [8, 21] stability restrictions that frequently trigger numerical blow-up in explicit differential schemes. This structural advantage explains why our method succeeds in extreme noise and sparse regimes where local differential-based methods experience catastrophic gradient explosion and severe truncation penalties. 6 Experiment 6.1 Experimental Setup To systematically evaluate our framework, we design experiments across three dynamical systems of increasing complexity. We use the Wave Equation to verify foundational PDE recovery, the Convection-Diffusion Equation for complex nonlinear transport, and the Incompressible Navier-Stokes equations to test high-dimensional fluid dynamics. All tasks uniformly employ an additive graph aggregation scheme. Detailed physical formulations are elaborated in subsequent subsections. Datasets are generated via high-fidelity numerical solvers on a 32 × 32 × 100 spatiotemporal grid (representing the x, y, and t dimensions). To simulate measurement inaccuracies, we inject Gaussian white noise following (0,σnoise2)N(0,σ^2_noise), where the noise standard deviation σnoise _noise is scaled by a prescribed ratio η against the clean data. Since SGN and the baseline (PDE-Net 2.0) output mathematically distinct forms (discrete symbolic rules versus continuous PDEs), direct structural comparisons are insufficient. Instead, we establish a unified metric by computing the Mean Squared Error (MSE) between the predicted states of the discovered formulas and the pristine observational data. We benchmark SGN against PDE-Net 2.0 on the two simpler systems under low-noise regimes. Because explicit finite differences cause PDE-Net 2.0 to suffer severe gradient explosions (NaN/Inf) under noise, we exclusively evaluate SGN on the complex Navier-Stokes system and all high-noise stress tests to probe its ultimate robustness limits. 6.2 Wave Equation We utilize the 2D homogeneous wave equation system to benchmark the fundamental differences between SGN and the baseline method (PDE-Net 2.0) when processing coarse-grid observational data. The experimental data is generated from the 2D wave equation ∂2u∂t2=c2Δu ∂^2u∂ t^2=c^2 u, where the wave speed parameter is set to c2=0.5c^2=0.5. Specifically, the observations are sampled from the analytical solution of a fundamental standing wave: u(x,y,t)=sin(πx)sin(πy)cos(πt)u(x,y,t)= (π x) (π y) (π t) (12) For this fundamental standing wave, the Laplace operator Δu u and the field variable u itself exhibit a specific eigenfunction relationship: Δu=−2π2u≈−19.74u. u=-2π^2u≈-19.74u. (13) To facilitate the first-order operator learning required by baseline methods, we introduce an auxiliary velocity variable v=utv=u_t. Consequently, under this specific spatial distribution, the continuous wave equation numerically degenerates toward a localized, spatially decoupled harmonic oscillator system governed by the following true evolution equations: ut u_t =v =v (14) vt v_t =c2Δu=0.5×(−2π2u)=−π2u≈−9.87u =c^2 u=5×(-2π^2u)=-π^2u≈-87u We conduct a comprehensive quantitative and qualitative comparison of the physical formulas extracted by both methods across varying noise levels. The core numerical results are summarized in Table 1. Table 1: Comparison of discovered formulas for the Wave Equation across different noise levels. (Note: Constants in the extracted formulas are rounded to appropriate significant figures, and expressions are moderately simplified for readability. The complete, unedited formulas and detailed variable nomenclature are provided in the Appendix.) Noise Level Method MSE Extracted Formula (Simplified) 0.0% PDE-Net 2.0 5.91×10−55.91× 10^-5 ut=0.998v−0.099uvt=−9.82−0.102v+1.17×10−5(uxx+uyy) aligned u_t&=0.998v-0.099u\\ v_t&=-9.82u-0.102v+1.17× 10^-5(u_x+u_y) aligned SGN (Ours) 2.54×10−42.54× 10^-4 u≈sin(3.08x)sin(3.10y)cos(t/−0.317)u≈ (3.08x) (3.10y) (t/-0.317) 0.1% PDE-Net 2.0 2.92×10−52.92× 10^-5 ut=0.996v−0.092uvt=−9.79−0.114v+7.32×10−6(uxx+uyy) aligned u_t&=0.996v-0.092u\\ v_t&=-9.79u-0.114v+7.32× 10^-6(u_x+u_y) aligned SGN (Ours) 5.41×10−45.41× 10^-4 u≈sin(3.14x)sin(y+…)cos(−3.14t)/sin(1.23)u≈ (3.14x) (y+…) (-3.14t)/ (1.23) 0.2% PDE-Net 2.0 NaN ut=0.507v−0.011uy+0.007uxvt=−9.10u+0.108−0.050(uxxv)−0.018(uyyuxx) aligned u_t&=0.507v-0.011u_y+0.007u_x\\ v_t&=-9.10u+0.108-0.050(u_xv)-0.018(u_yu_x) aligned SGN (Ours) 1.38×10−101.38× 10^-10 u=sin(3.1416)sin(3.1416)cos(3.1416)u= (3.1416x) (3.1416y) (3.1416t) 0.5% SGN (Ours) 2.97×10−62.97× 10^-6 u≈sin(3.14x)sin(3.14y)cos(t/−0.318)u≈ (3.14x) (3.14y) (t/-0.318) 1.0% SGN (Ours) 1.43×10−41.43× 10^-4 u≈uprev+0.062(uprev−0.008)sin(3.12t)cos(3.09t)u≈ u_prev+0.062(u_prev-0.008) (3.12t) (3.09t) 2.0% SGN (Ours) 1.27×10−41.27× 10^-4 u≈uprev−0.063sin(3.14)sin(3.14)sin(3.12)u≈ u_prev-0.063 (3.14x) (3.14y) (3.12t) 5.0% SGN (Ours) 6.71×10−46.71× 10^-4 u≈uprev−0.063sin(3.14)sin(3.15)sin(3.13)u≈ u_prev-0.063 (3.14x) (3.15y) (3.13t) 10.0% SGN (Ours) 7.84×10−27.84× 10^-2 u≈uprev−sin(sin(−3.10)sin(−3.13)⋅0.038)u≈ u_prev- ( (-3.10t) (-3.13y)· 0.038) Figure 3: Visual comparison of predicted wave fields (u(x,y,t)u(x,y,t)) at time t=1.0t=1.0 under different noise regimes. Top row: low noise (0.2%); Bottom row: severe noise (5.0%). Columns present Ground Truth, SGN (ours), and PDE-NET 2.0 (baseline). While all methods converge under low noise, SGN maintains structural stability and captures the physical wave front even at 5.0% noise, whereas PDE-NET 2.0 exhibits catastrophic numerical divergence (indicated by ‘Inf’ on a blank background). As shown in Table 1, under extremely low noise scenarios, both methods achieve relatively low prediction errors, yet the ”physics” they learn are fundamentally different. For PDE-Net 2.0, the diffusion coefficient extracted in its evolution equation (∼10−5 10^-5) is nearly zero. This indicates that, to circumvent the prohibitive second-order finite difference truncation errors on the coarse grid, the model overfits to the intrinsic linear relationship specific to this standing wave mode (the extracted coefficient −9.82-9.82 is highly proximate to −π2-π^2). By decoupling the learned system (ut≈vu_t≈ v, vt≈−9.82uv_t≈-9.82u), we deduce that PDE-Net degenerates the wave equation into a localized ordinary differential equation utt+9.82u=0u_t+9.82u=0. While its temporal solution proportional to cos(3.13t) (3.13t) aligns with the true frequency, PDE-Net fundamentally fails to capture the spatial coupling (sin(πx)sin(πy) (π x) (π y)). In stark contrast, SGN demonstrates the profound advantage of stepping outside the local operator approximation paradigm. Because the node update module explicitly incorporates spatiotemporal coordinates (x,y,t)(x,y,t), the model utilizes symbolic regression to directly reconstruct the closed-form analytical solution. Remarkably, at a 0.2% noise level as show in Figure 3, PDE-Net 2.0 rapidly collapses and produces NaNs due to the noise amplification of explicit differential operators, SGN successfully discovers the exact mathematical structure sin(πx)sin(πy)cos(πt) (π x) (π y) (π t) with an astonishing precision (3.1416≈π3.1416≈π), achieving an ultra-low MSE of 1.38×10−101.38× 10^-10. This architectural advantage of SGN is further magnified under high-noise conditions. By directly fitting the analytical form, SGN completely bypasses numerical spatial differentiation and time-stepping iterations, fundamentally immunizing itself against the accumulation of grid truncation errors. Even under a severe 10.0% noise corruption, although a residual term uprevu_prev is introduced to compensate for the disturbance, SGN steadfastly locks onto the core periodic skeleton of the physical system (accurately identifying core terms like sin(−3.10⋅t) (-3.10· t) and sin(−3.13⋅y) (-3.13· y)). This validates that the synergy of graph message passing and symbolic sparsity penalties empowers SGN to pierce through severe observational noise and directly converge upon the underlying physical truth. 6.3 Convection-Diffusion Equation We evaluate the models on the 2D Convection-Diffusion equation, presenting a significantly heightened challenge compared to the fundamental wave equation. While this system still possesses an analytical solution, its mathematical formulation is exceptionally complex. Because it simulates an unbounded transport process truncated to a finite observational domain, this configuration strictly tests the models’ capacity to capture generalized, local spatiotemporal transport phenomena without overfitting to specific global boundary constraints. The experimental spatiotemporal data simulates a moving Gaussian wave packet governed by the 2D convection-diffusion equation: ∂u∂t+cx∂u∂x+cy∂u∂y=ν(∂2u∂x2+∂2u∂y2) ∂ u∂ t+c_x ∂ u∂ x+c_y ∂ u∂ y=ν ( ∂^2u∂ x^2+ ∂^2u∂ y^2 ) (15) Based on the fundamental solution of the heat kernel, the exact analytical solution for this unforced propagating wave packet is given by: u(x,y,t)=exp(−(x−xc(t))2+(y−yc(t))22σ2(t))u(x,y,t)= (- (x-x_c(t))^2+(y-y_c(t))^22σ^2(t) ) (16) where the packet center drifts linearly due to convection (xc(t)=x0+cxt,yc(t)=y0+cyt)(x_c(t)=x_0+c_xt,\ y_c(t)=y_0+c_yt), and the spatial variance expands over time due to diffusion (σ2(t)=σ02+2νt)(σ^2(t)= _0^2+2ν t). In our high-fidelity dataset, the physical parameters are explicitly configured to represent an advection-dominated transport process: convection velocities are set to cx=0.18c_x=0.18 and cy=0.09c_y=0.09, the diffusion coefficient is heavily suppressed to ν=0.002ν=0.002, and the initial packet variance is σ0=0.1 _0=0.1. Consequently, the theoretical ground-truth evolution equation that operator learning baselines should ideally extract is: ut=−0.18ux−0.09uy+0.002(uxx+uyy)u_t=-0.18u_x-0.09u_y+0.002(u_x+u_y) (17) The quantitative MSE comparisons and the corresponding extracted formulas are presented in Table 2, Figure 4 shows some of the visualization results. Note that while PDE-Net extracts continuous spatial derivatives representing the governing PDE (utu_t), SGN’s symbolic engine outputs explicit discrete-time node update rules (u≈uprev+Δt⋅ℱu≈ u_prev+ t·F). Table 2: Comparison of discovered formulas for the Convection-Diffusion Equation across different noise levels. (Note: Constants in SGN’s extracted node formulas are moderately simplified for readability, with the global multiplier representing the learned time step Δt t. The complete, unedited formulas and detailed variable nomenclature are provided in the Appendix.) Noise Level Method MSE (on True) Extracted Evolution / Node Formula (Simplified) 0.0% PDE-Net 2.0 5.38×10−25.38× 10^-2 ut≈0.96−0.71ux−1.42uy+0.016uxx+0.024uyyu_t 0.96u-0.71u_x-1.42u_y+0.016u_x+0.024u_y SGN (Ours) 5.56×10−35.56× 10^-3 u≈uprev−0.004[ux−uprev+0.50uy+…]u≈ u_prev-0.004 [u_x-u_prev+0.50u_y+… ] 0.1% PDE-Net 2.0 5.36×10−25.36× 10^-2 ut≈1.35−0.71ux−1.42uy+0.013uxx+0.020uyy+0.010uxyu_t 1.35u-0.71u_x-1.42u_y+0.013u_x+0.020u_y+0.010u_xy SGN (Ours) 1.68×10−31.68× 10^-3 u≈uprev−0.003[ux+0.61uy−(ux−uprev)2+…]u≈ u_prev-0.003 [u_x+0.61u_y-(u_x-u_prev)^2+… ] 0.5% PDE-Net 2.0 4.03×10−24.03× 10^-2 ut≈0.83u−0.71ux−1.42uy+0.037()+0.014uxx+0.032uyyu_t≈ 0.83u-0.71u_x-1.42u_y+0.037(u_x)^2+0.014u_x+0.032u_y SGN (Ours) 1.43×10−31.43× 10^-3 u≈uprev−0.004[ux+0.51uy−uprev]u≈ u_prev-0.004 [u_x+0.51u_y-u_prev ] 1.0% PDE-Net 2.0 3.84×1043.84× 10^4 ut≈2.00u−0.73ux−1.41uy−0.031−0.061−0.038uuyu_t≈ 2.00u-0.73u_x-1.41u_y-0.031u_x-0.061u_y-0.038u_y SGN (Ours) 3.06×10−33.06× 10^-3 u≈uprev−0.004[ux−uprev+uy/(…)+…]u≈ u_prev-0.004 [u_x-u_prev+u_y/(…)+… ] 2.0% SGN (Ours) 2.31×10−32.31× 10^-3 u≈uprev+0.002[(uy+uxx)×⋯−(ux+2uy)+…]u≈ u_prev+0.002 [(u_y+u_x)×…-(u_x+2u_y)+… ] 5.0% SGN (Ours) 1.52×10−31.52× 10^-3 u≈uprev−0.004[ux+uy⋅y+cos(ux)−2e1+…]u≈ u_prev-0.004 [u_x+u_y· y+ (u_x)-2e_1+… ] 10.0% SGN (Ours) 1.42×10−21.42× 10^-2 u≈uprev−0.004[ux+sin(uprev/0.02)/(ux2+0.21)−0.30]u≈ u_prev-0.004 [u_x+ (u_prev/0.02)/(u_x^2+0.21)-0.30 ] Comparing the baseline outputs in Table 2 against the theoretical ground truth (Eq. 17) reveals significant physical distortions. Even in noise-free scenarios, PDE-Net hallucinates an artificial linear source term (e.g., 0.96u0.96u), attempting to incorrectly compensate for the non-periodic boundary truncations of the moving Gaussian wave. More critically, as noise escalates to 1.0%, the explicit finite difference schemes severely corrupt the extraction of second-order derivatives, leading PDE-Net to discover negative diffusion coefficients (−0.031uxx−0.061uyy-0.031u_x-0.061u_y). Physically, negative diffusion violates the second law of thermodynamics, representing an anti-dissipative process. Numerically, this ill-posed formulation destabilizes the forward integration scheme, leading to immediate exponential error amplification (MSE skyrocketing to 3.84×1043.84× 10^4) and total computational overflow. Figure 4: Visual comparison of predicted solutions for the Convection-Diffusion Equation with 0.5% noise. Spatial distributions are shown at two time steps: t=0.2t=0.2 (top row) and t=2.0t=2.0 (bottom row). Columns display Ground Truth, SGN, and PDE-NET 2.0. The figure illustrates the relative degradation of prediction coherent structures at longer simulation times for both learned models compared to Ground Truth. Conversely, SGN completely avoids the ill-posed inverse PDE problem. By framing the discovery as an explicit state-transition mapping through graph message passing, SGN learns highly robust discrete-time update rules. As shown in the extracted formulas, SGN consistently isolates a global multiplier acting as the numerical time step (Δt≈±0.004 t≈± 0.004), coupled with a stable spatial gradient combination dominated by local transport features. This structural stability enables SGN to consistently achieve MSEs in the 10−310^-3 range up to 5.0% noise, and maintain a highly credible error of 1.42×10−21.42× 10^-2 even under extreme 10.0% observational noise. The results confirm that discovering stable temporal update schemes via symbolic graph networks is vastly superior to explicit continuous operator approximation when resolving complex transport phenomena. 6.4 Incompressible Navier-Stokes Equations As the ultimate challenge for our framework, we evaluate SGN on the 2D Incompressible Navier-Stokes (N-S) equations. This chaotic fluid dynamics system is notoriously difficult for data-driven operator learning due to the presence of implicit, non-local constraints. The experimental data is generated based on the classic Taylor-Green Vortex decay problem, a renowned exact solution to the N-S equations characterized by a persistent spatial vortex structure whose kinetic energy decays exponentially over time due to viscous dissipation. The analytical ground truth for the velocity field (u,v)(u,v) is given by: u(x,y,t) u(x,y,t) =e−2νtcos(x)sin(y) =e^-2ν t (x) (y) (18) v(x,y,t) v(x,y,t) =−e−2νtsin(x)cos(y) =-e^-2ν t (x) (y) In our dataset, the kinematic viscosity is set to ν=0.05ν=0.05, and the temporal sampling interval on the 32×3232× 32 spatial grid is Δt=0.1 t=0.1. Consequently, the theoretical temporal decay multiplier for a single time step is λ=e−2νΔt=e−0.01≈0.9900498.λ=e^-2ν t=e^-0.01 0.9900498. (19) The momentum evolution in the incompressible N-S equations is governed by ∂t=−(⋅∇)+ν∇2−∇p ∂ t=-(u·∇)u+ν∇^2u-∇ p. Crucially, the pressure gradient term ∇p∇ p acts as a global, non-local projection to enforce the divergence-free constraint (∇⋅=0∇·u=0). Traditional baselines like PDE-Net 2.0 strictly rely on local convolution kernels (e.g., 3×33× 3 stencils) to approximate spatial derivatives. Mathematically, local spatial derivatives cannot explicitly represent the instantaneous, global physical effects of the Poisson pressure field. Attempting to force-fit this non-local phenomenon into local differential operators causes PDE-Net 2.0 to experience severe parameter oscillation and catastrophic divergence. Therefore, quantitative comparisons with PDE-Net 2.0 are omitted in this section. Table 3: Discovery results for the Navier-Stokes (Taylor-Green Vortex) system across different noise levels. The intermediate graph message functions are represented by e0e_0 and e1e_1. Constants are moderately simplified for readability. The complete, unedited formulas and detailed variable nomenclature are provided in the Appendix.) Noise MSE (on True) Node Formula (u,vu,v update rules) Message Formula (e0,e1e_0,e_1 spatial coupling) 0.0% 9.69×10−49.69× 10^-4 u≈cos()sin()⋅[−0.61sin(−0.13t)−0.98]u≈ (x) (y)· [-0.61 (-0.13t)-0.98 ] [Decoupled: Messages ignored by node formula] v≈−cos()⋅(ux+uy−sin(x))⋅[0.94cos(…)]v≈- (y)·(u_x+u_y- (x))· [0.94 (…) ] 0.1% 6.27×10−46.27× 10^-4 u≈u1⋅(0.9900558−0.0016⋅e0)u≈ u_1· (0.9900558-0.0016· e_0 ) e0≈0.015(v1−u2)e_0≈ 0.015(v_1-u_2) v≈v1⋅(0.9902582−0.0007⋅e1)v≈ v_1· (0.9902582-0.0007· e_1 ) e1≈0.159−0.026(t−u2)e_1≈ 0.159-0.026(t-u_2) 0.5% 6.30×10−46.30× 10^-4 u≈0.9900131⋅u1−0.0046⋅e1u 0.9900131· u_1-0.0046· e_1 e0≈0.237−0.212cos(u1−v1−uy1)e_0≈ 0.237-0.212 (u_1-v_1-u_y1) v≈0.9900553⋅v1+0.0030⋅e1v 0.9900553· v_1+0.0030· e_1 e1≈−0.005(uy1+v1)e_1≈-0.005(u_y1+v_1) 1.0% 6.53×10−46.53× 10^-4 u≈0.9897041⋅u1−0.0110⋅e0u 0.9897041· u_1-0.0110· e_0 e0≈−0.079(−)e_0≈-0.079(u_1-u_2) (Discrete Spatial Gradient) v≈0.9898316⋅v1−0.0050⋅e1v 0.9898316· v_1-0.0050· e_1 e1≈−0.007(v2/t)−0.006e_1≈-0.007(v_2/t)-0.006 2.0% 2.13×10−32.13× 10^-3 u≈0.9885945⋅u1+0.0076⋅e0u 0.9885945· u_1+0.0076· e_0 e0≈−0.309(u1−u2+v1−v2)−0.012e_0≈-0.309(u_1-u_2+v_1-v_2)-0.012 v≈0.9888680⋅v1−0.0038⋅e1v 0.9888680· v_1-0.0038· e_1 e1≈−0.195u1+0.224u2+0.391v1−0.429v2e_1≈-0.195u_1+0.224u_2+0.391v_1-0.429v_2 5.0% 6.87×10−36.87× 10^-3 u≈0.9843308⋅u1−0.0190⋅e0+e1u 0.9843308· u_1-0.0190· e_0+e_1 e0≈0.518sin(u1−u2+v1)−v2−0.014e_0≈ 0.518 (u_1-u_2+v_1)-v_2-0.014 v≈0.9815741⋅v1−0.0184⋅e0−…v 0.9815741· v_1-0.0184· e_0-… e1≈0.040(uy1−0.348)−0.517(…)e_1≈ 0.040(u_y1-0.348)-0.517(…) 10.0% 3.93×10−23.93× 10^-2 u≈0.9576391⋅u1+0.0856⋅e0u 0.9576391· u_1+0.0856· e_0 e0≈−0.278u1+0.341u2−0.256v1+0.278v2e_0≈-0.278u_1+0.341u_2-0.256v_1+0.278v_2 v≈v1−0.010[0.375x+(t+1.87)(v1−e0)+…]v≈ v_1-0.010 [0.375x+(t+1.87)(v_1-e_0)+… ] e1≈0.378(v1−u1)+0.423(u2−v2)+0.012e_1≈ 0.378(v_1-u_1)+0.423(u_2-v_2)+0.012 As shown in Table 3 and Figure 5, freed from the restrictive local operator approximation paradigm, SGN circumvents the pressure trap entirely. SGN demonstrates two distinct, highly advanced modes of physical discovery depending on the observational noise level. Under noise-free conditions, SGN leverages its broad symbolic search space to bypass numerical iteration entirely. As shown in Table 3, SGN successfully factorizes the spatiotemporal data, explicitly extracting the spatial coupling terms cos(x)sin(y) (x) (y) and cos(y)sin(x) (y) (x). This structural configuration perfectly aligns with the theoretical analytical solution F(t)⋅f(x,y)F(t)· f(x,y), demonstrating SGN’s profound capability to deduce global analytical forms directly from grid data without invoking the complex local PDE integration involving the implicit pressure term. Figure 5: Snapshots of the Navier-Stokes flow dynamics under a severe 10% noise level. The figure illustrates the Ground Truth alongside the SGN predictions. Even when the underlying data is heavily obscured by high-frequency artifacts, the proposed SGN architecture flawlessly maintains the structural stability of the fluid flow, achieving highly consistent physical predictions without relying on dense spatial sampling. When exposed to 0.1%0.1\% noise, attempting to fit a perfect analytical form becomes statistically prohibitive. Remarkably, SGN adaptively switches its strategy to learn a numerical state-transition update rule. Because the spatial topology (cos(x)sin(y) (x) (y)) of the Taylor-Green vortex remains invariant and only its amplitude decays, SGN’s symbolic engine discards the complex spatial operators and isolates a pure temporal decay multiplier. For the u velocity field at 0.1% noise, SGN discovers the update rule ut+1≈0.9900558⋅utu_t+1 0.9900558· u_t. When compared to the theoretical physical decay multiplier λ≈0.9900498λ 0.9900498 (derived from the viscosity ν=0.05ν=0.05), the relative error of the learned physical coefficient is <0.0006%<0.0006\%. This extraordinarily precise extraction of the fluid energy dissipation rate is consistently maintained across higher noise levels (e.g., yielding multipliers of 0.990010.99001 at 0.5% noise, and 0.989700.98970 at 1.0% noise). Beyond precise quantitative parameter estimation, SGN exhibits profound qualitative robustness under extreme conditions; as visualized in Figure 5, the model flawlessly maintains the coherent structural dynamics of the flow field even when subjected to a severe 10% noise regime. These results compellingly validate that SGN does not merely curve-fit local numerical gradients; rather, it possesses the architectural capacity to pierce through observational noise and directly capture the fundamental conservation and dissipation laws governing complex dynamical systems. 7 Conclusion In this work, we proposed the SGN framework for data-driven discovery of partial differential equations from noisy and sparsely sampled observations. By leveraging the inductive bias of graph neural networks, the proposed approach models local spatial interactions through message passing, providing a mesh-free representation that alleviates the reliance on numerical differentiation. Combined with symbolic regression, the framework enables the extraction of interpretable governing relations from data. The proposed method was evaluated on several representative PDE systems, including the wave equation, convection–diffusion equation, and incompressible Navier–Stokes equations. The results indicate that SGN can recover meaningful physical relations and maintain stable performance under varying noise levels, particularly in sparse data regimes. Despite its effectiveness in noise suppression and equation discovery, the current SGN framework exhibits several limitations. First, it faces a trade-off between the representational capacity of the graph model and the complexity of the symbolic regression stage. To control the downstream search space, the dimension of the latent message representation is restricted to a relatively low value. While this facilitates convergence, it may limit the ability to capture subtle but important coupled physical effects in more complex systems. Second, although the non-local aggregation mechanism of GNNs can effectively suppress high-frequency noise, it may also introduce over-smoothing. In scenarios involving sharp localized transitions, such as discontinuities or shock waves in fluid dynamics, this effect can degrade gradient fidelity and blur physical interfaces. To address these challenges, future work will focus on improving the balance between expressiveness and robustness in the SGN framework. In particular, we plan to investigate adaptive dimensionality reduction and feature selection strategies to better preserve essential physical information while maintaining tractable symbolic regression. In addition, we will explore the incorporation of physics-informed constraints and regularization techniques designed for discontinuous phenomena. By embedding conservation laws and related priors into the learning process, we aim to improve the model’s ability to resolve sharp physical structures while retaining robustness to noise in complex dynamical systems. Funding information The authors extend their appreciation to the supported by Chongqing science and technology bureau (No. CSTB2024TIAD-CYKJCXX0009), the Foundation of the Sichuan Provincial Department of Science and Technology (No. 2026NSFSC0138), the Key Laboratory of Numerical Simulation of Sichuan Provincial Universities (No. KLNS-2023SZFZ002 and KLNS-2023SZFZ005), and the Key Laboratory of Mathematical Meteorology (No. 2025Z0340). Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper. References [1] J. Berg and K. Nyström (2019-05) Data-driven discovery of pdes in complex datasets. Journal of Computational Physics 384, p. 239–252 (en). Cited by: §1. [2] C. M. Bishop (1995) Training with noise is equivalent to tikhonov regularization. Neural computation 7 (1), p. 108–116. Cited by: §4.1.2. [3] S. L. Brunton and J. N. Kutz (2024) Promising directions of machine learning for partial differential equations. Nature Computational Science 4 (7), p. 483–494. Cited by: §2.2. [4] S. L. Brunton, J. L. Proctor, and J. N. Kutz (2016) Discovering governing equations from data by sparse identification of nonlinear dynamical systems. Proceedings of the national academy of sciences 113 (15), p. 3932–3937. Cited by: §1, §1. [5] S. L. Brunton and J. N. Kutz (2024-06) Promising directions of machine learning for partial differential equations. Nature Computational Science 4 (7), p. 483–494 (en). Cited by: §1. [6] W. Cao and W. Zhang (2023-11) Machine learning of partial differential equations from noise data. Theoretical and Applied Mechanics Letters 13 (6), p. 100480 (en). Cited by: §2.1. [7] G. Corso, H. Stark, S. Jegelka, T. Jaakkola, and R. Barzilay (2024) Graph neural networks. Nature Reviews Methods Primers 4 (1), p. 17. Cited by: §1. [8] R. Courant, K. Friedrichs, and H. Lewy (1928) Über die partiellen differenzengleichungen der mathematischen physik. Mathematische annalen 100 (1), p. 32–74. Cited by: §5. [9] M. Cranmer, A. Sanchez Gonzalez, P. Battaglia, R. Xu, K. Cranmer, D. Spergel, and S. Ho (2020) Discovering symbolic models from deep learning with inductive biases. Advances in neural information processing systems 33, p. 17429–17442. Cited by: §2.1. [10] M. Cranmer (2023) Interpretable machine learning for science with pysr and symbolicregression.jl. External Links: 2305.01582 Cited by: §2.1. [11] M. Cranmer (2024) PySR: high-performance symbolic regression in python and julia. Astrophysics Source Code Library, p. ascl–2409. Cited by: §4. [12] B. de Silva, K. Champion, M. Quade, J. Loiseau, J. Kutz, and S. Brunton (2020) PySINDy: a python package for the sparse identification of nonlinear dynamical systems from data. Journal of Open Source Software 5 (49), p. 2104. Cited by: §1. [13] N. Doumèche, F. Bach, G. Biau, and C. Boyer (2024) Physics-informed kernel learning. External Links: 2409.13786 Cited by: §1. [14] A. A. Elngar, M. Arafa, A. Fathy, B. Moustafa, O. Mahmoud, M. Shaban, and N. Fawzy (2021) Image classification based on cnn: a survey. Journal of Cybersecurity and Information Management 6 (1), p. 18–50. Cited by: §2.2. [15] C. Fefferman, S. Mitter, and H. Narayanan (2016) Testing the manifold hypothesis. Journal of the American Mathematical Society 29 (4), p. 983–1049. Cited by: §4.2.1. [16] D. Floryan and M. D. Graham (2022) Data-driven discovery of intrinsic dynamics. Nature Machine Intelligence 4 (12), p. 1113–1120. Cited by: §1. [17] M. L. Gao, J. N. Kutz, and B. Font (2025) Mesh-free sparse identification of nonlinear dynamics. External Links: 2505.16058 Cited by: §1. [18] J. Gilmer, S. S. Schoenholz, P. F. Riley, O. Vinyals, and G. E. Dahl (2020) Message passing neural networks. In Machine learning meets quantum physics, p. 199–214. Cited by: §4.1.1. [19] A. A. Kaptanoglu, B. M. de Silva, U. Fasel, K. Kaheman, A. J. Goldschmidt, J. Callaham, C. B. Delahunt, Z. G. Nicolaou, K. Champion, J. Loiseau, J. N. Kutz, and S. L. Brunton (2022) PySINDy: a comprehensive python package for robust sparse system identification. Journal of Open Source Software 7 (69), p. 3994. Cited by: §1. [20] R. Kress, V. Maz’ya, and V. Kozlov (1989) Linear integral equations. Vol. 82, Springer. Cited by: §5. [21] R. J. LeVeque (2002) Finite volume methods for hyperbolic problems. Vol. 31, Cambridge university press. Cited by: §5. [22] Z. Li, F. Liu, W. Yang, S. Peng, and J. Zhou (2021) A survey of convolutional neural networks: analysis, applications, and prospects. IEEE transactions on neural networks and learning systems 33 (12), p. 6999–7019. Cited by: §2.2. [23] Z. Long, Y. Lu, and B. Dong (2019-12) PDE-net 2.0: learning pdes from data with a numeric-symbolic hybrid deep network. Journal of Computational Physics 399, p. 108925. Cited by: §1, §2.2. [24] Z. Long, Y. Lu, X. Ma, and B. Dong (2018) Pde-net: learning pdes from data. In International conference on machine learning, p. 3208–3216. Cited by: §2.2. [25] C. Meng, S. Griesemer, D. Cao, S. Seo, and Y. Liu (2025) When physics meets machine learning: a survey of physics-informed machine learning. Machine Learning for Computational Science and Engineering 1 (1), p. 20. Cited by: §2.2. [26] D. A. Messenger, J. W. Burby, and D. M. Bortz (2024) Coarse-graining hamiltonian systems using wsindy. Scientific Reports 14 (1), p. 14457. Cited by: §2.3, §2.3. [27] Z. Mo, Y. Fu, and X. Di (2024) PI-neugode: physics-informed graph neural ordinary differential equations for spatiotemporal trajectory prediction. In Proceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems, p. 1418–1426. Cited by: §2.1. [28] A. Pal, S. Bhowmick, and S. Nagarajaiah (2025) Physics-informed ai and ml-based sparse system identification algorithm for discovery of pde’s representing nonlinear dynamic systems. Mechanical Systems and Signal Processing 238, p. 113238. External Links: ISSN 0888-3270 Cited by: §1. [29] K. Raviprakash, B. Huang, and V. Prasad (2022) A hybrid modelling approach to model process dynamics by the discovery of a system of partial differential equations. Computers & Chemical Engineering 164, p. 107862. Cited by: §1. [30] P. A. Reinbold, D. R. Gurevich, and R. O. Grigoriev (2020) Using noisy or incomplete data to discover models of spatiotemporal dynamics. Physical Review E 101 (1), p. 010203. Cited by: §2.3. [31] W. Rudin (1987) Real and complex analysis. 3rd edition, McGraw-Hill. Cited by: §5. [32] S. H. Rudy, S. L. Brunton, J. L. Proctor, and J. N. Kutz (2017) Data-driven discovery of partial differential equations. Science Advances 3 (4), p. e1602614. Cited by: §1. [33] A. Savitzky and M. J. Golay (1964) Smoothing and differentiation of data by simplified least squares procedures. Analytical chemistry 36 (8), p. 1627–1639. Cited by: §4.1.2. [34] R. W. Schafer (2011) What is a savitzky-golay filter?. IEEE Signal processing magazine 28 (4), p. 111–117. Cited by: §4.1.2. [35] R. Stephany and C. Earls (2024) Weak-pde-learn: a weak form based approach to discovering pdes from noisy, limited data. Journal of Computational Physics 506, p. 112950. External Links: ISSN 0021-9991 Cited by: §1, §2.3. [36] P. Thanasutives, T. Morita, M. Numao, and K. Fukui (2023) Adaptive uncertainty-guided model selection for data-driven pde discovery. External Links: 2308.10283 Cited by: §2.1. [37] C. Wang, S. Li, D. He, and L. Wang (2022) Is lˆ2 physics informed loss always suitable for training physics informed neural network?. Advances in Neural Information Processing Systems 35, p. 8278–8290. Cited by: §4.1.2. [38] Y. Wang, N. Wagner, and J. M. Rondinelli (2019) Symbolic regression in materials science. MRS communications 9 (3), p. 793–805. Cited by: §2.1. [39] J. Wentz and A. Doostan (2023-08) Derivative-based sindy (dsindy): addressing the challenge of discovering governing equations from noisy data. Computer Methods in Applied Mechanics and Engineering 413, p. 116096 (en). Cited by: §1, §2.1. [40] Z. Wu, S. Pan, F. Chen, G. Long, C. Zhang, and P. S. Yu (2020) A comprehensive survey on graph neural networks. IEEE transactions on neural networks and learning systems 32 (1), p. 4–24. Cited by: §1. [41] C. Xu, B. T. Cao, Y. Yuan, and G. Meschke (2023) Transfer learning based physics-informed neural networks for solving inverse problems in engineering structures under different loading scenarios. Computer Methods in Applied Mechanics and Engineering 405, p. 115852. Cited by: §2.2. [42] H. Xu, Y. Chen, R. Cao, T. Tang, M. Du, J. Li, A. H. Callaghan, and D. Zhang (2025) Generative discovery of partial differential equations by learning from math handbooks. Nature Communications 16 (1), p. 10255. Cited by: §2.3. [43] H. Zheng and G. Lin (2024) LES-sindy: laplace-enhanced sparse identification of nonlinear dynamical systems. External Links: 2411.01719 Cited by: §1. Appendix A Complete experimental data results This appendix presents the complete, unsimplified mathematical expressions discovered by SGN across various physical systems and noise regimes. For the evaluation metrics in the subsequent tables, MSE(Noise) denotes the mean squared error evaluated on the corrupted data, whereas MSE(Original) measures the structural generalization error on the pristine ground truth. To facilitate interpretation of the raw expressions, which inherently adopt the programmatic naming conventions of our GNN architecture, Table 1 details the physical and architectural meanings of all symbolic variables. Table 1: Nomenclature and physical interpretations of the symbolic variables discovered by SGN. Variable Physical and Architectural Description t The temporal evolution step t of the dynamical system. x, y The spatial coordinates of the grid nodes. u_prev For scalar fields (Wave, Convection-Diffusion), the field value u of the central node at the previous time step. x_prev, y_prev For vector fields (Navier-Stokes), the two state components of the central node at the previous time step, corresponding to horizontal (u) and vertical (v) velocities. *_prev1 In message functions, the historical state variable of the neighbor node (source node, j). *_prev2 In message functions, the historical state variable of the central node (target node, i). ux, uy The first-order spatial derivatives (gradients) of the field variable at the central node with respect to x and y. uxx, uyy The second-order spatial derivatives (diffusion terms) of the field variable at the central node with respect to x and y. ux1, uy1, … The corresponding first/second-order spatial derivatives evaluated at the neighbor node. ux2, uy2, … The corresponding first/second-order spatial derivatives evaluated at the central node. e0, e1 The components of the aggregated low-dimensional message vector received by the central node, serving as inputs to the node update equation. Table 2: Unedited symbolic expressions discovered for the Wave Equation across varying noise levels. Noise MSE(Noise) MSE(Original) Type Equation 0.0% / 2.5431E-04 Msg sin(sin(sin(sin(cos(((t - t) + -0.56922716) + sin(((-1.8829098 - (t + ux2)) * -0.119090885) - u_prev1))) + -1.2227476))) + -0.0041820854 Node (-0.0049376567 + (sin(y * 3.0965688) * sin(x * 3.084605))) * cos(t / -0.31685314) 0.1% 5.4111E-04 5.4102E-04 Msg u_prev2*(2.405846 - u_prev1)*(u_prev1*(-0.14979462) + ux2*(-0.017480928) - (-1)*0.02454977/(t - 1*0.5184039)) Node sin(sin(x*3.1401417)*cos(t*(-3.1437826)))*sin(y + y + y/0.8408427)/sin(1.2261999) 0.2% 4.7517E-07 1.3837E-10 Msg ((cos(cos(sin(u_prev1 + sin(u_prev1 + u_prev2)) + ((uy1 + (ux1 - t)) * 0.20625228))) * cos(t)) - (u_prev2 - -0.08223844)) * ((u_prev1 - 1.3154284) * -0.19282961) Node sin(x*3.1416032)*sin(y*3.1416051)*cos(t*3.1416218) 0.5% 2.9693E-06 1.4776E-10 Msg (-u_prev2 + sin(t*(u_prev2*ux2 - 1*(-3.231915)))*0.06917196)*0.45004588 Node sin(x*3.1415567)*sin(y*3.1416364)*cos(t/(-0.31830963)) 1.0% 1.4262E-04 1.2472E-04 Msg cos((uy1*0.025896823 + 0.096348494)*cos(t - (-0.31524143)*ux1) - (sin(u_prev2)*0.66344243 + 0.48486096)) - 0.9074855 Node u_prev - (-0.06166134)*(0.008264106 - u_prev)*sin(t*3.123437)/cos(t*3.086291) 2.0% 1.2652E-04 3.8287E-05 Msg t*(-0.03245001) - u_prev2/(-1.6076635) + (u_prev1 - 1*0.46895114)*(u_prev1*0.23417574 + (-ux1 + uy1)*(-0.026761778) - 0.22886173) Node u_prev + sin(x*3.1438239)*(-0.06252605)*sin(t*3.1181076)* sin(y*3.1389296) 5.0% 6.7079E-04 6.6427E-05 Msg sin(-0.4441846*u_prev2 + (2.4497662*u_prev2*u_prev2 + sin(t/0.32815966 - 1*0.65609646))*0.022165552) Node ((((sin(t / 0.31949028) * -0.09182406) * sin(y / 0.3178091)) * 0.68839014) * sin(x / 0.3180929)) + (u_prev / 0.99954045) 10.0% 7.8418E-02 7.6029E-02 Msg (u_prev1 + (-sin(u_prev1) - 2.8246412)*sin(u_prev2))/(-6.849007) Node u_prev - sin((sin(t * -3.1070971) * sin(y * -3.1335773)) * 0.038306106) Table 3: Unedited symbolic expressions discovered for the Convection-Diffusion Equation across varying noise levels. Noise MSE(Noise) MSE(Original) Type Equation 0.0% / 5.5649E-03 Msg1 (0.26722786 - 0.37149414*u_prev2)*sin(u_prev2*(ux1 + 2.8515506)) Msg2 ux2*(sin(ux2)*0.04038851 + 0.02046432) Node u_prev + ((ux - ((u_prev + (((uy * -0.08273945) - 0.50470436) * uy)) - (sin(u_prev) * ((u_prev * t) + -1.7263777)))) * -0.0040644426) 0.1% 1.6789E-03 1.6792E-03 Msg1 (cos(uy2) - cos(uyy1*(-0.0684862)))*(cos(uyy2*(-0.075827084)) + 0.92147017)*(-0.10752371) Msg2 u_prev1*sin((u_prev1 - 1*0.82263863)*sin(ux1 + 0.73160326) + sin(t*u_prev2) + cos(u_prev1)/0.27612534) Node u_prev + ((-u_prev + ux*0.25640112)**2 - (ux - (-0.6112448)*uy))*0.0033418469 0.5% 1.4336E-03 1.4342E-03 Msg1 (-u_prev1/1.3110251 + u_prev2)*(t - cos(u_prev1/0.1322431))*(-0.30173266) Msg2 u_prev1*(u_prev2 - (-0.019017395)*(uyy1 - (-0.37160122)*(uxx1 - uyy2))) Node u_prev + (u_prev - ux + uy*(-0.50821894))*0.0040541845 1.0% 3.0767E-03 3.0589E-03 Msg1 ((u_prev1 + (u_prev1 - (uyy1 * 0.007043898))) + ((ux2 + (uy1 * -0.43381253)) * 0.08615916)) * ((sin(u_prev2 - (u_prev1 * 1.4456807)) + ((t * 0.087506555) + 0.26573014)) * 0.9458101) Msg2 sin(sin(sin(((sin(sin(u_prev1) - t) * -0.38601163) * u_prev1) - ((((-0.79170316 - u_prev1) * u_prev1) - (u_prev2 * -1.593338)) * (sin(u_prev2 / 0.68245786) * -1.6866558)))) Node u_prev - ((((e1 - square((ux + cos(uy)) - u_prev)) * 0.049613487) + ((ux - u_prev) + (uy / (sin((y + uy) * -0.3068668) + 2.3308575)))) * 0.0039940816) 2.0% 2.3087E-03 2.3142E-03 Msg1 0.010929662*(-ux1 + ux2) Msg2 u_prev1*(u_prev1 - (t*0.08562517 + sin(u_prev2)))/0.25098807 Node ((((e1 + sin(uyy)) * 0.37931076) + (u_prev - ((ux + uy) + ux))) * (0.0019810875 - ((ux - (uy + (uxx * -0.06388371))) * 0.000119407356))) + u_prev 5.0% 1.5740E-03 1.5162E-03 Msg1 (u_prev1 * ((-0.4607998 - u_prev1) + (t * 0.13325739))) + (u_prev2 / 0.87254393) Msg2 (((uxx2 * u_prev1) + t) + sin(uy2)) * ((t * 0.0018317241) - ((sin(u_prev2) - u_prev1) * -0.022658467)) Node u_prev + ((((uy * y) - e1) + (((cos(ux) - exp(u_prev * -107.679985)) + ux) - e1)) * -0.0040704776) 10.0% 1.3019E-02 1.4153E-02 Msg1 u_prev2 * (cos(sin(-0.037906747) * ((ux1 / -0.33668214) + uyy1)) * 0.39657518) Msg2 uy2*(-0.007133365) - (-0.07567513)*(u_prev2 - 0.0005025013*uyy1*uyy1) Node u_prev + ((ux + ((sin(u_prev / 0.021259548) / (square(ux) + 0.21166894)) + -0.30247244)) * -0.0043281894) Table 4: Unedited symbolic expressions discovered for the Incompressible Navier-Stokes Equations across varying noise levels. Noise MSE(Noise) MSE(Original) Type Equation 0.0% / 9.6899E-04 Msg1 -sin(t/(-6.233686) + y_prev2) - 0.6374596 Msg2 (cos((-0.118906446 * t) - cos(x_prev1 + 0.21283843)) - y_prev2) * 0.6010129 Node1 sin(y - 1.5701683)*(ux + uy - sin(x))*cos(sin(t*(-0.1646834)) - 1*0.1230068)*0.9365822 Node2 ((sin(t * -0.13451573) + 1.6071377) * sin(y)) * ((cos(x) - 0.007157283) * -0.6139257) 0.1% 6.2724E-04 6.2718E-04 Msg1 (-x_prev2 + y_prev1)*0.015265973 Msg2 0.15940122 + (t - x_prev2)*(-0.026457742) Node1 ((e0 * -0.0015770864) + 0.9900558) * x_prev Node2 (0.9902582 - (e1 * 0.00073871564)) * y_prev 0.5% 6.3068E-04 6.2790E-04 Msg1 0.23712291 + cos(-uy1 + x_prev1 - y_prev1)*(-0.21217735) Msg2 (uy1 + y_prev1)*(-0.004990364) Node1 (e1 * -0.0046218093) - (x_prev * -0.9900131) Node2 (e1 * 0.0029781198) + (y_prev * 0.9900553) 1.0% 6.5369E-04 6.4082E-04 Msg1 (-x_prev1 + x_prev2)*(-0.07927439) Msg2 (y_prev2 / (t / -0.006975147)) + -0.0055887355 Node1 (x_prev * 0.9897041) + (e0 * -0.011039681) Node2 ((e1 * -0.005140257) + y_prev) * 0.9898316 2.0% 1.0362E-03 2.1334E-03 Msg1 (x_prev1 - x_prev2 + y_prev1 - y_prev2)*(-0.3086815) - 1*0.012407416 Msg2 (-0.4978986*x_prev1 - (-0.5724906)*x_prev2 + y_prev1 - 1.0975764*y_prev2)/2.5570834 Node1 (x_prev * 0.9885945) - (e0 * -0.0075576548) Node2 ((e1 * -0.0038489283) + y_prev) * 0.98886806 5.0% 6.1238E-03 6.8773E-03 Msg1 (-y_prev2 + torch.sin(x_prev1 - x_prev2 + y_prev1))*0.5179765 - 1*0.01374288 Msg2 (0.34813726 - uy1)*(-0.040036205) + (-x_prev2 + y_prev2 + (x_prev1 - y_prev1)*0.9260883)*(-0.5167467) Node1 e1 + (((x_prev - e1) * 0.98433083) - (e0 * 0.019087898)) Node2 y_prev - ((e0 + (y_prev + torch.square(e1 + 0.27977818))) * 0.018425854) 10.0% 3.9270E-02 3.9270E-02 Msg1 x_prev2*0.34059486 + (x_prev1 + y_prev1*0.92205274 - y_prev2)*(-0.27819878) - 1*0.008375836 Msg2 (-x_prev1 + y_prev1 + (x_prev2 - y_prev2 + 0.02801064)*1.1195679)*0.37756348 Node1 (e0*(-0.089349516) - x_prev)*(-0.9576391) Node2 y_prev + (x*0.37457064 - (-t - 1.8729712)*(-e0 + e1*2.0220585 + y_prev) - 1*1.8696856)*(-0.010119061)