← Back to papers

Paper deep dive

Beyond Interleaving: Causal Attention Reformulations for Generative Recommender Systems

Hailing Cheng

Year: 2026Venue: arXiv preprintArea: cs.IRType: PreprintEmbeddings: 46

Intelligence

Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 96%

Last extracted: 3/13/2026, 1:09:53 AM

Summary

The paper introduces two novel architectures, AttnLFA and AttnMVP, to replace the inefficient token-interleaving paradigm in generative recommender systems. By explicitly modeling the causal dependency between items and actions through attention-based pooling, these models reduce sequence complexity by 50%, decrease training time, and improve recommendation accuracy by mitigating attention noise.

Entities (5)

AttnLFA · architecture · 100%AttnMVP · architecture · 100%Transformer · model-architecture · 100%Generative Recommender Systems · system-category · 95%HSTU · architecture · 95%

Relation Signals (3)

AttnLFA improves Generative Recommender Systems

confidence 95% · Experimental results show that AttnLFA and AttnMVP consistently outperform interleaved baselines

AttnMVP improves Generative Recommender Systems

confidence 95% · Experimental results show that AttnLFA and AttnMVP consistently outperform interleaved baselines

AttnLFA reduces Sequence Complexity

confidence 90% · AttnLFA and AttnMVP... eliminate interleaved dependencies to reduce sequence complexity by 50%

Cypher Suggestions (2)

Identify systems improved by the proposed architectures · confidence 95% · unvalidated

MATCH (a:Architecture)-[:IMPROVES]->(s:System) RETURN a.name, s.name

Find all architectures proposed in the paper · confidence 90% · unvalidated

MATCH (a:Architecture) WHERE a.name IN ['AttnLFA', 'AttnMVP'] RETURN a

Abstract

Abstract:Generative Recommender Systems (GR) increasingly model user behavior as a sequence generation task by interleaving item and action tokens. While effective, this formulation introduces significant structural and computational inefficiencies: it doubles sequence length, incurs quadratic overhead, and relies on implicit attention to recover the causal relationship between an item and its associated action. Furthermore, interleaving heterogeneous tokens forces the Transformer to disentangle semantically incompatible signals, leading to increased attention noise and reduced representation this http URL this work, we propose a principled reformulation of generative recommendation that aligns sequence modeling with underlying causal structures and attention theory. We demonstrate that current interleaving mechanisms act as inefficient proxies for similarity-weighted action pooling. To address this, we introduce two novel architectures that eliminate interleaved dependencies to reduce sequence complexity by 50%: Attention-based Late Fusion for Actions (AttnLFA) and Attention-based Mixed Value Pooling (AttnMVP). These models explicitly encode the $i_n \rightarrow a_n$ causal dependency while preserving the expressive power of Transformer-based sequence this http URL evaluate our framework on large-scale product recommendation data from a major social network. Experimental results show that AttnLFA and AttnMVP consistently outperform interleaved baselines, achieving evaluation loss improvements of 0.29% and 0.80%, and significant gains in Normalized Entropy (NE). Crucially, these performance gains are accompanied by training time reductions of 23% and 12%, respectively. Our findings suggest that explicitly modeling item-action causality provides a superior design paradigm for scalable and efficient generative ranking.

Tags

ai-safety (imported, 100%)csir (suggested, 92%)preprint (suggested, 88%)

Links

PDF not stored locally. Use the link above to view on the source site.

Full Text

45,735 characters extracted from source content.

Expand or collapse full text

Beyond Interleaving: Causal Attention Reformulations for Generative Recommender Systems Hailing Cheng haicheng@linkedin.com Linkedin Inc Mountain View, California, USA Abstract Generative recommender systems (GR), exemplified by Meta’s HSTU ranker architecture, model user behavior as a sequence gen- eration problem by interleaving item and action tokens. While effective, this formulation introduces fundamental limitations: it doubles sequence length, incurs quadratic computational overhead, and relies on implicit attention mechanisms to recover the causal relationship that an item interaction푖 푛 elicits a user action푎 푛 . More- over, interleaving heterogeneous item and action tokens forces Transformers to disentangle semantically incompatible signals, in- troducing attention noise and reducing representation efficiency. This work presents a principled reformulation of generative recommendation—specifically targeting architectures like Meta’s HSTU as a ranker—by aligning sequence modeling with causal structure and attention theory. We demonstrate that the interleav- ing mechanism prevalent in current models acts as an inefficient proxy for similarity-weighted action pooling. To address this, we propose a structural shift that explicitly encodes the푖 푛 → 푎 푛 causal dependency. We introduce two novel architectures, Attention- based Late Fusion for Actions (AttnLFA) and Attention-based Mixed Value Pooling (AttnMVP), which eliminate interleaved de- pendencies to reduce sequence complexity by 50%. Our framework enforces strict causal attention while preserving the expressive power of Transformer-based sequence modeling, providing a theo- retically grounded and computationally efficient path for generative ranking. AttnLFA performs causal attention pooling over historical ac- tions conditioned on item similarity, whereas AttnMVP further in- tegrates action signals early by mixing item and action embeddings in the Transformer value stream, progressively learning preference- aware item representations. From an information-theoretic perspec- tive, AttnMVP reduces attention noise by aligning the attention space with the true causal graph of user behavior, enabling more efficient representation learning. We evaluate our methods on large-scale product recommenda- tion data from a major social network. Compared with the inter- leaved ranker baseline, AttnLFA and AttnMVP achieve consistent Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from permissions@acm.org. KDD 2026, Jeju, Korea © 2026 Copyright held by the owner/author(s). Publication rights licensed to ACM. ACM ISBN 978-1-4503-X-X/2026/06 https://doi.org/X.X improvements in evaluation loss by 0.29% and 0.8%, and normal- ized entropy (NE) gains across multiple tasks, while reducing training time by 23% and 12% respectively. Ablation studies con- firm that early, causally constrained fusion of action signals is the primary driver of performance gains. Overall, our results demon- strate that explicitly modeling item–action causality yields more efficient, scalable, and accurate generative recommender systems, offering a new design paradigm beyond token interleaving. CCS Concepts • Information systems→ Recommender systems. Keywords Generative Recommenders ACM Reference Format: Hailing Cheng. 2026. Beyond Interleaving: Causal Attention Reformulations for Generative Recommender Systems. In Proceedings of The 32nd ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD 2026). ACM, New York, NY, USA, 9 pages. https://doi.org/X.X 1 Introduction Generative recommender(GR) systems (e.g., Meta’s Hierarchical Se- quential Transduction Units (HSTU) [19]) model user behavior as a sequential prediction problem. These systems adopt Transformer ar- chitectures [5,7,12,15,16] originally developed for large language models (LLMs) [4, 8, 20] and formulate user action prediction as a token generation process by interleaving item and action tokens. Under this formulation, user actions are generated autoregressively [3,9] following item tokens, enabling the model to naturally cap- ture temporal dependencies in user interaction sequences. This paradigm represents a significant departure from traditional rec- ommendation models [1,11,17] and has demonstrated superior performance over conventional deep neural networks, largely due to its transductive learning capability and effective utilization of long-range user histories. As a result, generative recommendation has been widely adopted as a new modeling paradigm in industrial recommender systems. Despite its success, the interleaving formulation introduces fun- damental limitations. Semantic Heterogeneity of Tokens: In natural language, to- kens share a common semantic space and are compositional by nature. In contrast, recommender systems operate over fundamen- tally heterogeneous entities: items (e.g., posts, videos, products) and actions (e.g., click, like, share). Let푖 푛 ∈Idenote an item token and푎 푛 ∈ Aan action token, whereIandAare disjoint semantic spaces. Interleaving these tokens into a single sequence arXiv:2603.10369v1 [cs.IR] 11 Mar 2026 KDD 2026, June 03–05, 2026, Jeju, KoreaHailing Cheng 푖 0 푎 0 푖 1 푎 1 푖 2 푎 2 Figure 1: Interleaved generative recommenders treat items and actions as a single token stream. Action푎 2 attends to all prior tokens, obscuring the direct causal dependency 푖 2 → 푎 2 and introducing attention noise. x=[푖 0 ,푎 0 ,푖 1 ,푎 1 , . . .,푖 푛 ,푎 푛 ]implicitly assumes a shared latent struc- ture acrossI∪A. This assumption is weak in practice: the seman- tic relationship between an item and an action is asymmetric and causal, rather than compositional. Treating them as homogeneous tokens forces the Transformer to learn artificial alignments that are not grounded in the underlying data-generating process. Missing Explicit Causality in Self-Attention: In our sequen- tial framework, we model user actions through a localized causal lens. While a user’s global state is informed by their interaction historyH <푛 = (푖 푘 ,푎 푘 ) 푘<푛 , we posit that the specific action푎 푛 is primarily a response to the proximal stimulus푖 푛 . Formally, we define the action probability as conditioned on the current item, where the historical sequence acts as a contextual moderator rather than a direct determinant: 푃(푎 푛 | 푖 푛 ,H <푛 ) ≈ 푃(푎 푛 | 푖 푛 ;휃 H <푛 ) Standard interleaved formulations often fail to explicitly articulate this functional mapping, treating푎 푛 and푖 푛 as part of a homogenous sequence. In the standard causal self-attention mechanism: Attn(푄 푛 ,퐾 ≤푛 ,푉 ≤푛 ) where items and actions up to index푛contribute symmetrically to the attention computation, two primary issues arise: •Causal Dilution: Action tokens attend to the entire histori- cal prefix, which dilutes the direct causal dependency on푖 푛 with unrelated historical signals. •Structural Ambiguity: Item tokens face difficulty in map- ping specific previous actions to their corresponding items due to the uniform distribution of attention weights across all historical tokens. Critically, positional encodings alone are insufficient to recover this structure, as they encode sequence order but lack the capacity to enforce the requisite item-action causal pairing. Attention Noise Induced by Interleaving: Even if high ca- pacity Transformer architectures are theoretically capable of ap- proximating latent item-action correspondences, the interleaved sequence format introduces systematic attention noise. Specifically, once the model establishes a strong causal dependency between푖 푛−1 and푎 푛−1 , the subsequent token푖 푛 —due to the locality-preserving nature of Rotary Position Embedding (RoPE) [14] or Relative At- tention Bias (RAB) [13]—inherits a nearly identical attention bias toward푎 푛−1 . This leads to the formation of spurious dependencies where푖 푛 attends to푎 푛−1 with an unwarranted inductive bias, re- gardless of their semantic or causal relevance. Such architectural artifacts impose an unnecessary burden on subsequent layers to ’correct’ these correlations, ultimately degrading sample efficiency and complicating the optimization landscape. 푖 0 푖 1 푖 2 푎 0 푎 1 푎 2 Ground-truth causal dependency Figure 2: True causal structure of user interactions. Each action푎 푛 is a response to the corresponding item 푖 푛 , conditioned on prior history. This structure is not explicitly represented by interleaved self-attention. Computational Inefficiency: Interleaving item and action to- kens increases the effective sequence length from푁to 2푁. Because self-attention has quadratic complexity in sequence length, this results in approximately a 4×increase in both memory and com- putational cost. Such overhead is especially detrimental in long- horizon recommendation settings, where scalability is a primary concern. The increased sequence length substantially prolongs both training and inference, reduces GPU utilization efficiency, and leads to higher energy consumption, making interleaving-based formula- tions less suitable for large-scale, production-grade systems. In this work, we propose a new formulation for Transformer- based generative recommender (GR) models that addresses the fun- damental limitations of interleaved item–action sequence modeling. We first provide an explicit interpretation of interleaving-based GR methods, showing that they implicitly perform attention-based aggregation over historical user actions conditioned on item repre- sentations. Based on this observation, we reformulate GR modeling as an attention pooling mechanism in which item embeddings are used to construct the query and key projections, while action em- beddings are incorporated exclusively through the value projections under a strict causal masking scheme. Building on this formulation, we introduce two improved Trans- former architectures that decouple item and action representations while preserving sequential dependency modeling. We evaluate the proposed models on large-scale product recommendation data collected from a major social networking platform and demonstrate consistent improvements over interleaving-based baselines in both recommendation accuracy and computational efficiency. After the Transformer encoder, the item representations are discarded, and the final action token representation is extracted. This representa- tion is concatenated with additional late-fusion features and fed into a task-specific prediction head, typically implemented using a Multi-gate Mixture-of-Experts (MMoE) architecture [10], to gener- ate predictions for the target actions associated with the current item. The item-side features comprise content and metadata repre- sentations associated with each item, including text embeddings, author embeddings, item type indicators, and related attributes. To incorporate personalization signals, we additionally include se- lected viewer-side contextual features, such as device type. The action features capture user feedback signals for each item, includ- ing click, skip, dwell time, like, share, and comment, etc. In our Beyond Interleaving: Causal Attention Reformulations for Generative Recommender SystemsKDD 2026, June 03–05, 2026, Jeju, Korea Figure 3: Traditional Generative Recommender (In- terleaving Item and Action Tokens) architecture: the item and action tokens are interleaved before the trans- former layers setting, user behavior is modeled using more than ten supervised action labels to represent engagement intensity and preference. 2 Architecture overview and attention mechanism Figure 3 illustrates a standard Transformer-based generative rec- ommender (GR) architecture. Raw item features are first concate- nated and passed through a projection network to produce item embeddings. Similarly, action features—corresponding to multi- task supervision signals such as click, dwell time, like, share, and comment—are concatenated and projected into action embeddings using a separate projection network. The resulting item and action embeddings are then interleaved to form an input token sequence of the form[푖 0 ,푎 0 ,푖 1 ,푎 1 , . . .], which is processed by a stack of Trans- former layers (12 layers in our implementation). In addition to sequence-level representations, we employ a set of late-fusion features, primarily consisting of counting statistics. These features are introduced to improve score calibration, which is critical in production ranking systems where items are ordered according to a value function that combines multiple predicted outcomes. Stable and well-calibrated prediction scores are therefore essential for reliable and controllable ranking behavior. For fairness and reproducibility, we use the same feature set across all model architectures evaluated in this paper. As feature engineering is not the focus of this work, we omit further details and concentrate on architectural and modeling differences in the subsequent sections. We use a toy example to provide an intuitive illustration of how the generative recommender (GR) architectures operate. As shown in Figure 4, we consider two users with distinct preference profiles. User A consistently exhibits positive engagement with dog-related items and negative engagement with cat-related items, whereas User B demonstrates the opposite behavior. Given a candidate item belonging to either category, the task is to predict the user’s next ac- tion and, consequently, decide whether a dog-related or cat-related item should be presented. Figure 4: Illustrative toy sequences for Users A and B. The sequences demonstrate contrasting behavioral patterns: User A consistently exhibits positive interac- tions (e.g., "Like") with dog-related items and negative interactions with cat-related items, while User B ex- hibits the inverse preference profile. This highlights the model’s task of capturing item-action dependen- cies for future state prediction. In a non-sequential point-wise recommendation setting without personalization, the model would assign approximately equal prob- ability to dog and cat items for both users, as predictions would be driven solely by aggregate item statistics across the population. To incorporate personalization in such settings, one must rely on hand- crafted historical features, such as per-user like rates for dog and cat items (e.g., 100% vs.0% for User A, and vice versa for User B). This approach places a heavy reliance on manual feature engineering and explicit content understanding. In particular, accurate personalization requires the model to dis- tinguish between fine-grained item semantics (e.g., dog versus cat). If the content representation collapses both categories into a coarse concept such as “animal,” the predictive signal is substantially weak- ened. Moreover, achieving fine-grained semantic understanding typically necessitates extensive item taxonomies and carefully engi- neered historical statistics, which are costly to build and maintain at scale. These limitations motivate sequence-based generative rec- ommender models, which aim to learn such preference patterns directly from user interaction sequences without relying on bespoke feature engineering. Interleaved GR architectures implicitly recover such personalized signals by modeling user behavior as a single token sequence of the form[푖 0 ,푎 0 ,푖 1 ,푎 1 , . . .]. Within this formulation, the self-attention KDD 2026, June 03–05, 2026, Jeju, KoreaHailing Cheng mechanism can associate an item token푖 푛 with its subsequent ac- tion token푎 푛 , allowing the model to infer user preferences through repeated item–action co-occurrences. In the toy example, atten- tion can learn that dog items are frequently followed by positive actions for User A but negative actions for User B, and vice versa for cat items. As a result, interleaving enables the model to encode user-specific preferences without explicit handcrafted features or predefined item taxonomies. We now examine the Transformer dynamics underlying inter- leaved GR models in more detail. After one Transformer layer, User A’s interleaved sequence may evolve from[푖 0 = dog,푎 0 = like] to contextualized representations[푑표푔 ′ 0 ,푙푖푘푒 ′ 0 + 훼 · 푑표푔 ′ 0 ], where the action token aggregates information from its associated item through self-attention. From this perspective, the Transformer’s attention mechanism effectively performs a similarity-weighted pooling over historical item and action representations. Consider a subsequent Transformer layer in which a dog-related item token푑표푔 ′ 푛 attends over the historical sequence to produce its updated representation푑표푔 ′ 푛 . When attending to the action token푙푖푘푒 ′ 0 + 훼 ·푑표푔 ′ 0 , the attention weight is amplified due to the high semantic similarity between푑표푔 ′ 푛 and푑표푔 ′ 0 . In contrast, action tokens associated with semantically dissimilar items (e.g., cats) receive lower attention weights. As stacking multiple Transformer layers compounds this effect, the model progressively disentangles fine-grained semantic concepts such as “liked dog,” “disliked cat,” “positive feedback on dog,” and “negative feedback on cat,” even though these abstractions are never explicitly encoded. This analysis suggests that the effectiveness of interleaved GR models stems from using self-attention as a structured pooling operator that implicitly associates items with their corresponding user actions via semantic similarity. However, this association is formed only indirectly and incurs substantial overhead. By forcing heterogeneous item and action tokens into a single sequence, the attention mechanism must disentangle fundamentally different semantic types, introducing spurious interactions and increasing representational noise, while simultaneously doubling the effective sequence length. For instance, in User A’s history, when a token corresponding to푐푎푡 1 attends to preceding tokens, it can only attend to푑표푔 0 and 푙푖푘푒 0 . As a result, its contextualized representation takes the form 푐푎푡 ′ 1 +훽·푑표푔 ′ 0 +훾·푙푖푘푒 ′ 0 . Even though we know User A consistently dislikes cats, this representation may still inherit positive signals associated with actions such as “like on dog” through attention, effectively encoding a “partially liked cat.” We refer to this phe- nomenon as attention noise, arising from indiscriminate mixing of heterogeneous tokens within interleaved self-attention. Prior work has sought to mitigate the infrastructure and com- putational costs induced by interleaved generative recommender architectures. For example, Huang et al. (2025) [6] propose an early- fusion formulation in which item and action signals are embedded into a unified feature space, and a dummy action token is injected for the target item to prevent label leakage. While effective in re- ducing sequence length, this approach fundamentally partitions the user sequence into a context segment and a candidate segment. In the context segment, actions are treated solely as input features and cannot serve as supervision, whereas in the candidate segment, actions are used exclusively as labels and are unavailable as features. As a result, long user histories cannot be trained end-to-end; instead, the sequence must be artificially decomposed and processed in a staged or progressive manner, requiring multiple training passes and introducing additional system overhead. Moreover, the injected dummy actions in the candidate segment differ substantially from the real action embeddings used in the context segment, creating a distributional mismatch. This discrepancy distorts the attention pat- terns learned across segments and introduces an additional source of modeling noise, potentially degrading representation quality and convergence stability. While Wei et al. (2025) [18] introduced Lagged Action Condition- ing (LAC) to utilize(푎 푛−1 ,푖 푛 )pairings as input tokens, we argue that such transitions lack inherent semantic coherence. In typical recommendation environments, item sequences are exogenously de- termined, rendering the(푎 푛−1 ,푖 푛 )pairing a structural artifact rather than a valid causal transition. Despite the widespread adoption of such interleaved formats for formal consistency, the literature lacks a first-principles analysis of how Transformer architectures process these synthetic dependencies and the potential noise they intro- duce. These limitations motivate a more principled formulation that preserves the expressive benefits of interleaving—namely, learning item–action associations directly from sequences—while avoiding its representational and computational drawbacks. In the next sec- tion, we introduce an alternative attention-based formulation that explicitly models this dependency without interleaving tokens. 3 AttnLFA: Attention-based Late Fusion for Action Architecture Building on the analysis in the preceding sections, we argue that an effective generative recommender must explicitly encode the causal relationship that an exposed item푖 푛 induces a subsequent user ac- tion푎 푛 . Guided by this principle, we depart from the conventional next-token prediction formulation of generative recommendation and adopt a fundamentally different perspective. Our key insight is that user actions can be modeled as a similarity-weighted ag- gregation over historical actions: if a target item is semantically similar to previously consumed items, then the user’s response to the target item should resemble the actions associated with those similar items. Under this formulation, the recommendation problem is cast as an item-conditioned action pooling task, where attention serves as a structured, similarity-based pooling operator. Based on this idea, we propose the attention-based late-fusion architecture illustrated in Figure 5. Item embeddings and action embeddings are maintained as separate representation streams. Item embeddings are processed by a stack of Transformer layers to produce contextualized item representations. The final-layer item embeddings are then used as both Queries and Keys, while action embeddings are supplied as Values in the attention operation, yielding an item-conditioned pooled action representation. To prevent label leakage, the proposed attention pooling is en- forced under a strict causal constraint: the representation corre- sponding to item푖 푛 may attend only to items at positions0, . . .,푛− 1, and is explicitly prohibited from attending to its own position푛. Beyond Interleaving: Causal Attention Reformulations for Generative Recommender SystemsKDD 2026, June 03–05, 2026, Jeju, Korea Figure 5: Attention-based Late Fusion for Action (At- tnLFA). Item embeddings are transformed through a series of Transformer blocks (labeled as "Transform- ers" for clarity) to generate latent sequence representa- tions. These representations serve as both Queries and Keys for the subsequent attention mechanism. In the final stage, action embeddings are integrated as Values via a causally-constrained attention pooling operation, conditioned on the sequence context. The resulting aggregated action representation is then passed to the prediction head for the final output. Although such constraints can be expressed via customized atten- tion masks, these masks are not efficiently supported by FlashAt- tention [2] kernels and significantly degrade kernel fusion and throughput. To leverage high-throughput GPU kernels while maintaining compatibility with standard FlashAttention implementations, we employ a query-shifting mechanism to enforce a strict causal constraint 6. Specifically, we set theis_causalflag in the module of scaled_dot_product_attentionand apply a one-step left-shift to the query sequence푞 1 , . . .,푞 푛 relative to the keys. This en- sures that each query푞 푖 is restricted to the preceding key prefix 푘 1 , . . .,푘 푖−1 , effectively preventing self-attention. Post-computation, we apply left-side zero-padding to the attention outputs to restore temporal alignment with the original sequence. Under this formula- tion, the first item푖 0 naturally produces a null value representation, reflecting the absence of prior context in the sequence. This ensures structural consistency and allows for seamless integration with downstream Transformer layers, while preserving computational efficiency and numerical stability. We refer to this architecture as Attention-based Late Fusion for Actions (AttnLFA). Figure 6: The query-shifting mechanism to enforce a strict causal constraint. To assess effectiveness, we compare AttnLFA against a strong interleaved-token baseline. Experiments are conducted on large- scale product recommendation logs collected from one of the largest professional social networks. User interaction sequences of up to 1024 events are constructed over the past 12 months and partitioned temporally. The partition immediately following the training win- dow is reserved for evaluation. Each evaluation sequence is further divided into a context segment, comprising interactions occurring before the training cutoff, and a candidate segment, comprising in- teractions occurring afterward. To ensure a rigorous and controlled comparison, we maintain identical hyperparameter configurations and architectural components—including embedding layers, Trans- former blocks, and projection heads—across all evaluated models, unless otherwise specified. Each model is trained for a single epoch. Furthermore, we employ Rotary Positional Embeddings (RoPE) [14] as the universal positional encoding scheme throughout our experiments to ensure consistency in spatial modeling. To faithfully approximate online serving conditions, we apply a timestamp-based label masking scheme during evaluation. Loss and metrics are computed exclusively on candidate items. For the context segment, standard causal masking is applied to interleaved baseline, while strict causal masking is enforced for (AttnLFA) . For the candidate segment, candidate items are prohibited from attending to one another. This evaluation protocol prevents information leakage while maintaining consistency with real-world recommendation constraints. We evaluate model performance across three primary engage- ment signals, formulated as binary classification tasks: •Long Dwell: A binary indicator 1(푑푤푒푙 _푡푖푚푒> 휏), where 푑푤푒푙 _푡푖푚푒is the user’s dwell time and휏is a predefined temporal threshold. •Contribution: A multi-action signal where the label is posi- tive if the user performs at least one non-click engagement (e.g., like, comment, or share). • Like: A specific binary task representing whether the user explicitly engaged with the "like" mechanism. KDD 2026, June 03–05, 2026, Jeju, KoreaHailing Cheng Model Eval Loss LongDwell NE Contribution NE Like NETime Baseline----- AttnLFA -0.29% -0.06%-0.49%-0.47% -22.8% Table 1: Performance comparison between AttnLFA and Baseline. We report the relative improvement across four key dimensions: (a) Multi-task Binary Cross Entropy (BCE) Loss, (b) evaluation Normalized Entropy (NE) for Long Dwell, Contribution, and Like actions, and (c) total training latency. AttnLFA demon- strates superior predictive accuracy (lower NE/Loss) while maintaining competitive computational effi- ciency. Table 1 summarizes the experimental results, where all multi-task models use the binary labels in each task, and are optimized using standard Binary Cross Entropy (BCE) loss. For brevity, we report performance on three representative major tasks; the remaining nine tasks exhibit consistent performance trends and are omitted for conciseness. AttnLFA achieves substantial improvements in evaluation loss and normalized entropy (NE) across the pri- mary prediction tasks. In addition, by eliminating the interleaving formulation, the proposed approach reduces end-to-end training time by 22.8%, demonstrating both modeling and computational ef- ficiency gains. The observed improvements corroborate our central hypothesis: accurately modeling the causal relationship from item 푖 푛 to action푎 푛 is critical for improving predictive performance. In AttnLFA, this causality is explicitly enforced by design—each ac- tion representation푎 푛 is obtained by attention-based pooling driven solely by the corresponding item representation푖 푛 . This mecha- nism operationalizes a simple but powerful principle: items that are semantically similar should induce similar action distributions. Encouraged by the effectiveness of AttnLFA, we further ex- tend this idea to a more expressive variant. Specifically, instead of applying action embeddings through late fusion, we explore an early-fusion formulation that integrates item–action interactions at earlier stages of representation learning. 4 AttnMVP: Attention based Mixed Value Pooling Architecture Building on the effectiveness of AttnLFA, we introduce an early- fusion variant termed Attention-based Mixed Value Pooling (Attn- MVP), shown in Figure 7. AttnMVP reformulates action-conditioned sequential recommendation as a causally constrained representa- tion learning problem, in which item representations are iteratively refined by integrating historical user actions through attention- based value mixing. Leti 푡 푇 푡=1 denote the sequence of item embeddings anda 푡 푇 푡=1 the corresponding action embeddings (e.g., click, dwell, like), follow- ing the conventions of SASRec. At Transformer layerℓ, AttnMVP applies self-attention over item representations using Q (ℓ) = K (ℓ) = H (ℓ−1) , where H (0) = i 푡 and H (ℓ) denotes the item representations after layerℓ. The value vectors are constructed via mixed-value fusion: V (ℓ) 푡 =H (ℓ−1) 푡 + 휆a 푡 , where휆 ≥0 controls the contribu- tion of action signals. We adopt an additive fusion for H (ℓ−1) 푡 and Figure 7: Attention-based Mixed Value Pooling (Attn- MVP) architecture. Item embeddings serve as Queries and Keys in each Transformer layer, while item and action embeddings are additively fused as mixed Val- ues. Across stacked layers, action signals are progres- sively injected into item representations under strict causal constraints. In the final stage, action embed- dings are pooled via causally masked attention condi- tioned on the sequence-level item representations, and the pooled action representation is fused with the final item embedding to produce action predictions. 휆 a 푡 primarily to prioritize computational efficiency and maintain a lightweight architecture. While this linear combination minimizes overhead, exploring more sophisticated fusion mechanisms—such as gated pooling —remains a promising avenue for future research and is expected to yield further performance gains. In practice, we set휆=1, and preliminary sensitivity analysis over휆 ∈ [0.5,1.0] indicates stable performance. After푇 (= 12) layers of Transformer blocks, at the final layer, we apply an action pooling operation identical to AttnLFA. This formulation enables each Transformer layer to perform causally masked, attention-weighted aggregation of historical action signals into item representations, conditioned on item similarity. As item embeddings propagate through succes- sive layers, they evolve from encoding generic content semantics (e.g., dog versus cat) to capturing user-conditioned semantics (e.g., preferred dog versus disfavored cat). Through this progressive inte- gration, user preference signals are implicitly embedded into the Beyond Interleaving: Causal Attention Reformulations for Generative Recommender SystemsKDD 2026, June 03–05, 2026, Jeju, Korea item representations, amplifying item–item contrast in a person- alized manner. Notably, this personalization emerges end-to-end from the attention mechanism itself, without requiring explicit user profiling or handcrafted personalization features. From a representation learning perspective, AttnMVP explicitly encodes the inductive bias that semantically related items elicit analogous user responses. By decoupling item and action represen- tations, the model circumvents the heterogeneous token entangle- ment and quadratic computational overhead inherent in interleaved generative frameworks. Consequently, AttnMVP offers a principled and scalable alternative to token-level interleaving for sequential behavioral modeling. Model Eval Loss LongDwell NE Contribution NE Like NETime Baseline----- AttnMVP -0.80% -0.41%-1.1%-1.1% -12.3% AttnMVP - LFA -0.78% -0.40%-1.0%-1.0% -13.02% Table 2: Relative performance improvement of Eval Loss, major tasks’ Eval Normalized Entropies (NEs) and Training Time for Baseline, AttnMVP and AttnMVP without LFA. (AttnMVP-LFA) Table 2 reports the relative improvements in evaluation loss, normalized entropy (NE) for major prediction tasks, and training time. Compared with the interleaved ranker baseline, AttnMVP delivers consistent and larger gains in both loss and NE across all tasks, while also reducing training time by 12.3%. These im- provements exceed those achieved by AttnLFA, indicating the significant benefit of integrating action information earlier in the representation learning process. To isolate the contribution of early fusion, we further evaluate a variant of AttnMVP that retains mixed-value fusion within Trans- former layers but removes the late fusion attention (denoted as AttnMVP–LFA). This variant achieves performance comparable to the full model, with only marginal degradation in loss and NE. The result suggests that the majority of the gains stem from early, causally constrained integration of action signals into item repre- sentations. This finding supports our hypothesis that explicitly en- coding preference-aware semantics—e.g., distinguishing “preferred” versus “disfavored” items—within the sequence representations is critical for effective generative recommendation modeling. 5 Future Work: AttnDHN - Attention based Dual-Helix Network Motivated by the strong empirical performance of AttnMVP, we further propose a symmetric dual-stream architecture, termed Attention- based Dual-Helix Network (AttnDHN). In AttnMVP, item repre- sentations are updated via self-attention with mixed item–action values, using(푄 푡 ,퐾 푡 ,푉 푡 )=(푖 푡 ,푖 푡 ,푖 푡 + 푎 푡 ). AttnDHN extends this formulation by introducing a complementary action-centric up- date, in which action representations are simultaneously refined using(푄 푡 ,퐾 푡 ,푉 푡 )=(푎 푡 ,푎 푡 ,푖 푡 +푎 푡 ). The details can be referred from Figure 8. Within each Transformer block, item and action streams are updated sequentially in a paired manner, forming a tightly cou- pled interaction unit. This alternating update mechanism induces a bidirectional flow of information between item and action repre- sentations, analogous to a double-helix structure, enabling more expressive and coherent co-evolution of user preference signals across network depth. To date, we do not observe AttnDHN to consistently outperform AttnMVP. We attribute this outcome to three primary factors. First, AttnDHN exhibits reduced training stability relative to AttnMVP; in practice, stable optimization requires halving the learning rate. When both models are trained for the same number of optimiza- tion steps, this constraint leads to weaker convergence and inferior performance for AttnDHN. Second, AttnDHN effectively doubles the number of Transformer updates per layer due to its dual-stream design, making direct comparisons with AttnMVP less straight- forward. Improvements observed at a fixed depth (e.g., 12 layers) cannot be cleanly attributed to architectural superiority, as similar gains might be achievable by increasing the depth of AttnMVP. Third, and more fundamentally, item and action tokens reside in highly heterogeneous semantic spaces: the action vocabulary is small (on the order of tens), whereas the item space is effectively unbounded. As a result, action-centric representations tend to be less expressive and noisier than item-centric representations (e.g., “preferred dog” versus a diffuse mixture of preferences over many unrelated items). Despite these limitations, we include AttnDHN as an exploratory architecture. Its symmetric dual-stream design may be better suited to settings in which the two representation spaces are more ho- mogeneous, such as multimodal recommendation scenarios that jointly model text and visual embeddings. We leave a systematic investigation of such applications to future work. 6 Conclusion We revisit the interleaved-token formulation common in generative recommendation and offer a first-principles critique of its opera- tional mechanics. Our analysis reveals that while self-attention effectively acts as a latent pooling mechanism for user actions via item-level semantics, the standard interleaved approach remains suboptimal. Specifically, this formulation introduces representa- tional noise and computational inefficiency by interleaving hetero- geneous tokens, which not only doubles sequence length but also complicates the attention landscape. Guided by the causal structure that an item interaction푖 푡 induces its corresponding user action푎 푡 , we propose a family of attention- based architectures that explicitly encode this causality without interleaving. We introduce AttnLFA, which formulates action mod- eling as causally masked attention-based late fusion, and AttnMVP, which further generalizes this idea through mixed-value early fu- sion that progressively integrates historical action information into item representations. From an information-theoretic perspective, these designs reduce attention noise by constraining aggregation to causally valid and semantically aligned interactions, enabling more efficient and expressive representation learning. Empirically, both architectures consistently outperform a strong interleaved ranker baseline on large-scale real-world recommen- dation data, achieving lower evaluation loss and normalized entropy across major tasks while substantially reducing training time and computational cost. These gains validate our central thesis: KDD 2026, June 03–05, 2026, Jeju, KoreaHailing Cheng Figure 8: Attention-based Dual-Helix Network (At- tnDHN) architecture. Action embedding and Item em- beddings are updated in pair-wise sequence in in- dividual Transformer layer. Both transformer layer use either action and item embedding as query and key, and use a combination of item + action embed- ding as value. In the final stage, action embeddings are pooled via causally masked attention conditioned on the sequence-level item representations, and the pooled action representation is fused with the final item embedding to produce action predictions. explicitly modeling the causal relationship from items to actions leads to both improved predictive accuracy and better system effi- ciency. Finally, we explore a symmetric dual-stream extension, AttnDNA, and discuss its limitations in standard recommender settings due to semantic heterogeneity between item and action spaces, while highlighting its potential applicability to more homogeneous mul- timodal scenarios. Overall, our results suggest that moving beyond interleaving toward causality-aware attention formulations offers a principled and scalable path forward for generative recommender systems. Acknowledgments The author gratefully acknowledges Samaneh Moghaddam and Ying Xuan for funding and support. Special thanks to Chen Zhu, Tao Huang, and Antonio Alonso for their guidance and discussions on model training. The author also thanks the LinkedIn Feed AI team, Apollo Engineering team and Core AI team for valuable discussions and infrastructure support. References [1]Heng-Tze Cheng, Levent Koc, Jeremiah Harmsen, Tal Shaked, Tushar Chandra, Hrishi Aradhye, Glen Anderson, Greg Corrado, Wei Chai, Mustafa Ispir, et al. 2016. Wide & Deep Learning for Recommender Systems. In Proceedings of the 1st Workshop on Deep Learning for Recommender Systems (DLRS). [2]Tri Dao, Daniel Y. Fu, Stefano Ermon, Atri Rudra, and Christopher Ré. 2022. FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness. In Advances in Neural Information Processing Systems (NeurIPS). https://arxiv. org/abs/2205.14135 [3]Jaixin Deng, Shiyao Wang, Kuo Cai Li, Qigen Ren, Qigen Hu, Weifeng Ding, Qiang Luo, and Guorui Zhou. 2025. OneRec: Unifying Retrieve and Rank with Generative Recommender and Iterative Preference Alignment. arXiv preprint arXiv:2502.18596 (2025). https://arxiv.org/abs/2502.18596 [4]Shijie Geng, Shuchang Liu, Zuohui Fu, Yingqiang Ge, and Yongfeng Zhang. 2022. Recommendation as language processing (rlp): A unified pretrain, personalized prompt & predict paradigm (p5). In Proceedings of the 16th ACM conference on recommender systems. 299–315. [5] Ruidong Han, Bin Yin, Shangyu Chen, He Jiang, Fei Jiang, Xiang Li, Chi Ma, Mincong Huang, Xiaoguang Li, Chunzhen Jing, Yueming Han, Menglei Zhou, Lei Yu, Chuan Liu, and Wei Lin. 2025. MTGR: Industrial-Scale Genera- tive Recommendation Framework in Meituan. In Proceedings of the 34th ACM International Conference on Information and Knowledge Management (CIKM). https://arxiv.org/abs/2505.18654 [6]Yanhua Huang, Yuqi Chen, Xiong Cao, Rui Yang, Mingliang Qi, Yinghao Zhu, Qingchang Han, Yaowei Liu, Zhaoyu Liu, Xuefeng Yao, Yuting Jia, Leilei Ma, Yinqi Zhang, Taoyu Zhu, Liujie Zhang, Lei Chen, Weihang Chen, Min Zhu, Ruiwen Xu, and Lei Zhang. 2025. Towards Large-scale Generative Ranking. arXiv:2505.04180 [cs.IR] https://arxiv.org/abs/2505.04180 [7]Wang-Cheng Kang and Julian McAuley. 2018. Self-Attentive Sequential Recom- mendation. In Proceedings of the IEEE International Conference on Data Mining (ICDM). https://arxiv.org/abs/1808.09781 [8] Wang-Cheng Kang, Jianmo Ni, Nikhil Mehta, Maheswaran Sathiamoorthy, Lichan Hong, Ed Chi, and Derek Zhiyuan Cheng. 2023. Do llms understand user prefer- ences? evaluating llms on user rating prediction. arXiv preprint arXiv:2305.06474 (2023). [9]Zida Liang, Changfa Wu, Dunxian Huang, Weiqiang Sun, Ziyang Wang, Yuliang Yan, Jian Wu, Yuning Jiang, Bo Zheng, Ke Chen, Silu Zhou, and Yu Zhang. 2025. TBGRecall: A Generative Retrieval Model for E-commerce Recommendation Scenarios. arXiv preprint arXiv:2508.11977 (2025). https://arxiv.org/abs/2508. 11977 [10]Jiaqi Ma, Zhe Zhao, Xinyang Yi, Jilin Chen, Lichan Hong, and Ed H Chi. 2018. Modeling task relationships in multi-task learning with multi-gate mixture-of- experts. In Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining. 1930–1939. [11]Maxim Naumov, Deepak Mudigere, Hao-Jun Michael Shi, Jianyu Huang, Narayanan Sundaraman, Jongsoo Park, Xiaodong Wang, Udit Gupta, Carole-Jean Wu, Alisson G. Azzolini, et al.2019. Deep Learning Recommendation Model for Personalization and Recommendation Systems. arXiv preprint arXiv:1906.00091 (2019). https://arxiv.org/abs/1906.00091 [12] Qiji Pei, Changhua Lv, Chao Li, Junfeng Ge, and Wenwu Ou. 2021. End-to-end User Behavior Retrieval in Click-through Rate Prediction Model. In Proceedings of the 28th International Conference on Machine Learning (ICML). [13]Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2023.Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. arXiv:1910.10683 [cs.LG] https://arxiv.org/abs/1910.10683 [14]Jianlin Su, Yu Lu, Shengfeng Pan, Ahmed Murtadha, Bo Wen, and Yunfeng Liu. 2023. RoFormer: Enhanced Transformer with Rotary Position Embedding. arXiv:2104.09864 [cs.CL] https://arxiv.org/abs/2104.09864 Beyond Interleaving: Causal Attention Reformulations for Generative Recommender SystemsKDD 2026, June 03–05, 2026, Jeju, Korea [15]Fei Sun, Jun Liu, Jian Wu, Changhua Pei, Xiao Lin, Wenwu Ou, and Peng Jiang. 2019. BERT4Rec: Sequential Recommendation with Bidirectional En- coder Representations from Transformer. In Proceedings of the 28th ACM Inter- national Conference on Information and Knowledge Management (CIKM). https: //arxiv.org/abs/1904.06690 [16]Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems 30 (2017). [17] Ruoxi Wang, Bin Fu, Gang Fu, and Ming Wang. 2021. DCN V2: Improved Deep & Cross Network and Practical Lessons for Web-scale Learning to Rank Systems. In Proceedings of the Web Conference (W). [18]Xiaokai Wei, Jiajun Wu, Daiyao Yi, Reza Shirkavand, and Michelle Gong. 2025. The Layout Is the Model: On Action-Item Coupling in Generative Recommenda- tion. arXiv:2510.16804 [cs.IR] https://arxiv.org/abs/2510.16804 [19] Jiaqi Zhai, Lucy Liao, Xing Liu, Yueming Wang, Rui Li, Xuan Cao, Leon Gao, Zhao- jie Gong, Fangda Gu, Michael He, Yinghai Lu, and Yu Shi. 2024. Actions Speak Louder than Words: Trillion-Parameter Sequential Transducers for Generative Recommendations. arXiv:2402.17152 [cs.LG] https://arxiv.org/abs/2402.17152 [20]Junjie Zhang, Ruobing Xie, Yupeng Hou, Xin Zhao, Leyu Lin, and Ji-Rong Wen. 2025. Recommendation as instruction following: A large language model em- powered recommendation approach. ACM Transactions on Information Systems 43, 5 (2025), 1–37. Received x 2026; revised x 2026; accepted x 2026