← Back to papers

Paper deep dive

Adaptive Decoding via Test-Time Policy Learning for Self-Improving Generation

Asmita Bhardwaj, Yuya Jeremy Ong, Eelaaf Zahid, Basel Shbita

Year: 2026Venue: arXiv preprintArea: cs.CLType: PreprintEmbeddings: 33

Abstract

Abstract:Decoding strategies largely determine the quality of Large Language Model (LLM) outputs, yet widely used heuristics such as greedy or fixed temperature/top-p decoding are static and often task-agnostic, leading to suboptimal or inconsistent generation quality across domains that demand stylistic or structural flexibility. We introduce a reinforcement learning-based decoder sampler that treats decoding as sequential decision-making and learns a lightweight policy to adjust sampling parameters at test-time while keeping LLM weights frozen. We evaluated summarization datasets including BookSum, arXiv, and WikiHow using Granite-3.3-2B and Qwen-2.5-0.5B. Our policy sampler consistently outperforms greedy and static baselines, achieving relative gains of up to +88% (BookSum, Granite) and +79% (WikiHow, Qwen). Reward ablations show that overlap-only objectives underperform compared to composite rewards, while structured shaping terms (length, coverage, repetition, completeness) enable stable and sustained improvements. These findings highlight reinforcement learning as a practical mechanism for test-time adaptation in decoding, enabling domain-aware and user-controllable generation without retraining large models.

Tags

ai-safety (imported, 100%)cscl (suggested, 92%)preprint (suggested, 88%)

Links

Your browser cannot display the PDF inline. Open PDF directly →

Intelligence

Status: not_run | Model: - | Prompt: - | Confidence: 0%

Entities (0)

No extracted entities yet.

Relation Signals (0)

No relation signals yet.

Cypher Suggestions (0)

No Cypher suggestions yet.

Full Text

33,183 characters extracted from source content.

Expand or collapse full text

AI with Recursive Self-Improvement Workshop @ ICLR 2026 ADAPTIVEDECODINGVIATEST-TIMEPOLICY LEARNING FOR SELF-IMPROVING GENERATION Asmita Bhardwaj ∗ UC San Diego San Diego, CA asbhardwaj@ucsd.edu Yuya Jeremy Ong Plastic Labs San Jose, CA yuya@plasticlabs.ai Eelaaf Zahid IBM Research Cambridge, MA eelaaf.zahid@ibm.com Basel Shbita IBM Research San Jose, CA basel@ibm.com ABSTRACT Decoding strategies largely determine the quality of Large Language Model (LLM) outputs,yet widely used heuristics such as greedy or fixed temperature/top-p decoding are static and often task-agnostic, leading to subop- timal or inconsistent generation quality across domains that demand stylistic or structural flexibility. We introduce a reinforcement learning–based decoder sam- pler that treats decoding as sequential decision-making and learns a lightweight policy to adjust sampling parameters at test-time while keeping LLM weights frozen. We evaluated summarization datasets including BookSum, arXiv, and WikiHow using Granite-3.3-2B and Qwen-2.5-0.5B. Our policy sampler consis- tently outperforms greedy and static baselines, achieving relative gains of up to +88% (BookSum, Granite) and +79% (WikiHow, Qwen). Reward ablations show that overlap-only objectives underperform compared to composite rewards, while structured shaping terms (length, coverage, repetition, completeness) enable sta- ble and sustained improvements. These findings highlight reinforcement learning as a practical mechanism for test-time adaptation in decoding, enabling domain- aware and user-controllable generation without retraining large models. 1INTRODUCTION Large Language Models (LLMs) have become powerful generative systems, capable of producing contextually rich text and achieving impressive performance in a wide range of tasks (Naveed et al., 2025; Yang et al., 2024; Wang et al., 2019; Brown et al., 2020). Text generation involves decoding, the process of selecting output tokens from the model’s predicted probability distribution. The de- coding strategy provides heuristics for token selection and plays a decisive role in shaping the final generated output. Conventional decoding methods such as greedy sampling, top-k sampling (Fan et al., 2018), or nucleus sampling (Holtzman et al., 2019) rely on fixed heuristics, while other meth- ods such as beam search (Lowerre, 1976), contrastive decoding (Li et al., 2023), or Mirostat (Basu et al., 2020) introduce controlled randomness on the same principle. Given a practically infinite search space for possible word sequences, these decoding strategies play a crucial role in the qual- ity, diversity, and robustness of the generated text. However, these strategies do not adapt to the evolving context of generation or the demands of different tasks, limiting the degree of freedom of the model’s capability to generate text with more variations. Often such heuristics are limited by their lack of understanding of the control capabilities and effects of the text generation process, making it unreliable and ad-hoc. As a result, they often underperform when balancing trade-offs such as fluency, diversity, factual accuracy, and stylistic control. Moreover, improvements in decoding behavior typically require retraining or fine-tuning the LLM itself, which is computationally expensive and impractical at scale. From the perspec- tive of recursive self-improvement, decoding offers a particularly attractive intervention point: it allows a model to monitor its own generation behavior, evaluate outcomes, and adjust future actions without modifying internal parameters. Such test-time adaptation enables improvement to occur within a single episode of generation, aligning with emerging views of self-improving systems that operate through feedback, control, and policy updates rather than continual retraining. Despite this ∗ This work was conducted during a Summer Internship at IBM Research. 1 arXiv:2603.18428v1 [cs.CL] 19 Mar 2026 AI with Recursive Self-Improvement Workshop @ ICLR 2026 potential, principled methods for adaptive decoding that are both lightweight and measurable remain underexplored. We propose a learnable reinforcement learning-based decoder sampler, in which an agent learns a policy that dynamically controls and adapts decoding parameters of the LLM at test-time while keeping the LLM weights fixed. In this context, we frame decoding as a sequential decision-making problem: the agent observes the evolving generation state (i.e., prompt, generated text, and model logits) and selects an action (i.e., setting the temperature hyperparameter) which influences the de- coding behavior of the LLM. The reward signals are derived from task-specific metrics to guide the agent toward producing higher-quality output. This approach offers three main advantages. First, it decouples the sampling capabilities of the decoding sampling controls from the LLM training process, which avoids shifting the underlying learned distributions of the LLM. Second, it adapts in real time, allowing flexible control and near-real-time adaption across various tasks and domains. Third, by formulating decoding as a Markov Decision Process (MDP) (Puterman, 2014), it provides a guiding framework for defining states, actions, and rewards that generalize beyond a single task. Viewed through a recursive improvement lens, the proposed decoder forms a closed feedback loop: generation produces intermediate outputs, these outputs are evaluated via reward signals, and the re- sulting feedback informs subsequent decoding decisions within the same generation process. While we focus on lightweight policy learning rather than long-horizon model updates, this formulation captures a core mechanism of self-improving systems under fixed model capacity. In this work, we demonstrate the effectiveness of this method in the context of summarization, a task where decoding needs vary widely between various domains and text styles, as a demonstration of this capability. Experiments with Granite-3.3 (Granite Team, 2024) at 2B and Qwen-2.5 (Yang et al., 2025) at 0.5B demonstrate consistent improvements over static baselines such as greedy decoding. Ablation studies also reveal the contribution of different reward components and their combinations. Our key contributions are threefold: (i) we introduce a reinforcement learning decoder sampler that formalizes decoding as a test-time decision process with explicit state, action, and reward spaces, enabling adaptive decoding without modifying LLM weights; (i) we perform initial evaluations across two models and multiple datasets demonstrating gains over static strategies; and (i) we conduct ablation studies that highlight how different reward signals influence decoding behavior. 2RELATED WORK Numerous decoding strategies have been developed for large language models (LLMs), encompass- ing both deterministic methods (e.g., beam search (Lowerre, 1976), contrastive decoding (Li et al., 2023)) and stochastic approaches (e.g., top-k, top-p, Mirostat (Basu et al., 2020)). These meth- ods primarily rely on heuristic or sampling-based mechanisms to balance fluency, diversity, and alignment with user intent. A comprehensive taxonomy and empirical evaluation of such decoding techniques is presented by Zhang et al. (2024), highlighting both the strengths and limitations of conventional approaches in different tasks and model settings. However, these decoding strategies are typically hand-designed and task-agnostic, making it difficult to adapt generation behavior to specific downstream objectives or contextual preferences. To address these limitations, recent work has explored learnable and controllable decoding strate- gies, where decoding behavior is shaped by trainable components rather than fixed heuristics. Zhang et al. (2022) provide a comprehensive overview of controllable text generation using transformer- based models, highlighting methods that steer generation via prompt design, decoding constraints, or auxiliary models. Hu et al. (2025) further categorize speculative and refinement-based decoding schemes that enhance generation quality or efficiency through multi-stage processes. These works underscore a common theme: most learnable decoding approaches operate by either post-hoc con- trol of token selection or re-ranking candidate generations. In contrast, our method introduces a reinforcement-learned decoder sampler that directly learns a token selection policy over a frozen LLM, enabling decoding decisions to be optimized end-to-end with respect to downstream reward signals. This differs from prior work by tightly coupling decoding behavior with task-specific feed- back, rather than relying on predefined constraints or sampling heuristics. Our method shares conceptual similarities with “Controlled Decoding”, which introduces a learned prefix scorer to guide decoding over a frozen base model Mudgal et al. (2024). However, while their 2 AI with Recursive Self-Improvement Workshop @ ICLR 2026 scorer is trained offline to approximate a static reward function, our approach learns a decoding policy directly through reinforcement learning, enabling it to receive task-specific feedback at gen- eration time. This allows the sampler to optimize decoding behavior end-to-end through interaction with the environment, rather than relying on precomputed supervision or reward proxies. The rein- forcement signal thus serves as a dynamic learning signal, adapting generation to complex objectives that may not be easily captured by hand-crafted scoring functions. By combining the strengths of frozen LLMs with reinforcement-learned decoding policies, our method bridges the gap between fixed decoding heuristics and fully fine-tuned models. 3RL-BASED LEARNABLE DECODER SAMPLER We propose to formulate decoding control as a reinforcement learning (RL) problem, where the goal is to learn an adaptive sampler that dynamically steers the output of a frozen LLM without altering its parameters. Unlike static heuristics such as greedy search, which are hand-designed and context-agnostic, our approach introduces a learnable policy that can adapt decoding behavior to task demands in real time. Figure 1 illustrates this adaptive control loop, in which decoding decisions are continuously informed by feedback from the generated output while the underlying LLM remains fixed. Figure 1: Overview of the RL-based decoder sampler. The agent observes the decoding state from a frozen LLM, selects decoding parameters (e.g., temperature and top-p), and receives rewards guiding adaptive sampling. 3.1DECODING AS A MARKOV DECISION PROCESS We frame the decoding process as a Markov Decision Process (MDP). At each decoding step, the LLM produces a probability distribution over the vocabulary, conditioned on the prompt and gen- erated prefix. The decoder sampler (agent) observes this evolving state and selects an action that determines how the next token is sampled. The episode terminates when an end-of-sequence token is generated or a length limit is reached, yielding a complete feedback cycle over a single generation trajectory. State. The state representation encapsulates information about the decoding context. In our frame- work, this includes the prompt, generated prefix, model logits, and hidden representations. Such representations provide a rich view of both the semantic history and the distributional uncertainty of the model. In practice, we concatenate (i) mean-pooled final hidden states, (i) top-k logits (here k = 50), (i) prefix length, and (iv) normalized entropy of the next-token distribution. These are linearly projected and layer-normalized to match the policy input dimension, enabling the agent to condition decisions on the unfolding generation. Action. The action space corresponds to decoding hyper-parameters that govern token selection. While conventional methods fix these parameters globally (e.g., setting temperature to 0.7), our agent learns to adjust them based off of context. In this work, we focus on a two-dimensional action space controlling temperature and top-p. The policy outputs Gaussian-distributed actions, which are 3 AI with Recursive Self-Improvement Workshop @ ICLR 2026 squashed with a sigmoid and affinely mapped to parameter ranges T ∈ [0.2, 1.2] and p∈ [0.8, 1.0]. While we restrict to these two for tractability, the framework naturally extends to additional knobs (e.g., top-k, length penalties, repetition penalties). Crucially, actions taken by the policy influence not only the immediate next token but also the fu- ture state distribution through the evolving prefix, creating a closed-loop interaction between policy decisions and model behavior. This recursive dependency allows the decoder to refine its strategy within an episode, adapting sampling behavior based on the consequences of earlier decisions. Reward. The reward function provides task-specific supervision that evaluates generated outputs and guides subsequent decoding decisions. A detailed formulation of the reward is provided in Sec- tion 4. In this work, we focus on summarization as the target task, but emphasize that summarization is not monolithic: summarizing a news article, a fictional chapter, or a Reddit post each requires dif- ferent stylistic and structural trade-offs. Static heuristics cannot easily adapt across these domains, whereas our framework is designed to flexibly incorporate reward signals that capture general qual- ities of a good summary (e.g., faithfulness, coverage, fluency). This flexibility allows the same policy to adapt its decoding behavior across diverse summarization styles without hand-tuning for each case. However, the effectiveness of RL training critically depends on how well the reward cap- tures the intended quality criteria; we therefore dedicate ablation studies to analyzing how different reward formulations impact policy learning and final performance (see Section 4). 3.2POLICY LEARNING We employ Proximal Policy Optimization (PPO) to train the decoder sampler. PPO offers a stable on-policy optimization procedure well-suited to low-resource RL training, where sample efficiency and stability are crucial. The policy is parameterized as a lightweight multi-layer perceptron that maps state embeddings through a 2-layer MLP (hidden size 256, GELU activations) to a Gaussian distribution over the 2D action space. This design ensures that the additional learnable component is compact, trainable with modest resources, and decoupled from the frozen LLM backbone. Objective. Formally, the clipped PPO objective is: L clip (θ) =E t h min ρ t (θ)A t , clip ρ t (θ), 1− ε, 1 + ε A t i (1) where ρ t (θ) = π θ (a t |s t ) π θ old (a t |s t ) is the probability ratio and A t is the advantage. We estimate advantages with GAE(γ = 0.99,λ = 0.95) for variance reduction. The full loss adds value and entropy terms: J (θ) =L clip (θ)− c v E t (V θ (s t )− R t ) 2 + c ent E t H(π θ (·| s t )) (2) and in practice we minimize−J (θ). 3.3FRAMEWORK FLEXIBILITY A key advantage of our approach is modularity: the same RL formulation can be instantiated with different state features, action knobs, and reward functions depending on the target application. For example, abstractive summarization may benefit from rewards emphasizing conciseness and factual coverage, whereas story generation could prioritize diversity and creativity. By learning a decoding policy end-to-end, the sampler adapts dynamically to input prompts and domain variation, moving beyond static, one-size-fits-all heuristics. This modularity makes the decoder sampler a natural substrate for recursive self-improvement at test-time, where evaluation signals can be iteratively incorporated without modifying model parameters. In Section 4, we instantiate this framework on summarization tasks using Granite-3.3-2B-Base and Qwen-2.5-0.5B, comparing against conventional baselines and analyzing the impact of different reward designs. 4 AI with Recursive Self-Improvement Workshop @ ICLR 2026 4EVALUATION 4.1EXPERIMENTAL SETUP We evaluate the proposed RL-based decoder sampler on abstractive summarization, a natural testbed that spans diverse domains and styles. This setting allows us to assess whether adaptive sampling improves robustness across tasks where static decoding heuristics are commonly brittle. This eval- uation targets a constrained but measurable form of self-improvement: a lightweight controller that updates its behavior through feedback while the base model remains unchanged. We therefore emphasize (i) within-episode adaptation (state-dependent decoding actions) and (i) across-training improvement (early→late PPO gains) as complementary evidence that the system is not merely a re-parameterized heuristic. Datasets. We use three domains: (1) BookSum (narrative chapters, long-form); (2) arXiv (sci- entific articles/abstracts); and (3) WikiHow (instructional “how-to”). These datasets differ sub- stantially in length, factuality requirements, and stylistic variation, providing a broad stress test for adaptive decoding. Models. We evaluate Granite-3.3 (2B Base) and Qwen-2.5 (0.5B). The LLMs remain frozen throughout; only the lightweight policy network is trained. Training details. Each PPO run uses 100 sampled prompts per dataset, reflecting a low-resource RL setting. For each prompt we generate a candidate summary with the current policy, compute the reward, and update periodically. Unless noted otherwise, prompts are sampled without overlap be- tween training and evaluation splits to avoid overestimating reward-driven adaptation. We report both (i) absolute average rewards and (i) early→late changes to show whether PPO training im- proves policy quality rather than collapsing to a static strategy. We define early→late as the relative change in average reward between the first and last evaluation checkpoints within a run (computed under the same reward variant), which serves as a direct indicator of whether policy learning yields improvement over training time. Hyperparameters follow the defaults in Section 3: PPO with clip- ping parameter ε = 0.2, γ = 0.99, λ = 0.95, value coefficient c v = 0.5, and entropy coefficient c ent = 0.01. Baselines. We compare against two common heuristics: • Greedy decoding: argmax at each step. • Static sampling: fixed temperature of 0.3, held constant across prompts. We choose a conservative static temperature to represent a common “safe” setting for summariza- tion that typically trades diversity for precision; the RL policy is evaluated under the same decoding infrastructure for a controlled comparison. These baselines provide strong static references for as- sessing whether learning a decoding policy yields tangible benefits. Rewards. Our Proposed Reward is defined as a weighted, normalized combination of: • ROUGE-L F1 (weight 0.5): overlap with reference summaries. • Length penalty (weight 0.2): deviation from an ideal source-derived length. • Coverage bonus (weight 0.1, capped): overlap with important source tokens (> 4 chars). • Repetition penalty: applied if more than 30% of words are repeated. • Completeness penalty (−0.05): applied if the summary lacks sentence-ending punctuation. The raw score is linearly normalized to [0, 1] via RAW MIN,RAWMAX. This composite reward is designed to balance multiple, sometimes competing, aspects of summary quality, rather than optimizing a single metric in isolation. Because reinforcement learning is highly sensitive to the shape and scale of the reward signal, small design choices can substantially influence both training stability and final behavior. To better understand these effects, we systematically ablate individual reward components as described next. 5 AI with Recursive Self-Improvement Workshop @ ICLR 2026 We perform several ablations to isolate the impact of reward design: • ROUGE-only: remove all shaping terms (length, coverage, repetition, completeness). • Core shaping: keep ROUGE + length + repetition, drop coverage. • Proposed w/o Coverage: remove the coverage bonus term only. • Proposed w/ Softer Repetition: reduce the weight of the repetition penalty. • Proposed w/ Sigmoid Scaling: replace the linear normalization with a sigmoid squashing. Because reward scales differ, we compare results in relative terms: (i) PPO vs. baselines under the same reward, and (i) PPO early→late change. This protocol isolates the role of evaluation signals in enabling test-time controllability: changing the reward changes what the controller learns to optimize, without altering the base LLM. 4.2RESULTS AND DISCUSSION Training signal (proof of learning). Before comparing against baselines, we first verify that PPO training yields meaningful improvements. Figure 2 shows the reward trajectory for WikiHow + Granite-3.3 under the Proposed Reward. The moving average rises steadily during training, con- firming that the policy learns to improve generation quality rather than collapsing to a static heuristic. This provides empirical evidence that the feedback loop in Figure 1 supports learnable, non-trivial control over decoding behavior. Figure 2: Reward over time for WikiHow + Granite-3.3 under the Proposed Reward. The upward trend in the moving average demonstrates PPO learns a useful sampling policy. Main results (comparison to baselines). Table 1 compares PPO-trained policies against greedy and static baselines across all datasets and models. RL consistently outperforms heuristics, with especially large improvements on BookSum and WikiHow, where stylistic diversity and length vari- ation make fixed parameters brittle. Early→late changes are generally modest but positive, confirm- ing that PPO training improves policy quality rather than collapsing to static strategies. Reward ablations (impact of design). Table 2 focuses on a representative case, arXiv + Granite- 3.3, to isolate the effect of reward design. ROUGE-only yields negligible or even negative gains, confirming that overlap metrics alone are insufficient. Core shaping improves over baselines but shows instability. Removing coverage produces large gains with positive early→late improvements, while sigmoid scaling is modest but stable. Our Proposed Reward balances strong gains with ro- bustness, supporting our claim that structured shaping terms are critical. 6 AI with Recursive Self-Improvement Workshop @ ICLR 2026 Table 1: Main evaluation results with the Proposed Reward. RL achieves consistent improvements over greedy and static baselines. Scores are reported alongside percentage gains and PPO early→late improvements. Granite-3.3-2BQwen-2.5-0.5B WikiHowarXivBookSumWikiHowarXivBookSum Greedy0.3330.2200.1470.1140.1540.014 Static0.3690.2750.2120.1400.1740.083 RL (Proposed)0.390 0.353 0.277 0.204 0.2260.147 ∆ vs Greedy (%)+17.17+60.45+88.44+78.95+46.75+950.00 ∆ vs Static (%)+5.69+28.36+30.66+45.71+29.89+77.11 Early→Late (%)+22.88-4.38+22.06-43.60+54.09+57.14 Table 2: Reward ablations on arXiv (Granite-3.3-2B). We report absolute rewards, percentage im- provements over baselines, and PPO early→late change. Reward VariantPPOGreedyStatic∆ vs Greedy (%) ∆ vs Static (%) Early → Late (%) ROUGE-only0.1690.1580.170+6.96-0.59+15.38 Core shaping0.4480.3350.400+33.73+12.00-4.56 No coverage bonus0.3490.2190.275+59.36+26.91+24.91 Soft repetition penalty0.5030.4310.462+16.72+8.88+6.81 Sigmoid scaling0.3820.3020.323+26.49+18.27-4.27 Proposed (ours)0.3530.2200.275+60.45+28.36-4.38 Takeaways. Our findings yield several insights into the effectiveness of RL-based decoding and the role of reward design: 1. PPO training is stable and effective. The reward-over-time curve (Figure 2) demonstrates that PPO steadily increases reward under the Proposed Reward, confirming the agent learns a non-trivial policy rather than collapsing to a static strategy. 2. RL policies outperform static heuristics. Across datasets and models (Table 1), PPO- trained policies consistently achieve higher rewards than greedy and static decoding. The magnitude of these gains depends on the domain: in stylistically diverse tasks like Book- Sum and WikiHow, PPO delivers substantial improvements (e.g., +88% on BookSum Granite-3.3; +79% on WikiHow Qwen-2.5), showing that adaptive parameter control is particularly valuable where one-size-fits-all heuristics fail. Even where improvements are modest (e.g., arXiv), PPO at least matches static strategies without degradation, suggesting robustness across domains. 3. Reward design is critical. The ablation study (Table 2) reveals that reward shaping strongly influences learning outcomes. ROUGE-only rewards yield minimal gains and sometimes underperform static baselines, confirming that overlap alone is insufficient to capture summary quality. Adding shaping terms improves performance but affects stabil- ity differently: (i) Core shaping boosts absolute reward but exhibits negative early→late trends, indicating instability; (i) Removing coverage produces large, stable gains, sug- gesting coverage is not always beneficial and may introduce noise; (i) Softer repetition penalties dampen improvements, showing that repetition control must be sufficiently strict; 7 AI with Recursive Self-Improvement Workshop @ ICLR 2026 and (iv) Sigmoid scaling smooths rewards but reduces training sensitivity, leading to mod- erate but stable results; Our Proposed Reward achieves a balance of gains and robustness, supporting the claim that well-designed, multi-component rewards are necessary for con- trollable generation. 4. Domain effects. Improvements vary with domain. BookSum and WikiHow, which demand stylistic flexibility and length adaptation, benefit most from adaptive decoding. In contrast, arXiv summaries are more formulaic and yield smaller but consistent gains, showing that RL-based control is most useful when the target style diverges from the assumptions of fixed heuristics. This suggests that adaptive samplers could be especially impactful in open-domain or user-facing applications where prompt style varies widely. 5. Model effects. Even small models like Qwen-2.5-0.5B benefit significantly from adaptive sampling. Relative gains are often larger than for Granite-3.3-2B, demonstrating that RL- based decoding control is not limited to large-scale LLMs and may be even more impactful in constrained settings. This broadens the applicability of our approach to scenarios where fine-tuning is infeasible but decoding-time adaptation can deliver tangible improvements. 6. Latency and deployment overhead. The policy is a 2-layer MLP with hidden size 256, so per-token compute is negligible relative to a full LLM forward pass. The primary state fea- tures are already produced during standard decoding and do not require additional forward passes. For resource-constrained settings, either k in the top-k logit features or hidden-state inputs can be reduced at a potential cost to policy expressiveness. 7. Broader implications. RL-based samplers can reliably improve generation quality without retraining the LLM, and reward design is as important as the RL algorithm itself. Rather than hand-crafting decoding heuristics, lightweight agents can learn domain-specific strate- gies guided by modular reward definitions. We note that this study targets short-horizon improvement in a bounded setting (summarization) rather than long-horizon autonomous self-modification, but it provides a concrete and measurable stepping stone toward loop-based systems. 5CONCLUSION We introduced a reinforcement learning–based decoder sampler that adaptively adjusts decoding parameters of frozen LLMs at inference time. By framing decoding as a sequential decision-making problem, our method learns lightweight policies that improve generation quality without requiring model fine-tuning. Experiments across two models and three datasets show PPO-trained samplers consistently outperforming greedy and static baselines, with reward ablations confirming that struc- tured shaping terms are critical for stable learning. These findings underscore two key themes: reinforcement learning is a practical tool for controllable decoding, and reward design is as important as the RL algorithm itself. Looking ahead, we envision extending this framework to richer action spaces (e.g., top-k, top-p, penalties), exploring preference- based or LLM-as-a-judge reward signals, and applying adaptive decoding to broader generative tasks such as dialogue or story generation. By decoupling decoding control from model parameters, we hope this line of work opens new pathways toward flexible, domain-aware, and user-controllable language generation. All experiments target summarization; transfer to dialogue, code generation, or reasoning, where reward specification is harder, remains untested. The evaluated models (0.5B–2B) represent the lower end of current LLM scale, and out-of-distribution robustness of the learned policy is an open empirical question. We identify both as directions for future work. Since the controller operates at inference time and does not modify model parameters, it avoids per- manent regressions to the base model and enables straightforward rollback by disabling the learned policy. Moreover, policy behavior can be monitored and constrained through explicit reward defini- tions and action bounds, providing a clear surface for auditing and governance. Together, these prop- erties suggest that test-time decoder adaptation offers a lower-risk pathway toward self-improving generation compared to weight-level updates, particularly in settings where stability and reversibility are critical. 8 AI with Recursive Self-Improvement Workshop @ ICLR 2026 REFERENCES Sourya Basu, Govardana Sachitanandam Ramachandran, Nitish Shirish Keskar, and Lav R Varsh- ney. Mirostat: A neural text decoding algorithm that directly controls perplexity. arXiv preprint arXiv:2007.14966, 2020. Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020. Angela Fan, Mike Lewis, and Yann Dauphin. Hierarchical neural story generation. arXiv preprint arXiv:1805.04833, 2018. IBM Granite Team. Granite 3.0 language models. URL: https://github.com/ibm-granite/granite-3.0- language-models, 2024. Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, and Yejin Choi. The curious case of neural text degeneration. arXiv preprint arXiv:1904.09751, 2019. Zhiwei Hu, Tianyi Zhang, Kevin Lin, Maosong Sun, and Xiang Li. Speculative decoding and beyond: An in-depth survey of techniques. arXiv preprint arXiv:2502.19732, 2025. URL https://arxiv.org/abs/2502.19732. Xiang Lisa Li, Ari Holtzman, Daniel Fried, Percy Liang, Jason Eisner, Tatsunori Hashimoto, Luke Zettlemoyer, and Mike Lewis. Contrastive decoding: Open–ended text generation as optimiza- tion. arXiv preprint arXiv:2210.15097, 2023. URL https://arxiv.org/abs/2210. 15097. version 2, revised 10 July 2023. Bruce T Lowerre. The harpy speech recognition system. Carnegie Mellon University, 1976. Shrey Desai Mudgal, Sai Vemprala, Chi Chi, Abhilasha Ravichander, Bryan McCann, and Colin Raffel. Controlled decoding: Prompt engineering for decoding algorithms. In Proceedings of the 41st International Conference on Machine Learning (ICML), volume 235 of Proceedings of Ma- chine Learning Research, p. 24148–24166. PMLR, 2024. URL https://proceedings. mlr.press/v235/mudgal24a.html. Humza Naveed, Asad Ullah Khan, Shi Qiu, Muhammad Saqib, Saeed Anwar, Muhammad Usman, Naveed Akhtar, Nick Barnes, and Ajmal Mian. A comprehensive overview of large language models. ACM Transactions on Intelligent Systems and Technology, 16(5):1–72, 2025. Martin L Puterman. Markov decision processes: discrete stochastic dynamic programming. John Wiley & Sons, 2014. Alex Wang, Yada Pruksachatkun, Nikita Nangia, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel Bowman. Superglue: A stickier benchmark for general-purpose language understanding systems. Advances in neural information processing systems, 32, 2019. Jingfeng Yang, Hongye Jin, Ruixiang Tang, Xiaotian Han, Qizhang Feng, Haoming Jiang, Shaochen Zhong, Bing Yin, and Xia Hu. Harnessing the power of llms in practice: A survey on chatgpt and beyond. ACM Transactions on Knowledge Discovery from Data, 18(6):1–32, 2024. Qwen: An Yang, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chengyuan Li, Dayiheng Liu, Fei Huang, Haoran Wei, et al. Qwen2.5 technical report, 2025. URL https: //arxiv.org/abs/2412.15115. Tianyi Zhang, Yi Sun, Naveen Subramani, Xin Geng, Xisen Li, Maosong Sun, and Xiang Li. A thorough examination of decoding methods in the era of llms. arXiv preprint arXiv:2402.06925, 2024. Yizhong Zhang, Hongyu Sun, Sudha Rao, Samira Shaikh, James Pustejovsky, Yejin Choi, and Noah A. Smith. A survey of controllable text generation using transformer-based pre-trained lan- guage models. arXiv preprint arXiv:2201.05337, 2022. URL https://arxiv.org/abs/ 2201.05337. 9