ICLR 2026 Orals

$p\textrm{-less}$ Sampling: A Robust Hyperparameter-Free Approach for LLM Decoding

Runyan Tan, Shuang Wu, Phillip Howard

LLMs & Reasoning Sat, Apr 25 · 11:30 AM–11:40 AM · Amphitheater Avg rating: 6.00 (6–6)
Author-provided TL;DR

P-less Sampling: A parameterless sampling strategy grounded in information theory, where the truncation threshold adapts to the entire token probability distribution, is bounded and valid, and dynamically adjusts with temperature.

Abstract

Obtaining high-quality outputs from Large Language Models (LLMs) often depends upon the choice of a sampling-based decoding strategy to probabilistically choose the next token at each generation step. While a variety of such sampling methods have been proposed, their performance can be sensitive to the selection of hyperparameters which may require different settings depending upon the generation task and temperature configuration. In this work, we introduce $p\textrm{-less}$ sampling: an information-theoretic approach to sampling which dynamically sets a truncation threshold at each decoding step based on the entire token probability distribution. Unlike existing methods, $p\textrm{-less}$ sampling has no hyperparameters and consistently produces high-quality outputs as temperature increases. We provide theoretical perspectives on $p$-less sampling to ground our proposed method and conduct experiments to empirically validate its effectiveness across a range of math, logical reasoning, and creative writing tasks. Our results demonstrate how $p\textrm{-less}$ sampling consistently outperforms existing sampling approaches while exhibiting much less degradation in text quality at higher temperature values. We further show how $p$-less achieves greater inference-time efficiency than alternative methods through lower average token sampling times and shorter generation lengths, without sacrificing accuracy. Finally, we provide analyses to highlight the benefits of $p\textrm{-less}$ through qualitative examples, case studies, and diversity assessments.

One-sentence summary·Auto-generated by claude-haiku-4-5-20251001(?)

p-less sampling dynamically sets truncation threshold using information theory for hyperparameter-free LLM decoding with robust quality at high temperatures.

Contributions·Auto-generated by claude-haiku-4-5-20251001(?)
  • Information-theoretic approach to sampling with no hyperparameters
  • Dynamically sets truncation threshold based on entire token probability distribution
  • Consistently produces high-quality outputs across varying temperatures
Methods used·Auto-generated by claude-haiku-4-5-20251001(?)
  • Sampling methods
  • Information theory
  • LLM decoding
  • Token selection
Limitations (author-stated)·Auto-generated by claude-haiku-4-5-20251001(?)

Authors did not state explicit limitations.

Future work (author-stated)·Auto-generated by claude-haiku-4-5-20251001(?)

Authors did not state explicit future directions.

Author keywords

  • LLM
  • decoding
  • sampling
  • truncation
  • inference
  • information-theoretic
  • information-theory
  • hyperparameterless
  • hyperparameter-free
  • entropy
  • entropy-aware
  • distribution-aware
  • adaptive
  • efficient
  • generation

Related orals

Something off? Let us know →