Beyond Fixed: Variable-Length Denoising for Diffusion Large Language Models
Jinsong Li, Xiaoyi Dong, Yuhang Zang, Yuhang Cao, Jiaqi Wang, Dahua Lin
2025-08-04
Summary
This paper talks about DAEDAL, a new strategy that helps Diffusion Large Language Models (LLMs) handle text of different lengths more efficiently without needing extra training.
What's the problem?
The problem is that Diffusion LLMs usually generate text with a fixed length, which limits their ability to handle varied or longer texts and reduces their performance and efficiency.
What's the solution?
DAEDAL introduces a training-free method that allows Diffusion LLMs to adapt dynamically to different input and output lengths during generation by using a special denoising process that adjusts token lengths on the fly.
Why it matters?
This matters because it makes Diffusion LLMs more flexible and practical, improving their ability to generate natural and varied text while reducing the computing cost, which is useful for many language-related applications.
Abstract
DAEDAL, a novel training-free denoising strategy, enables dynamic length adaptation in Diffusion Large Language Models, improving performance and computational efficiency.