Sink-Aware Pruning for Diffusion Language Models
Diffusion Language Models (DLMs) incur high inference cost due to iterative denoising, motivating efficient pruning.
Academic or research source. Check the methodology, sample size, and whether it's been replicated.
Diffusion Language Models (DLMs) incur high inference cost due to iterative denoising, motivating efficient pruning.
TLDR
Diffusion Language Models (DLMs) incur high inference cost due to iterative denoising, motivating efficient pruning.