Self-Speculative Decoding Speeds for Multi-Token LLMs Post date June 6, 2025 Post author By Large Models (dot tech) Post categories In ai-efficiency, code-generation, inference-optimization, llm-decoding-speed, llm-inference, multi-token-models, multi-token-prediction, self-speculative-decoding