Moneycontrol
HomeTechnologyApple’s new language model can write long texts at lightning speed

Apple’s new language model can write long texts at lightning speed

Apple researchers, in collaboration with Ohio State University, have unveiled a breakthrough language model that can generate long-form text up to 128 times faster than existing systems.

October 15, 2025 / 17:56 IST
Story continues below Advertisement
Apple

Apple’s latest study introduces a fresh take on how generative AI can process and produce language. Traditionally, large language models (LLMs) like ChatGPT rely on an autoregressive approach — they generate content one token at a time, using both the input prompt and the sequence of previously produced words to decide what comes next. While this method ensures coherence, it’s inherently slow because it processes text sequentially.

In contrast, diffusion models take a parallel approach. Instead of predicting one word at a time, they generate multiple tokens simultaneously and then refine them across several steps. This iterative process gradually transforms random noise into coherent text. Building upon that, flow-matching models, a subset of diffusion models, simplify this process even further by learning to generate the final output in one smooth step, eliminating the need for thousands of refinement iterations.

Story continues below Advertisement

Apple’s new approach, FS-DFM, is designed to combine the best of both worlds: high output quality with significantly faster generation speed. In their paper titled “FS-DFM: Fast and Accurate Long Text Generation with Few-Step Diffusion Language Models,” the researchers demonstrate that FS-DFM can create complete passages in just eight refinement rounds while matching the quality of diffusion models that typically require over a thousand steps.

To achieve this, the researchers implemented a three-stage training process. First, they trained FS-DFM to handle different levels of refinement efficiently. Then, a “teacher” model was used to guide the process, helping the system make more precise and stable updates with each iteration. Finally, they fine-tuned how each iteration behaves to reduce overshooting and accelerate convergence toward the final text.