MARS Method Accelerates Autoregressive Model Generation Speeds
- •MARS enables autoregressive models to predict multiple tokens simultaneously without architectural changes.
- •Method achieves 1.5-1.7x throughput improvements while maintaining baseline-level prediction accuracy.
- •Real-time speed control allows dynamic adjustment via confidence thresholds during inference.
The challenge of generating text with autoregressive language models—those that predict the next word in a sequence one step at a time—has long been a bottleneck for deployment. While these models are powerful, their 'one-token-at-a-time' nature limits their speed, creating latency issues when users expect instantaneous responses. Researchers from Nanyang Technological University have introduced MARS (Mask AutoRegression), a new fine-tuning method designed to bypass this limitation by training models to predict multiple tokens within a single computational pass.
What sets MARS apart from previous efforts like speculative decoding or multi-head architectures is its simplicity. Unlike existing methods that often require separate draft models or additional architectural layers, MARS functions as a lightweight fine-tuning step. It requires no changes to the underlying model structure and introduces no extra parameters, making it highly compatible with existing checkpoints. By reusing instruction-tuning data, the researchers have managed to teach models to predict several tokens at once, essentially 'batching' the prediction process while keeping the accuracy of the original model intact.
The performance gains are notable, with the system delivering a 1.5x to 1.7x increase in throughput on standard tasks. Furthermore, the team implemented a block-level KV caching strategy, which allows for even faster processing in batch inference scenarios. This approach, tested on models like Qwen2.5-7B, proves that significant efficiency gains are possible through algorithmic cleverness rather than just throwing more hardware at the problem.
One of the most practical features of MARS is its support for real-time speed adjustment. Using a confidence thresholding mechanism, a serving system can decide on the fly whether to output single tokens or multiple tokens based on current request loads. This acts as a 'latency-quality knob,' allowing system administrators to prioritize speed during peak demand without the need to restart the model or swap it out for a different version. This flexibility represents a significant step forward for developers looking to balance user experience with computational overhead.