Running Qwen 397B Locally via Apple Flash Techniques
- •Qwen 397B model runs at 5.5 tokens/second on 48GB MacBook Pro.
- •Technique utilizes SSD weight streaming based on Apple's 'LLM in a Flash' research.
- •AI-driven 'autoresearch' optimized the code through 90 automated experimental cycles.
Running a massive artificial intelligence model—one that usually requires a room full of expensive servers—on a standard high-end laptop is now a reality. Researcher Dan Woods recently achieved this feat by deploying the Qwen3.5-397B-A17B model on a MacBook Pro with only 48GB of memory, even though the model itself occupies over 200GB on disk.
The breakthrough relies on a technique called 'LLM in a Flash,' originally proposed by Apple researchers to solve memory limitations. Most AI models need to keep all their 'weights'—the numerical values that determine how the AI processes information—inside the computer's fast-access memory (RAM). However, this specific model uses a Mixture-of-Experts (MoE) architecture. In an MoE setup, the AI only activates a small fraction of its total brainpower for any given token or piece of text.
By keeping the most essential parts in RAM and streaming the specialized 'expert' weights from the laptop's slower storage (SSD) only when needed, the system maintained a speed of 5.5 tokens per second. To refine this complex process, Woods utilized an 'autoresearch' method where an AI coding tool ran 90 different experiments to find the most efficient code automatically.
While the model's output quality at such high compression levels is still being evaluated, the experiment marks a significant shift in AI accessibility. It suggests a future where the most powerful models can run locally on personal hardware, offering users better privacy and lower costs without relying on massive cloud data centers.