BotBeat
...
← Back

> ▌

Qwen (Alibaba)Qwen (Alibaba)
RESEARCHQwen (Alibaba)2026-03-22

Flash-MoE: Researchers Run 397B Parameter Model on MacBook Pro with 48GB RAM at 4.4 Tokens/Second

Key Takeaways

  • ▸A 397B parameter Mixture-of-Experts model now runs on consumer MacBook Pro hardware (48GB RAM) at 4.4+ tokens/second, making large models accessible without specialized infrastructure
  • ▸The implementation uses custom C/Metal code with no Python frameworks, streaming 209GB of model weights from SSD on-demand and loading only active experts per layer to achieve efficient resource utilization
  • ▸Multiple optimization techniques including 4-bit quantization, FMA-optimized dequantization, hand-tuned GPU kernels, and deferred compute pipelines demonstrate how understanding hardware constraints enables substantial performance gains
Source:
Hacker Newshttps://github.com/danveloper/flash-moe↗

Summary

Researchers have demonstrated a breakthrough in efficient inference by running Qwen3.5-397B-A17B, a 397 billion parameter Mixture-of-Experts model, on a MacBook Pro with just 48GB of RAM at 4.4+ tokens per second with production-quality output. The achievement uses a pure C/Metal inference engine that streams the entire 209GB model from SSD through custom Metal compute pipelines, requiring no Python or standard frameworks. The system leverages Apple Silicon's unified memory architecture and implements several optimization techniques including 4-bit quantization, FMA-optimized dequantization kernels, deferred GPU expert compute, and intelligent expert streaming that loads only the K=4 active experts per layer on demand.

The technical implementation combines CPU and GPU workloads in a carefully orchestrated pipeline, with GPU compute, SSD I/O, and memory operations serialized due to Apple Silicon's shared memory controller constraints. Key optimizations include hand-tuned Metal shaders for matrix operations, fused kernels for activation and normalization, and leveraging the OS page cache for expert weight caching rather than implementing custom caching solutions. The project was completed in 24 hours and demonstrates that large-scale language models can run efficiently on consumer hardware through careful architectural design and low-level hardware optimization.

  • The project validates 'trust the OS' principles by using OS-level page caching instead of custom caching mechanisms, achieving ~71% cache hit rates naturally while reducing complexity

Editorial Opinion

Flash-MoE represents a significant shift in how the AI community thinks about model deployment, proving that billion-parameter models are no longer confined to data centers with specialized GPUs. The achievement is impressive not just for its performance metrics, but for demonstrating that careful hardware-aware engineering and algorithmic optimization can make cutting-edge models practical on commodity consumer devices. However, the reliance on 4-bit quantization for production use (as 2-bit breaks tool calling) and the MacBook-specific optimizations suggest there are still meaningful trade-offs between portability and capability that practitioners must navigate.

Large Language Models (LLMs)Generative AIMLOps & InfrastructureAI HardwareOpen Source

Comments

Suggested

AnthropicAnthropic
RESEARCH

Inside Claude Code's Dynamic System Prompt Architecture: Anthropic's Complex Context Engineering Revealed

2026-04-05
Google / AlphabetGoogle / Alphabet
RESEARCH

Deep Dive: Optimizing Sharded Matrix Multiplication on TPU with Pallas

2026-04-05
GitHubGitHub
PRODUCT LAUNCH

GitHub Launches Squad: Open Source Multi-Agent AI Framework to Simplify Complex Workflows

2026-04-05
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us