MacBook running local AI inference with performance graph showing speed boost
Technology & People

DFlash + MLX Delivers 4x LLM Speed Boost on Apple Silicon — Zero Accuracy Loss

A new block diffusion technique combined with Apple's MLX framework lets M-series Macs run local LLMs four times faster, with no quality tradeoff.

local AIApple SiliconMLXDFlashLLM inference

Running a large language model locally on your Mac used to mean choosing between speed and quality. DFlash changes that equation — and it does it without cheating.

What Is DFlash?

DFlash is a speculative decoding technique that uses a tiny block diffusion model to draft tokens in parallel, then has the main LLM verify them in a single forward pass. Think of it like having a fast junior writer produce a rough draft that a senior editor approves in one read-through — except the junior writer is a lightweight diffusion model and the editor is your full-size LLM.

The result: 4x faster inference on Apple Silicon with zero accuracy degradation. The output is bit-for-bit identical to running the model without DFlash. No shortcuts, no approximation.

Why MLX Matters

Apple’s MLX framework is the engine that makes this work on Mac hardware. MLX is designed specifically for Apple Silicon — it handles the unified memory architecture, the GPU/CPU switching, and the Metal acceleration that M-series chips are built for.

DFlash now has native MLX support, meaning you don’t need CUDA, you don’t need a data centre GPU, and you don’t need to cloud-anything. It runs on the Mac you already own.

The implementation is open-source and available on GitHub at z-lab/dflash.

Why This Matters

Three reasons this is more than a benchmark result:

  1. On-device AI becomes practical. Four times faster means the difference between a sluggish local model and one that feels responsive enough for daily use. That changes the calculus on whether you need cloud AI at all.

  2. Privacy by default. When local inference is fast enough, there’s no reason to send your data to someone else’s server. Your prompts, your documents, your conversations — they stay on your machine.

  3. The Mac as AI workstation trend accelerates. M-series Macs already had the unified memory advantage. Now they have the inference speed to match. This is another data point showing consumer hardware catching up to cloud infrastructure for inference workloads.

The Bigger Picture

We’ve been tracking the shift toward local AI for a while. Open-source models keep getting smaller and smarter. Hardware keeps getting more capable. And now optimisation layers like DFlash are closing the gap from the other direction — not by making models smaller, but by making them run smarter.

The combination of MLX + DFlash is specifically significant because it targets the hardware that millions of people already own. You don’t need to buy anything. You just need to run it.

For Singularity.Kiwi readers already running local models through tools like Ollama or OpenClaw, this is worth watching. As DFlash support integrates into more inference engines, that 4x speedup could become the new baseline.

Sources

Sources: z-lab/dflash GitHub, @outsource_ on X