Apple silicon M5, faster on device AI with MLX

Nov 20, 2025

Apple silicon M5, faster on device AI with MLX
Apple silicon M5, faster on device AI with MLX

Apple silicon M5, faster on device AI with MLX

The new M5 MacBook Pro is built for people who care about speed on their own machine. Apple’s latest chips add Neural Accelerators in the GPU and more memory bandwidth, so modern AI models can run faster, directly on your Mac.

Apple’s MLX framework sits on top of that hardware. It is the piece that makes it easy to run and experiment with large language models on Apple silicon, without needing to send data to a remote server.

Fenn uses MLX under the hood so it can take advantage of Apple silicon, including the new M5 chips, while keeping your file search private and on device.

MLX in simple terms

MLX is Apple’s open source array framework designed for Apple silicon.

In plain words:

  • It understands how to use the CPU, GPU, and now the Neural Accelerators in the M5.

  • It lets developers run and tune language models and other AI models directly on a Mac.

  • It takes advantage of unified memory, so data does not need to be copied back and forth between components.

You do not see MLX in your dock, but if you use an app that is built on it, you benefit from the speed and efficiency of Apple silicon without thinking about tensor shapes or kernels.

What M5 changes for on device AI

In the M5 MacBook Pro, Apple added Neural Accelerators in the GPU that are dedicated to matrix math. This is the core operation behind many AI models, including large language models.

With the latest macOS release, MLX can use those Neural Accelerators when you are on M5, which means:

  • Faster time to first response when a model runs on your Mac

  • Better performance on larger models, including quantized versions

  • Improved throughput for workloads that fit in unified memory

Apple’s own benchmarks show that, for a range of language models, M5 cuts the time to first token by several times compared to M4 in many cases, and also improves generation speed, especially when the model fits well in memory.

Source: https://machinelearning.apple.com/research/exploring-llms-mlx-m5

You do not have to understand the details to feel the effect. On an M5 Mac, AI workloads that once felt heavy become much more responsive.

How this helps Fenn users

Fenn is a file search engine for macOS that understands your content, PDFs, Office files, images with text, audio, video, and Apple Mail. It runs on device by default and uses MLX under the hood to take advantage of Apple silicon.

On an M5 Mac, this matters in two places:

  1. Indexing and understanding your files

  2. Agent Mode and heavier AI style queries

Faster indexing on powerful Macs

When you first point Fenn at your folders, it needs to read and understand your documents, images, audio, and video. On an M5:

  • Large libraries of PDFs and Office files are processed faster

  • Text is extracted and embedded more quickly

  • Image and screenshot understanding benefits from the improved hardware

The result is less time waiting for initial indexing and faster updates when you add new content.

Faster Agent Mode and semantic search

Agent Mode and semantic search rely on model based understanding. On an M5 Mac using MLX:

  • Complex questions that look across many files feel more responsive

  • Larger context windows and heavier prompts are more practical

  • Multi step queries (search plus lightweight analysis) complete faster

For example, queries like:

  • “All contracts that mention termination within 30 days, show the pages”

  • “Invoices from vendor X above 500 dollars in 2024, grouped by month”

  • “Spreads that show the teal gradient and mention Summer Lookbook”

become easier to run frequently, because the model work behind them is better served by the M5 hardware.

Agent Mode works best on higher memory Macs, especially when you keep many large files indexed. It also runs on 16 GB. If you want help tuning settings on your Mac, contact us.

What this looks like in daily work

Creatives with heavy assets

If you keep thousands of Photoshop, Illustrator, and InDesign exports, plus RAW images and reference boards:

  • Indexing completes faster on your M5 Mac

  • Searching for “hero image with teal gradient and tall headline” returns results quickly

  • Agent Mode can scan a whole project folder to list spreads that match a visual and text brief

Lawyers and contract heavy teams

If you manage folders full of contracts, scanned PDFs, and email threads:

  • Clause level search feels snappier, even across many PDFs

  • Agent Mode can triage contracts by termination, governing law, or liability caps more comfortably

  • Searching email plus contracts for a specific promise feels closer to realtime

Finance and ops

If your world is invoices, statements, models, and closing binders:

  • Large close folders are indexed more quickly

  • Queries like “invoices from vendor X above 500 dollars in 2024” or “documents that mention DSCR covenant” return faster

  • You can run richer Agent Mode questions during audits and reviews without slowing your Mac to a crawl

Researchers and product teams

If you live in dense PDFs, notes, and recordings:

  • Long reports index faster, including pages and figures

  • Model based search over notes and documents feels more interactive

  • Agent Mode can scan more material to summarize themes or surface key pages

Getting the most from an M5 Mac with Fenn

You do not need to change much to benefit from M5 and MLX, but a few habits help:

  1. Keep macOS up to date
    The Neural Accelerators in M5 are used by MLX on newer macOS releases, so updating keeps performance improvements flowing.

  2. Index the folders that matter most
    Start with contracts, finance, creative projects, and research folders, rather than your entire disk. You get the biggest gains where you work every day.

  3. Use Semantic and Agent Mode for heavy questions
    Simple filename lookups can stay in Keyword or Exact. Save Semantic and Agent Mode for questions that really benefit from AI understanding.

  4. Keep your Mac plugged in for big indexing jobs
    When you point Fenn at a huge archive, plug in and let M5 do the work while you continue using your Mac.

Why powerful Mac users should care

If you invested in an M5 MacBook Pro or another modern Apple silicon machine, you already have a strong on device AI platform. MLX is how Apple exposes that power to developers. Fenn is how you turn it into something practical, private, and concrete in your day to day work.

  • You keep your data on your own Mac

  • You get faster indexing and faster answers

  • You can ask richer questions about your files, not just the web

Download Fenn for Mac. Private on device. Find the moment, not the file.

See also