Running local models on Macs gets faster with Ollama’s MLX support



Ollama, a runtime system for operating large language models on a local computer, has introduced support for Apple’s open source MLX framework for machine learning. Additionally, Ollama says it has improved caching performance and now supports Nvidia’s NVFP4 format for model compression, making for much more efficient memory usage in certain models.

Combined, these developments promise significantly improved performance on Macs with Apple Silicon chips (M1 or later)—and the timing couldn’t be better, as local models are starting to gain steam in ways they haven’t before outside researcher and hobbyist communities.

The recent runaway success of OpenClaw—which raced its way to over 300,000 stars on GitHub, made headlines with experiments like Moltbook and became an obsession in China in particular—has many people experimenting with running models on their machines.

As developers get frustrated with rate limits and the high cost of top-tier subscriptions to tools like Claude Code or ChatGPT Codex, experimentation with local coding models has heated up. (Ollama also expanded Visual Studio Code integration recently.)

The new support is available in preview (in Ollama 0.19) and currently supports only one model—the 35-billion-parameter variant of Alibaba’s Qwen3.5. Hardware requirements are intense by normal users’ standards. Users need an Apple Silicon-equipped Mac, sure, but they also need at least 32GB of RAM, according to Ollama’s announcement.



Source link

  • Related Posts

    Robotaxi Outage in China Leaves Passengers Stranded on Highways

    An unknown technical problem caused a number of robotaxis owned by the Chinese tech giant Baidu to freeze on Tuesday in the middle of traffic, trapping some passengers in the…

    Mercor says it was hit by cyberattack tied to compromise of open-source LiteLLM project

    Mercor, a popular AI recruiting startup, has confirmed a security incident linked to a supply chain attack involving the open-source project LiteLLM. The AI startup told TechCrunch on Tuesday that…

    Leave a Reply

    Your email address will not be published. Required fields are marked *

    You Missed

    What value does the U.S. see in a return to the moon?

    What value does the U.S. see in a return to the moon?

    Robotaxi Outage in China Leaves Passengers Stranded on Highways

    Robotaxi Outage in China Leaves Passengers Stranded on Highways

    Alberta NDP not considering name change — for now…

    Global Beauty Market Grows 10% as AI and E-commerce Reshape Consumer Buying

    Trump’s order to limit mail-in voting: does he have the authority? | US news

    Trump’s order to limit mail-in voting: does he have the authority? | US news

    PSL 2026 – Lahore Qalandars – Fakhar Zaman handed two-match ban for ball tampering

    PSL 2026 – Lahore Qalandars – Fakhar Zaman handed two-match ban for ball tampering