<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0"><channel><title><![CDATA[🛠️ Ollama: Native MLX Backend for Apple Silicon]]></title><description><![CDATA[<p><img src="https://board.circlewithadot.net/assets/plugins/nodebb-plugin-emoji/emoji/android/1f6e0.png?v=28325c671da" class="not-responsive emoji emoji-android emoji--hammer_and_wrench" style="height:23px;width:auto;vertical-align:middle" title="🛠" alt="🛠" />️ Ollama: Native MLX Backend for Apple Silicon</p><p>Ollama now runs on Apple MLX natively. On M5 Max + Qwen3.5-35B-A3B: 1851 tok/s prefill, 134 tok/s decode. Also adds NVFP4 quantization for production parity with NVIDIA inference and improved KV cache reuse for agentic workloads.</p><p>solomonneas.dev/intel</p><p><a href="https://infosec.exchange/tags/Ollama" rel="tag">#<span>Ollama</span></a> <a href="https://infosec.exchange/tags/LLM" rel="tag">#<span>LLM</span></a> <a href="https://infosec.exchange/tags/AppleSilicon" rel="tag">#<span>AppleSilicon</span></a> <a href="https://infosec.exchange/tags/DevTools" rel="tag">#<span>DevTools</span></a></p>]]></description><link>https://board.circlewithadot.net/topic/d7799564-ef93-4066-b0cd-fa6f2c1c4f0e/ollama-native-mlx-backend-for-apple-silicon</link><generator>RSS for Node</generator><lastBuildDate>Mon, 06 Apr 2026 09:23:15 GMT</lastBuildDate><atom:link href="https://board.circlewithadot.net/topic/d7799564-ef93-4066-b0cd-fa6f2c1c4f0e.rss" rel="self" type="application/rss+xml"/><pubDate>Wed, 01 Apr 2026 16:03:46 GMT</pubDate><ttl>60</ttl></channel></rss>