Running local models on Macs gets faster with Ollama’s MLX support 

Apple Silicon Macs get a performance boost thanks to better unified memory usage.  Ollama, a runtime system for operating large language models on a local computer, has introduced support for Apple’s open source MLX framework for machine learning. Additionally, Ollama says it has improved caching performance and now supports…