info@news-matic.com

details

Running local models on Macs gets faster with Ollama's MLX support

Apple Silicon Macs get a performance boost thanks to better unified memory usage. Ollama, a runtime system for operating large language models on a local computer, has introduced support for Apple’s open source MLX framework for machine learning. Additionally, Ollama says it has improved caching performance and now supports Nvidia... [2442 chars]

Cookie Consent + Tracking