Ollama 0.19 has pushed out a preview update that leans on Apple’s MLX framework to make local AI models run faster on Macs with Apple silicon. The headline numbers are hard to ignore: prompts are processed about 1.6 times faster, and response generation is nearly twice as quick. The biggest gains are on Macs with […]

