This version of Ollama will change the architecture to directly support llama.cpp instead of building on top of GGML, and allows for compatibility with GGUF file format. MLX is used to accelerate model inference on Apple Silicon.
While in pre-release we'd love feedback on:
- Performance improvements or degradation
- Errors or crashes that did not previously occur
- Memory utilization improvements or degradation
Known issues
laguna-xs.2is not supported yet on this pre-releasellama3.2-visionis not supported yet on this pre-release
Installing
Mac/Linux
curl -fsSL https://ollama.com/install.sh | OLLAMA_VERSION=0.30.0-rc15 sh


