📦 Ollama – v0.11.2: ggml: Prevent kv cache quanitization on gpt-oss
📢 Hey AI Developers!
Ollama v0.11.2 is out now! This release includes a fix to prevent incorrect KV cache quantization when using the `gpt-oss` model. Get the update and keep those LLMs running smoothly!