Ollama v0.3.11 Release

Ollama 0.3.11 is here! Update to run more models, and lots more quality-of-life changes.

New models

  • Solar-Pro-Preview: an advanced large language model (LLM) with 22 billion parameters designed to fit into a single GPU.
  • Qwen 2.5: new multilingual Qwen models pretrained on Alibaba’s latest large-scale dataset, encompassing up to 18 trillion tokens with support for a context window of up to 128K tokens.
  • Bespoke-Minicheck: a state-of-the-art fact-checking model developed by Bespoke Labs.
  • Mistral-Small: a lightweight 22B model designed for cost-effective use in tasks like translation and summarization.
  • Reader-LM: A series of models that convert HTML content to Markdown content, which is useful for content conversion tasks.

What’s Changed

  • New ollama stop command to unload a running model
  • Ollama will now show an error when importing a model with an invalid number of tokens in the vocabulary
  • The ollama/ollama container image will now start running almost immediately, leading to 5s faster start times
  • Fixed issue where ollama show would show excessive whitespace in the output