Ollama 0.3.11 is here! Update to run more models, and lots more quality-of-life changes.
New models
- Solar-Pro-Preview: an advanced large language model (LLM) with 22 billion parameters designed to fit into a single GPU.
- Qwen 2.5: new multilingual Qwen models pretrained on Alibaba’s latest large-scale dataset, encompassing up to 18 trillion tokens with support for a context window of up to 128K tokens.
- Bespoke-Minicheck: a state-of-the-art fact-checking model developed by Bespoke Labs.
- Mistral-Small: a lightweight 22B model designed for cost-effective use in tasks like translation and summarization.
- Reader-LM: A series of models that convert HTML content to Markdown content, which is useful for content conversion tasks.
What’s Changed
- New
ollama stop
command to unload a running model - Ollama will now show an error when importing a model with an invalid number of tokens in the vocabulary
- The
ollama/ollama
container image will now start running almost immediately, leading to 5s faster start times - Fixed issue where
ollama show
would show excessive whitespace in the output