Mixtral

mistral logo

A set of Mixture of Experts (MoE) model with open weights by Mistral AI in 8x7b and 8x22b parameter sizes. The Mixtral large language models (LLMs) are a set of pretrained generative Sparse Mixture of Experts (SMoE). Sizes: Mixtral 8x22B: To run: ollama run mixtral:8x22b Mixtral 8x22B sets a new benchmark for performance and efficiency … Read more

Qwen2.5

Qwen2.5

Qwen2.5 is the latest generation of the Qwen language models, offering a variety of base and instruction-tuned models with parameter sizes ranging from 0.5 to 72 billion. Key improvements in Qwen2.5 compared to Qwen2 include: Note: All models except the 3B and 72B are released under the Apache 2.0 license, while the 3B and 72B … Read more

Mistral

Mistral 7b

Mistral is a 7B parameter model, distributed with the Apache license. It is available in both instruct (instruction following) and text completion. The Mistral AI team has noted that Mistral 7B: Performance in details We compared Mistral 7B to the Llama 2 family, and re-run all model evaluations ourselves for fair comparison. The performance comparison … Read more

Phi-3.5

Phi-3.5

Phi-3.5-mini is a lightweight, cutting-edge open model based on the same datasets as Phi-3, utilizing synthetic data and carefully curated publicly available websites with an emphasis on high-quality, reasoning-rich information. Part of the Phi-3 model family, it supports a 128K token context length. The model has undergone significant improvements through supervised fine-tuning, proximal policy optimization, … Read more

Qwen2

Qwen2

Qwen2 is trained on data across 29 languages, including English and Chinese. It comes in four different parameter sizes: 0.5B, 1.5B, 7B, and 72B. The 7B and 72B models feature an extended context length of up to 128k tokens. Model Qwen2-0.5B Qwen2-1.5B Qwen2-7B Qwen2-72B Params 0.49B 1.54B 7.07B 72.71B Non-Emb Params 0.35B 1.31B 5.98B 70.21B … Read more

Gemma 2

ollama Gemma 2

Google’s Gemma 2 model comes in three sizes: 2B, 9B, and 27B, featuring a cutting-edge architecture built for exceptional performance and efficiency. With 27 billion parameters, Gemma 2 outperforms models more than twice its size in benchmarks. Its impressive efficiency sets a new benchmark in the open model space. Available in 3 sizes: 2B, 9B, … Read more

mistral-small

mistral-small

Mistral Small is a lightweight model designed for cost-effective use in tasks like translation and summarization. Available under the Mistral Research License, this model offers customers a cost-efficient, fast, and reliable solution for tasks like translation, summarization, and sentiment analysis, without the need for full-scale general-purpose models. With 22 billion parameters, Mistral Small v24.09 serves … Read more

Llama 3.1

Llama 3.1

The Llama 3.1 model family includes the following sizes: Llama 3.1 405B stands out as the first open-access model that matches the top-tier AI models in areas like general knowledge, adaptability, mathematics, tool utilization, and multilingual translation. The upgraded 8B and 70B models are multilingual, feature an extended context length of 128K, and offer advanced … Read more

Ollama UI: A Comprehensive Guide

Ollama UI: A Comprehensive Guide

Ollama UI is a fully-featured front end for large language models (LLMs). It is an open-source platform that allows you to work with both local and open-source models. In this article, I’ll walk you through its features and how to set it up, so let’s dive in. What Is Ollama UI? OpenWebUI for Ollama is … Read more