aya-expanse

aya-expanse

Aya Expanse marks a major breakthrough in multilingual AI capabilities. It combines Cohere’s Command model family with a year of dedicated research in multilingual optimization, resulting in powerful 8B and 32B parameter models that excel at understanding and generating text across 23 languages, while delivering high performance across the board. Run aya-expanse with Ollama: Key … Read more

Swift Client

Ollama Swift Client

Ollama Swift Client is a library for interacting with the Ollama API. Requirements Installation Swift Package Manager Add the following to your Package.swift file: .package(url: “https://github.com/mattt/ollama-swift.git”, from: “1.0.0”) Usage The tests and example code for this library use the llama3.2 model. Run the following command to download the model to run them yourself: Initializing the client import Ollama // Use … Read more

Nemotron

nvidia banner

Llama-3.1-Nemotron-70B-Instruct is a large language model customized by NVIDIA to improve the helpfulness of LLM generated responses to user queries. Llama-3.1-Nemotron-70B-Instruct is a large language model customized by NVIDIA to improve the helpfulness of LLM generated responses to user queries. This model was trained using RLHF (specifically, REINFORCE), Llama-3.1-Nemotron-70B-Reward and HelpSteer2-Preference prompts on a Llama-3.1-70B-Instruct model as the initial policy. … Read more

Use Ollama on Hugging Face

Use Ollama on Hugging Face

Ollama is an application built on llama.cpp, enabling direct interaction with LLMs through your computer. It supports any GGUF quant models created by the community, such as those from bartowski, MaziyarPanahi, and many others, available on Hugging Face. With Ollama, there’s no need to create a new Modelfile—you can use any of the 45,000+ public … Read more

Ollama 0.3.13 is here

Ollama 0.3.13 is here

Ollama 0.3.13 is here! There are new safety models available! (Meta’s Llama Guard 3 and Google’s ShieldGemma). For contributors, devs have been working on a new Go runner for improved reliability and caching. Please give it a try! New safety models What’s Changed New Go runner (for contributors): Ollama is transitioning to a new Go-based … Read more

Ollama OpenAI Compatibility

Ollama OpenAI Compatibility

Ollama now offers built-in support for the OpenAI Chat Completions API, enabling seamless integration with more tools and applications locally. Setup Start by downloading Ollama and pulling a model, such as Llama 2 or Mistral: Usage Using cURL To interact with Ollama’s API, which is compatible with OpenAI’s format, adjust the hostname to http://localhost:11434: Using … Read more

Mistral-nemo

mistral logo

A cutting-edge 12B model featuring a 128k context length, developed by Mistral AI in partnership with NVIDIA. Mistral NeMo is a 12B model built in collaboration with NVIDIA. Mistral NeMo offers a large context window of up to 128k tokens. Its reasoning, world knowledge, and coding accuracy are state-of-the-art in its size category. As it … Read more

Llama 3.2

Llama 3.2

Llama 3.2 is available on Ollama! It’s lightweight and multimodal! It’s so fast and good! Meta’s Llama 3.2 goes small with 1B and 3B models. The Meta Llama 3.2 series of multilingual large language models (LLMs) consists of pretrained and instruction-tuned generative models available in 1B and 3B parameter configurations (text in/text out). Optimized for … Read more

Codestral

Codestral

Codestral is Mistral AI‘s inaugural code model, built specifically for code generation tasks, featuring a 22-billion parameter architecture. Proficient in over 80 programming languages. Codestral has been trained on a diverse dataset encompassing more than 80 languages, such as Python, Java, C, C++, JavaScript, Swift, Fortran, and Bash. It can complete functions, generate tests, and … Read more

LLava

LLava

LLaVA is an innovative end-to-end large multimodal model, integrating a vision encoder with Vicuna to enable comprehensive visual and language understanding. It has been updated to version 1.6. Run LLava LLaVA, which stands for Large Language and Vision Assistant, is a multimodal model that combines both visual and language processing abilities. It integrates a vision … Read more