Ollama Blog https://ollama.com/blog/feed_ollama.xml Latest updates from Ollama http://www.rssboard.org/rss-specification python-feedgen https://ollama.com/public/icon-64x64.png Ollama Blog https://ollama.com/blog/feed_ollama.xml en Tue, 13 Jan 2026 22:03:16 +0000 Run Llama 2 uncensored locally https://ollama.com/blog/run-llama2-uncensored-locally This post will give some example comparisons running Llama 2 uncensored model versus its censored model. https://ollama.com/blog/run-llama2-uncensored-locally Tue, 01 Aug 2023 00:00:00 +0000 Run Code Llama locally https://ollama.com/blog/run-code-llama-locally Meta's Code Llama is now available on Ollama to try. https://ollama.com/blog/run-code-llama-locally Thu, 24 Aug 2023 00:00:00 +0000 How to prompt Code Llama https://ollama.com/blog/how-to-prompt-code-llama This guide walks through the different ways to structure prompts for Code Llama and its different variations and features including instructions, code completion and fill-in-the-middle (FIM). https://ollama.com/blog/how-to-prompt-code-llama Sat, 09 Sep 2023 00:00:00 +0000 Leveraging LLMs in your Obsidian Notes https://ollama.com/blog/llms-in-obsidian This post walks through how you could incorporate a local LLM using Ollama in Obsidian, or potentially any note taking tool. https://ollama.com/blog/llms-in-obsidian Thu, 21 Sep 2023 00:00:00 +0000 Ollama is now available as an official Docker image https://ollama.com/blog/ollama-is-now-available-as-an-official-docker-image Ollama can now run with Docker Desktop on the Mac, and run inside Docker containers with GPU acceleration on Linux. https://ollama.com/blog/ollama-is-now-available-as-an-official-docker-image Thu, 05 Oct 2023 00:00:00 +0000 Building LLM-Powered Web Apps with Client-Side Technology https://ollama.com/blog/building-llm-powered-web-apps Recreate one of the most popular LangChain use-cases with open source, locally running software - a chain that performs Retrieval-Augmented Generation, or RAG for short, and allows you to “chat with your documents” https://ollama.com/blog/building-llm-powered-web-apps Fri, 13 Oct 2023 00:00:00 +0000 Python & JavaScript Libraries https://ollama.com/blog/python-javascript-libraries The initial versions of the Ollama Python and JavaScript libraries are now available, making it easy to integrate your Python or JavaScript, or Typescript app with Ollama in a few lines of code. Both libraries include all the features of the Ollama REST API, are familiar in design, and compatible with new and previous versions of Ollama. https://ollama.com/blog/python-javascript-libraries Tue, 23 Jan 2024 00:00:00 +0000 Vision models https://ollama.com/blog/vision-models New vision models are now available: LLaVA 1.6, in 7B, 13B and 34B parameter sizes. These models support higher resolution images, improved text recognition and logical reasoning. https://ollama.com/blog/vision-models Fri, 02 Feb 2024 00:00:00 +0000 OpenAI compatibility https://ollama.com/blog/openai-compatibility Ollama now has initial compatibility with the OpenAI Chat Completions API, making it possible to use existing tooling built for OpenAI with local models via Ollama. https://ollama.com/blog/openai-compatibility Thu, 08 Feb 2024 00:00:00 +0000 Windows preview https://ollama.com/blog/windows-preview Ollama is now available on Windows in preview, making it possible to pull, run and create large language models in a new native Windows experience. Ollama on Windows includes built-in GPU acceleration, access to the full model library, and serves the Ollama API including OpenAI compatibility. https://ollama.com/blog/windows-preview Thu, 15 Feb 2024 00:00:00 +0000 Ollama now supports AMD graphics cards https://ollama.com/blog/amd-preview Ollama now supports AMD graphics cards in preview on Windows and Linux. All the features of Ollama can now be accelerated by AMD graphics cards on Ollama for Linux and Windows. https://ollama.com/blog/amd-preview Thu, 14 Mar 2024 00:00:00 +0000 Embedding models https://ollama.com/blog/embedding-models Embedding models are available in Ollama, making it easy to generate vector embeddings for use in search and retrieval augmented generation (RAG) applications. https://ollama.com/blog/embedding-models Mon, 08 Apr 2024 00:00:00 +0000 Llama 3 https://ollama.com/blog/llama3 Llama 3 is now available to run on Ollama. This model is the next generation of Meta's state-of-the-art large language model, and is the most capable openly available LLM to date. https://ollama.com/blog/llama3 Thu, 18 Apr 2024 00:00:00 +0000 Llama 3 is not very censored https://ollama.com/blog/llama-3-is-not-very-censored Compared to Llama 2, Llama 3 feels much less censored. Meta has substantially lowered false refusal rates. Llama 3 will refuse less than 1/3 of the prompts previously refused by Llama 2. https://ollama.com/blog/llama-3-is-not-very-censored Fri, 19 Apr 2024 00:00:00 +0000 Google announces Firebase Genkit with Ollama support https://ollama.com/blog/firebase-genkit At Google IO 2024, Google announced Ollama support in Firebase Genkit, a new open-source framework for developers to build, deploy and monitor production-ready AI-powered apps. https://ollama.com/blog/firebase-genkit Mon, 20 May 2024 00:00:00 +0000 An entirely open-source AI code assistant inside your editor https://ollama.com/blog/continue-code-assistant Continue enables you to easily create your own coding assistant directly inside Visual Studio Code and JetBrains with open-source LLMs. https://ollama.com/blog/continue-code-assistant Fri, 31 May 2024 00:00:00 +0000 Google Gemma 2 https://ollama.com/blog/gemma2 Gemma 2 is now available on Ollama in 3 sizes - 2B, 9B and 27B. https://ollama.com/blog/gemma2 Thu, 27 Jun 2024 00:00:00 +0000 Tool support https://ollama.com/blog/tool-support Ollama now supports tool calling with popular models such as Llama 3.1. This enables a model to answer a given prompt using tool(s) it knows about, making it possible for models to perform more complex tasks or interact with the outside world. https://ollama.com/blog/tool-support Thu, 25 Jul 2024 00:00:00 +0000 Reduce hallucinations with Bespoke-Minicheck https://ollama.com/blog/reduce-hallucinations-with-bespoke-minicheck Bespoke-Minicheck is a new grounded factuality checking model developed by Bespoke Labs that is now available in Ollama. It can fact-check responses generated by other models to detect and reduce hallucinations. https://ollama.com/blog/reduce-hallucinations-with-bespoke-minicheck Wed, 18 Sep 2024 00:00:00 +0000 Llama 3.2 goes small and multimodal https://ollama.com/blog/llama3.2 Ollama partners with Meta to bring Llama 3.2 to Ollama. https://ollama.com/blog/llama3.2 Wed, 25 Sep 2024 00:00:00 +0000 IBM Granite 3.0 models https://ollama.com/blog/ibm-granite Ollama partners with IBM to bring Granite 3.0 models to Ollama. https://ollama.com/blog/ibm-granite Mon, 21 Oct 2024 00:00:00 +0000 Llama 3.2 Vision https://ollama.com/blog/llama3.2-vision Llama 3.2 Vision 11B and 90B models are now available in Ollama. https://ollama.com/blog/llama3.2-vision Wed, 06 Nov 2024 00:00:00 +0000 Ollama Python library 0.4 with function calling improvements https://ollama.com/blog/functions-as-tools With Ollama Python library version 0.4, functions can now be provided as tools. The library now also has full typing support and new examples have been added. https://ollama.com/blog/functions-as-tools Mon, 25 Nov 2024 00:00:00 +0000 Structured outputs https://ollama.com/blog/structured-outputs Ollama now supports structured outputs making it possible to constrain a model's output to a specific format defined by a JSON schema. The Ollama Python and JavaScript libraries have been updated to support structured outputs. https://ollama.com/blog/structured-outputs Fri, 06 Dec 2024 00:00:00 +0000 Minions: where local and cloud LLMs meet https://ollama.com/blog/minions Avanika Narayan, Dan Biderman, and Sabri Eyuboglu from Christopher Ré's Stanford Hazy Research lab, along with Avner May, Scott Linderman, James Zou, have developed a way to shift a substantial portion of LLM workloads to consumer devices by having small on-device models (such as Llama 3.2 with Ollama) collaborate with larger models in the cloud (such as GPT-4o). https://ollama.com/blog/minions Tue, 25 Feb 2025 00:00:00 +0000 Ollama's new engine for multimodal models https://ollama.com/blog/multimodal-models Ollama now supports new multimodal models with its new engine. https://ollama.com/blog/multimodal-models Thu, 15 May 2025 00:00:00 +0000 Streaming responses with tool calling https://ollama.com/blog/streaming-tool Ollama now supports streaming responses with tool calling. This enables all chat applications to stream content and also call tools in real time. https://ollama.com/blog/streaming-tool Wed, 28 May 2025 00:00:00 +0000 Thinking https://ollama.com/blog/thinking Ollama now has the ability to enable or disable thinking. This gives users the flexibility to choose the model’s thinking behavior for different applications and use cases. https://ollama.com/blog/thinking Fri, 30 May 2025 00:00:00 +0000 Secure Minions: private collaboration between Ollama and frontier models https://ollama.com/blog/secureminions Secure Minions is a secure protocol built by Stanford's Hazy Research lab to allow encrypted local-remote communication. https://ollama.com/blog/secureminions Tue, 03 Jun 2025 00:00:00 +0000 Ollama's new app https://ollama.com/blog/new-app Ollama's new app is now available for macOS and Windows. https://ollama.com/blog/new-app Wed, 30 Jul 2025 00:00:00 +0000 OpenAI gpt-oss https://ollama.com/blog/gpt-oss Ollama partners with OpenAI to bring gpt-oss to Ollama and its community. https://ollama.com/blog/gpt-oss Tue, 05 Aug 2025 00:00:00 +0000 Cloud models https://ollama.com/blog/cloud-models Cloud models are now in preview, letting you run larger models with fast, datacenter-grade hardware. You can keep using your local tools while running larger models that wouldn’t fit on a personal computer. https://ollama.com/blog/cloud-models Fri, 19 Sep 2025 00:00:00 +0000 New model scheduling https://ollama.com/blog/new-model-scheduling Ollama now includes a significantly improved model scheduling system, reducing crashes due to out of memory issues, maximizing GPU utilization and performance, especially on multi-GPU systems. https://ollama.com/blog/new-model-scheduling Tue, 23 Sep 2025 00:00:00 +0000 Web search https://ollama.com/blog/web-search A new web search API is now available in Ollama. Ollama provides a generous free tier of web searches for individuals to use, and higher rate limits are available via Ollama’s cloud. https://ollama.com/blog/web-search Wed, 24 Sep 2025 00:00:00 +0000 NVIDIA DGX Spark https://ollama.com/blog/nvidia-spark The latest NVIDIA DGX Spark is here! Ollama has partnered with NVIDIA to ensure it runs fast and efficiently out-of-the-box. https://ollama.com/blog/nvidia-spark Mon, 13 Oct 2025 00:00:00 +0000 Qwen3-VL https://ollama.com/blog/qwen3-vl Ollama now supports Alibaba's Qwen3-VL. https://ollama.com/blog/qwen3-vl Tue, 14 Oct 2025 00:00:00 +0000 New coding models & integrations https://ollama.com/blog/coding-models GLM-4.6 and Qwen3-coder-480B are available on Ollama’s cloud service with easy integrations to the tools you are familiar with. Qwen3-Coder-30B has been updated for faster, more reliable tool calling in Ollama’s new engine. https://ollama.com/blog/coding-models Thu, 16 Oct 2025 00:00:00 +0000 NVIDIA DGX Spark performance https://ollama.com/blog/nvidia-spark-performance We ran performance tests on release day firmware and an updated Ollama version to see how Ollama performs. https://ollama.com/blog/nvidia-spark-performance Thu, 23 Oct 2025 00:00:00 +0000 MiniMax M2 https://ollama.com/blog/minimax-m2 MiniMax M2 is now available on Ollama's cloud. It's a model built for coding and agentic workflows. https://ollama.com/blog/minimax-m2 Tue, 28 Oct 2025 00:00:00 +0000 OpenAI gpt-oss-safeguard https://ollama.com/blog/gpt-oss-safeguard Ollama is partnering with OpenAI and ROOST (Robust Open Online Safety Tools) to bring the latest gpt-oss-safeguard reasoning models to users for safety classification tasks. gpt-oss-safeguard models are available in two sizes: 20B and 120B, and are permissively licensed under the Apache 2.0 license. https://ollama.com/blog/gpt-oss-safeguard Wed, 29 Oct 2025 00:00:00 +0000