Ollama allows you to run open-source large language models,
such as gpt-oss, locally.
Ollama
bundles model weights, configuration, and data into a single package, defined by a Modelfile.
It optimizes setup and configuration details, including GPU usage.
For a complete list of supported models and model variants, see the Ollama model library.
See this guide for more details
on how to use ollama
with LangChain.
Installation and Setup
Ollama installation
Follow these instructions to set up and run a local Ollama instance. Ollama will start as a background service automatically, if this is disabled, run:ollama pull <name-of-model>
to download a model from the Ollama model library:
- This will download the default tagged version of the model. Typically, the default points to the latest, smallest sized-parameter model.
- To view all pulled (downloaded) models, use
ollama list
langchain-ollama
partner package and run a model.
Ollama LangChain partner package install
Install the integration package with:LLM
Chat Models
Chat Ollama
Ollama tool calling
Ollama tool calling uses the OpenAI compatible web server specification, and can be used with the defaultBaseChatModel.bind_tools()
methods
as described here.
Make sure to select an ollama model that supports tool calling.