Skip to main content
Open In ColabOpen on GitHub

OllamaLLM

caution

You are currently on a page documenting the use of Ollama models as text completion models. Many popular Ollama models are chat completion models.

You may be looking for this page instead.

This page goes over how to use LangChain to interact with Ollama models.

Installation

# install package
%pip install -U langchain-ollama

Setup

First, follow these instructions to set up and run a local Ollama instance:

  • Download and install Ollama onto the available supported platforms (including Windows Subsystem for Linux)
  • Fetch available LLM model via ollama pull <name-of-model>
    • View a list of available models via the model library
    • e.g., ollama pull llama3
  • This will download the default tagged version of the model. Typically, the default points to the latest, smallest sized-parameter model.

On Mac, the models will be download to ~/.ollama/models

On Linux (or WSL), the models will be stored at /usr/share/ollama/.ollama/models

  • Specify the exact version of the model of interest as such ollama pull vicuna:13b-v1.5-16k-q4_0 (View the various tags for the Vicuna model in this instance)
  • To view all pulled models, use ollama list
  • To chat directly with a model from the command line, use ollama run <name-of-model>
  • View the Ollama documentation for more commands. Run ollama help in the terminal to see available commands too.

Usage

from langchain_core.prompts import ChatPromptTemplate
from langchain_ollama.llms import OllamaLLM

template = """Question: {question}

Answer: Let's think step by step."""

prompt = ChatPromptTemplate.from_template(template)

model = OllamaLLM(model="llama3.1")

chain = prompt | model

chain.invoke({"question": "What is LangChain?"})

ChatOllama

In some cases, you will need a model with a "Chat" interface. (for more info see - https://python.langchain.com/docs/concepts/chat_models/).

from langchain_core.messages import ChatMessage
from langchain_ollama.chat_models import ChatOllama

llm = ChatOllama(
base_url="http://localhost:11434",
model="granite3.2",
verbose=True,
format=None,
num_predict=20,
)


messages = [
ChatMessage( # We can use ChatMessage objects to pass custom roles to a model in Ollama.
role="control", content="thinking"
),
# ChatMessage(
# role="user",
# content="What is the meaning of life?"
# ),
{ # Dicts also work for messages with standard roles.
"role": "user",
"content": "What is the meaning of life?",
},
]

llm.invoke(messages)
API Reference:ChatMessage | ChatOllama

Multi-modal

Ollama has support for multi-modal LLMs, such as bakllava and llava.

ollama pull bakllava

Be sure to update Ollama so that you have the most recent version to support multi-modal.

import base64
from io import BytesIO

from IPython.display import HTML, display
from PIL import Image


def convert_to_base64(pil_image):
"""
Convert PIL images to Base64 encoded strings

:param pil_image: PIL image
:return: Re-sized Base64 string
"""

buffered = BytesIO()
pil_image.save(buffered, format="JPEG") # You can change the format if needed
img_str = base64.b64encode(buffered.getvalue()).decode("utf-8")
return img_str


def plt_img_base64(img_base64):
"""
Display base64 encoded string as image

:param img_base64: Base64 string
"""
# Create an HTML img tag with the base64 string as the source
image_html = f'<img src="data:image/jpeg;base64,{img_base64}" />'
# Display the image by rendering the HTML
display(HTML(image_html))


file_path = "../../../static/img/ollama_example_img.jpg"
pil_image = Image.open(file_path)
image_b64 = convert_to_base64(pil_image)
plt_img_base64(image_b64)
from langchain_ollama import OllamaLLM

llm = OllamaLLM(model="bakllava")

llm_with_image_context = llm.bind(images=[image_b64])
llm_with_image_context.invoke("What is the dollar based gross retention rate:")
API Reference:OllamaLLM

Was this page helpful?