Skip to main content

Code

cookbook/11_models/ollama/async_basic_stream.py
import asyncio
from typing import Iterator  # noqa

from agno.agent import Agent, RunOutput  # noqa
from agno.models.ollama import Ollama

agent = Agent(model=Ollama(id="llama3.1:8b"), markdown=True)

# Get the response in a variable
# run_response: Iterator[RunOutputEvent] = agent.run("Share a 2 sentence horror story", stream=True)
# for chunk in run_response:
#     print(chunk.content)

# Print the response in the terminal
asyncio.run(agent.aprint_response("Share a 2 sentence horror story", stream=True))

Usage

1

Set up your virtual environment

uv venv --python 3.12
source .venv/bin/activate
2

Install Ollama

Follow the Ollama installation guide and run:
ollama pull llama3.1:8b
3

Install dependencies

uv pip install -U ollama agno
4

Run the example

python examples/models/ollama/async_basic_stream.py