Multimodal
Video Caption Agent
Examples
- Introduction
- Getting Started
- Agents
- Workflows
- Applications
Agent Concepts
- Multimodal
- RAG
- Knowledge
- Memory
- Teams
- Async
- Hybrid Search
- Storage
- Tools
- Vector Databases
- Embedders
Models
- Anthropic
- AWS Bedrock Claude
- Azure OpenAI
- Cohere
- DeepSeek
- Fireworks
- Gemini
- Groq
- Hugging Face
- Mistral
- NVIDIA
- Ollama
- OpenAI
- Together
- Vertex AI
- xAI
Multimodal
Video Caption Agent
Code
from agno.agent import Agent
from agno.models.openai import OpenAIChat
from agno.tools.moviepy_video import MoviePyVideoTools
from agno.tools.openai import OpenAITools
video_tools = MoviePyVideoTools(
process_video=True, generate_captions=True, embed_captions=True
)
openai_tools = OpenAITools()
video_caption_agent = Agent(
name="Video Caption Generator Agent",
model=OpenAIChat(
id="gpt-4o",
),
tools=[video_tools, openai_tools],
description="You are an AI agent that can generate and embed captions for videos.",
instructions=[
"When a user provides a video, process it to generate captions.",
"Use the video processing tools in this sequence:",
"1. Extract audio from the video using extract_audio",
"2. Transcribe the audio using transcribe_audio",
"3. Generate SRT captions using create_srt",
"4. Embed captions into the video using embed_captions",
],
markdown=True,
)
video_caption_agent.print_response(
"Generate captions for {video with location} and embed them in the video"
)
Usage
1
Create a virtual environment
Open the Terminal
and create a python virtual environment.
2
Set your API key
export OPENAI_API_KEY=xxx
3
Install libraries
pip install -U openai moviepy ffmpeg agno
4
Run Agent