NVIDIA offers a suite of high-performance language models optimized for advanced NLP tasks. These models are part of the NeMo framework, which provides tools for training, fine-tuning and deploying state-of-the-art models efficiently. NVIDIA’s language models are designed to handle large-scale workloads with GPU acceleration for faster inference and training. We recommend experimenting with NVIDIA’s models to find the best fit for your application.

Explore NVIDIA’s models here.

Authentication

Set your NVIDIA_API_KEY environment variable. Get your key from Nvidia here.

export NVIDIA_API_KEY=***

Example

Use Nvidia with your Agent:

from agno.agent import Agent, RunResponse
from agno.models.nvidia import Nvidia

agent = Agent(model=Nvidia(), markdown=True)

# Print the response in the terminal
agent.print_response("Share a 2 sentence horror story")

View more examples here.

Params

ParameterTypeDefaultDescription
idstr"nvidia/llama-3.1-nemotron-70b-instruct"The specific model ID used for generating responses.
namestr"Nvidia"The name identifier for the Nvidia agent.
providerstr-The provider of the model, combining "Nvidia" with the model ID.
api_keyOptional[str]-The API key for authenticating requests to the Nvidia service. Retrieved from the environment variable NVIDIA_API_KEY.
base_urlstr"https://integrate.api.nvidia.com/v1"The base URL for making API requests to the Nvidia service.

Nvidia also supports the params of OpenAI.