Skip to main content
This integration method is maintained but no longer actively developed. For the best experience and latest features, use our new AI Gateway with unified API access to 100+ models.
This integration is used to log usage with the Nvidia NIM API. For other Nvidia inference providers that are OpenAI-compatible, such as Dynamo, see here.
1

Create an account and generate an API key

Log into Helicone or create an account. Once you have an account, you can generate an API key here.
2

Set up your Helicone API key in your .env file

HELICONE_API_KEY=<your-helicone-api-key>
NVIDIA_API_KEY=<your-nvidia-api-key>
3

Modify the base URL path and set up authentication

OpenAI SDK
import OpenAI from "openai";

const openai = new OpenAI({
  apiKey: process.env.NVIDIA_API_KEY,
  baseURL: "https://nvidia.helicone.ai/v1",
  defaultHeaders: {
    "Helicone-Auth": `Bearer ${process.env.HELICONE_API_KEY}`
  }
});

const response = await openai.chat.completions.create({
  model: "nvidia/llama-3.1-nemotron-70b-instruct",
  messages: [{ role: "user", content: "Hello, how are you?" }],
  max_tokens: 1024,
  temperature: 0.7
});

console.log(response);
4

Verify your requests in Helicone

With the above setup, any calls to Nvidia will automatically be logged and monitored by Helicone. Review them in your Helicone dashboard.