This integration is used to log usage with the Nvidia NIM API. For other Nvidia inference providers that are OpenAI-compatible, such as Dynamo, see here.
1

Create an account and generate an API key

Log into Helicone or create an account. Once you have an account, you can generate an API key here.
2

Set up your Helicone API key in your .env file

HELICONE_API_KEY=<your-helicone-api-key>
NVIDIA_API_KEY=<your-nvidia-api-key>
3

Modify the base URL path and set up authentication

OpenAI SDK
import OpenAI from "openai";

const openai = new OpenAI({
  apiKey: process.env.NVIDIA_API_KEY,
  baseURL: "https://nvidia.helicone.ai/v1",
  defaultHeaders: {
    "Helicone-Auth": `Bearer ${process.env.HELICONE_API_KEY}`
  }
});

const response = await openai.chat.completions.create({
  model: "nvidia/llama-3.1-nemotron-70b-instruct",
  messages: [{ role: "user", content: "Hello, how are you?" }],
  max_tokens: 1024,
  temperature: 0.7
});

console.log(response);
4

Verify your requests in Helicone

With the above setup, any calls to Nvidia will automatically be logged and monitored by Helicone. Review them in your Helicone dashboard.