We recently added support for custom models like llama or GPT-Neo. This allows you to use your own models with Helicone. This is currently in beta, so please let us know if you have any issues.

Add HELICONE_API_KEY to your environment variables.

export HELICONE_API_KEY=sk-<your-api-key>
# You can also set it in your code (See below)
import {
  IHeliconeAsyncClientOptions,
  HeliconeLogBuilder,
  HeliconeLogger,
  ResponseBody,
} from "@helicone/helicone";

const heliconeApiKey = process.env.HELICONE_API_KEY;

const config: IHeliconeAsyncClientOptions = {
  heliconeMeta: {
    apiKey: heliconeApiKey,
    baseUrl: "https://api.hconeai.com/custom/v1/logs",
  },
};

const logger = new HeliconeLogger(config);

const llmArgs = {
  model: "llama-2",
  prompt: "Say hi!",
};
const builder = new HeliconeLogBuilder(llmArgs);

/*
result = callToLLM(llmArgs)
*/

const result: ResponseBody = {
  text: "This is my response",
  usage: {
    total_tokens: 13,
    prompt_tokens: 5,
    completion_tokens: 8,
  },
};

builder.addResponse(result);
builder.addUser("test-user");
const response = await logger.submit(builder);
if (response.status !== 200) {
  throw new Error(response.data);
}