PostHog Node AI
TypeScript SDK for LLM observability with PostHog.
SEE FULL DOCS
Installation
npm install @posthog/ai
Direct Provider Usage
import { OpenAI } from '@posthog/ai'
import { PostHog } from 'posthog-node'
const phClient = new PostHog('<YOUR_PROJECT_API_KEY>', { host: 'https://us.i.posthog.com' })
const client = new OpenAI({
apiKey: '<YOUR_OPENAI_API_KEY>',
posthog: phClient,
})
const completion = await client.chat.completions.create({
model: 'gpt-5-mini',
messages: [{ role: 'user', content: 'Tell me a fun fact about hedgehogs' }],
posthogDistinctId: 'user_123',
posthogTraceId: 'trace_123',
posthogProperties: { conversation_id: 'abc123', paid: true },
posthogGroups: { company: 'company_id_in_your_db' },
posthogPrivacyMode: false,
})
console.log(completion.choices[0].message.content)
await phClient.shutdown()
OpenTelemetry
@posthog/ai/otel provides two ways to send AI traces to PostHog via OpenTelemetry. Both automatically filter to AI-related spans only (gen_ai.*, llm.*, ai.*, traceloop.*) and PostHog converts them into $ai_generation events server-side. This works with any LLM provider SDK that supports OpenTelemetry.
npm install @posthog/ai @opentelemetry/sdk-node @opentelemetry/sdk-trace-base @opentelemetry/exporter-trace-otlp-http
PostHogSpanProcessor (recommended)
A self-contained SpanProcessor that handles batching and export internally. Use this when your setup accepts a span processor.
import { NodeSDK } from '@opentelemetry/sdk-node'
import { PostHogSpanProcessor } from '@posthog/ai/otel'
import { generateText } from 'ai'
import { openai } from '@ai-sdk/openai'
const sdk = new NodeSDK({
spanProcessors: [
new PostHogSpanProcessor({
apiKey: '<YOUR_PROJECT_API_KEY>',
host: 'https://us.i.posthog.com',
}),
],
})
sdk.start()
const result = await generateText({
model: openai('gpt-5-mini'),
prompt: 'Write a short haiku about debugging',
experimental_telemetry: {
isEnabled: true,
functionId: 'my-awesome-function',
metadata: {
posthog_distinct_id: 'user_123',
conversation_id: 'abc123',
},
},
})
await sdk.shutdown()
PostHogTraceExporter
A TraceExporter for APIs that only accept an exporter, such as Vercel's registerOTel.
import { PostHogTraceExporter } from '@posthog/ai/otel'
import { registerOTel } from '@vercel/otel'
registerOTel({
serviceName: 'my-app',
traceExporter: new PostHogTraceExporter({
apiKey: '<YOUR_PROJECT_API_KEY>',
host: 'https://us.i.posthog.com',
}),
})
LLM Observability docs
Please see the main PostHog docs.
Questions?