Groq provides low-latency inference for open-source large language models — Llama, Mixtral, Gemma, and others — through the Groq Python SDK. Arize AX captures every Groq SDK call — chat completions, tool calls, and token usage — via theDocumentation Index
Fetch the complete documentation index at: https://arize-ax.mintlify.dev/docs/llms.txt
Use this file to discover all available pages before exploring further.
openinference-instrumentation-groq package.
Prerequisites
- Python 3.10+
- An Arize AX account (sign up)
- A
GROQ_API_KEYfrom the Groq Console
Launch Arize AX
- Sign in to your Arize AX account.
- From Space Settings, copy your Space ID and API Key. You will set them as
ARIZE_SPACE_IDandARIZE_API_KEYbelow.
Install
Configure credentials
Setup tracing
Run Groq
Expected output
Verify in Arize AX
- Open your Arize AX space and select project
groq-tracing-example. - You should see a new trace within ~30 seconds containing a
CompletionsLLM span with the prompt, response, and token usage attached. - If no traces appear, see Troubleshooting.
Troubleshooting
- No traces in Arize AX. Confirm
ARIZE_SPACE_IDandARIZE_API_KEYare set in the same shell that runsexample.py. Enable OpenTelemetry debug logs withexport OTEL_LOG_LEVEL=debugand re-run. - Groq spans missing but other spans present.
GroqInstrumentor().instrument(...)must run before anyfrom groq import Groqimport. Make sureinstrumentation.pyis the first import in your entry point. 401from Groq. VerifyGROQ_API_KEYis set and active in the Groq Console.404 NOT_FOUNDfor the model. Groq retires older model aliases periodically. Ifllama-3.3-70b-versatilereturns 404, see the Groq supported models list and pick a current one.