Phoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting. It provides:
- Tracing - Trace your LLM application's runtime using OpenTelemetry-based instrumentation.
- Evaluation - Leverage LLMs to benchmark your application's performance using response and retrieval evals.
- Datasets - Create versioned datasets of examples for experimentation, evaluation, and fine-tuning.
- Experiments - Track and evaluate changes to prompts, LLMs, and retrieval.
- Playground- Optimize prompts, compare models, adjust parameters, and replay traced LLM calls.
- Prompt Management- Manage and test prompt changes systematically using version control, tagging, and experimentation.
Phoenix is vendor and language agnostic with out-of-the-box support for popular frameworks (π¦LlamaIndex, π¦βLangChain, Haystack, π§©DSPy, π€smolagents) and LLM providers (OpenAI, Bedrock, MistralAI, VertexAI, LiteLLM, Google GenAI and more). For details on auto-instrumentation, check out the OpenInference project.
Phoenix runs practically anywhere, including your local machine, a Jupyter notebook, a containerized deployment, or in the cloud.
Install Phoenix via pip or conda
pip install arize-phoenixPhoenix container images are available via Docker Hub and can be deployed using Docker or Kubernetes. Arize AI also provides cloud instances at app.phoenix.arize.com.
The arize-phoenix package includes the entire Phoenix platfom. However if you have deployed the Phoenix platform, there are light-weight Python sub-packages and TypeScript packages that can be used in conjunction with the platfrom.
| Package | Version & Docs | Description |
|---|---|---|
| arize-phoenix-otel | Provides a lightweight wrapper around OpenTelemetry primitives with Phoenix-aware defaults | |
| arize-phoenix-client | Lightweight client for interacting with the Phoenix server via its OpenAPI REST interface | |
| arize-phoenix-evals | Tooling to evaluate LLM applications including RAG relevance, answer relevance, and more |
| Package | Version & Docs | Description |
|---|---|---|
| @arizeai/phoenix-otel | Provides a lightweight wrapper around OpenTelemetry primitives with Phoenix-aware defaults | |
| @arizeai/phoenix-client | Client for the Arize Phoenix API | |
| @arizeai/phoenix-evals | TypeScript evaluation library for LLM applications (alpha release) | |
| @arizeai/phoenix-mcp | MCP server implementation for Arize Phoenix providing unified interface to Phoenix's capabilities |
Phoenix is built on top of OpenTelemetry and is vendor, language, and framework agnostic. For details about tracing integrations and example applications, see the OpenInference project.
Python Integrations
| Integration | Package | Version Badge |
|---|---|---|
| OpenAI | openinference-instrumentation-openai |
|
| OpenAI Agents | openinference-instrumentation-openai-agents |
|
| LlamaIndex | openinference-instrumentation-llama-index |
|
| DSPy | openinference-instrumentation-dspy |
|
| AWS Bedrock | openinference-instrumentation-bedrock |
|
| LangChain | openinference-instrumentation-langchain |
|
| MistralAI | openinference-instrumentation-mistralai |
|
| Google GenAI | openinference-instrumentation-google-genai |
|
| Google ADK | openinference-instrumentation-google-adk |
|
| Guardrails | openinference-instrumentation-guardrails |
|
| VertexAI | openinference-instrumentation-vertexai |
|
| CrewAI | openinference-instrumentation-crewai |
|
| Haystack | openinference-instrumentation-haystack |
|
| LiteLLM | openinference-instrumentation-litellm |
|
| Groq | openinference-instrumentation-groq |
|
| Instructor | openinference-instrumentation-instructor |
|
| Anthropic | openinference-instrumentation-anthropic |
|
| Smolagents | openinference-instrumentation-smolagents |
|
| Agno | openinference-instrumentation-agno |
|
| MCP | openinference-instrumentation-mcp |
|
| Pydantic AI | openinference-instrumentation-pydantic-ai |
|
| Autogen AgentChat | openinference-instrumentation-autogen-agentchat |
|
| Portkey | openinference-instrumentation-portkey |
Normalize and convert data across other instrumentation libraries by adding span processors that unify data.
| Package | Description | Version |
|---|---|---|
openinference-instrumentation-openlit |
OpenInference Span Processor for OpenLIT traces. | |
openinference-instrumentation-openllmetry |
OpenInference Span Processor for OpenLLMetry (Traceloop) traces. |
| Integration | Package | Version Badge |
|---|---|---|
| OpenAI | @arizeai/openinference-instrumentation-openai |
|
| LangChain.js | @arizeai/openinference-instrumentation-langchain |
|
| Vercel AI SDK | @arizeai/openinference-vercel |
|
| BeeAI | @arizeai/openinference-instrumentation-beeai |
|
| Mastra | @mastra/arize |
| Integration | Package | Version Badge |
|---|---|---|
| LangChain4j | openinference-instrumentation-langchain4j |
|
| SpringAI | openinference-instrumentation-springAI |
| Platform | Description | Docs |
|---|---|---|
| BeeAI | AI agent framework with built-in observability | Integration Guide |
| Dify | Open-source LLM app development platform | Integration Guide |
| Envoy AI Gateway | AI Gateway built on Envoy Proxy for AI workloads | Integration Guide |
| LangFlow | Visual framework for building multi-agent and RAG applications | Integration Guide |
| LiteLLM Proxy | Proxy server for LLMs | Integration Guide |
We take data security and privacy very seriously. For more details, see our Security and Privacy documentation.
By default, Phoenix does not collect any telemetry from self-hosted instances. However, in the future we may enable tracking for usage analytics.
This helps us to:
- Understand how Phoenix is used and improve the most relevant features
- Track overall usage for internal and external reporting
When telemetry is enabled, only basic web analytics are collected (e.g., page views, UI interactions). None of your trace data, evaluation results, or any sensitive information is ever collected.
You can permanently opt-out of telemetry by setting the environment variable: PHOENIX_TELEMETRY_ENABLED=false
Join our community to connect with thousands of AI builders.
- π Join our Slack community.
- π Read our documentation.
- π‘ Ask questions and provide feedback in the #phoenix-support channel.
- π Leave a star on our GitHub.
- π Report bugs with GitHub Issues.
- π Follow us on π.
- πΊοΈ Check out our roadmap to see where we're heading next.
- π§βπ« Deep dive into everything Agents and LLM Evaluations on Arize's Learning Hubs.
See the migration guide for a list of breaking changes.
Copyright 2025 Arize AI, Inc. All Rights Reserved.
Portions of this code are patent protected by one or more U.S. Patents. See the IP_NOTICE.
This software is licensed under the terms of the Elastic License 2.0 (ELv2). See LICENSE.
