README.md

May 3, 2026 Β· View on GitHub

phoenix banner

Add Arize Phoenix MCP server to Cursor

Phoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting. It provides:

  • Tracing - Trace your LLM application's runtime using OpenTelemetry-based instrumentation.
  • Evaluation - Leverage LLMs to benchmark your application's performance using response and retrieval evals.
  • Datasets - Create versioned datasets of examples for experimentation, evaluation, and fine-tuning.
  • Experiments - Track and evaluate changes to prompts, LLMs, and retrieval.
  • Playground- Optimize prompts, compare models, adjust parameters, and replay traced LLM calls.
  • Prompt Management- Manage and test prompt changes systematically using version control, tagging, and experimentation.

Phoenix is vendor and language agnostic with out-of-the-box support for popular frameworks (OpenAI Agents SDK, Claude Agent SDK, LangGraph, Vercel AI SDK, Mastra, CrewAI, LlamaIndex, DSPy) and LLM providers (OpenAI, Anthropic, Google GenAI, Google ADK, AWS Bedrock, OpenRouter, LiteLLM, and more). For details on auto-instrumentation, check out the OpenInference project.

Phoenix runs practically anywhere, including your local machine, a Jupyter notebook, a containerized deployment, or in the cloud.

Installation

Install Phoenix via pip or conda

pip install arize-phoenix

Phoenix container images are available via Docker Hub and can be deployed using Docker or Kubernetes. Arize AI also provides cloud instances at app.phoenix.arize.com.

Packages

The arize-phoenix package includes the entire Phoenix platform. However, if you have deployed the Phoenix platform, there are lightweight Python sub-packages and TypeScript packages that can be used in conjunction with the platform.

Python Subpackages

PackageVersion & DocsDescription
arize-phoenix-otelPyPI Version DocsProvides a lightweight wrapper around OpenTelemetry primitives with Phoenix-aware defaults
arize-phoenix-clientPyPI Version DocsLightweight client for interacting with the Phoenix server via its OpenAPI REST interface
arize-phoenix-evalsPyPI Version DocsTooling to evaluate LLM applications including RAG relevance, answer relevance, and more

TypeScript Subpackages

PackageVersion & DocsDescription
@arizeai/phoenix-otelNPM Version DocsProvides a lightweight wrapper around OpenTelemetry primitives with Phoenix-aware defaults
@arizeai/phoenix-clientNPM Version DocsClient for the Arize Phoenix API
@arizeai/phoenix-evalsNPM Version DocsTypeScript evaluation library for LLM applications (alpha release)
@arizeai/phoenix-mcpNPM Version DocsMCP server implementation for Arize Phoenix providing unified interface to Phoenix's capabilities
@arizeai/phoenix-cliNPM Version DocsCLI for fetching traces, datasets, and experiments for use with Claude Code, Cursor, and other coding agents

Tracing Integrations

Phoenix is built on top of OpenTelemetry and is vendor, language, and framework agnostic. For details about tracing integrations and example applications, see the OpenInference project.

Python Integrations

IntegrationPackageVersion
OpenAIopeninference-instrumentation-openaiPyPI Version
OpenAI Agentsopeninference-instrumentation-openai-agentsPyPI Version
LlamaIndexopeninference-instrumentation-llama-indexPyPI Version
DSPyopeninference-instrumentation-dspyPyPI Version
AWS Bedrockopeninference-instrumentation-bedrockPyPI Version
LangChainopeninference-instrumentation-langchainPyPI Version
MistralAIopeninference-instrumentation-mistralaiPyPI Version
Google GenAIopeninference-instrumentation-google-genaiPyPI Version
Google ADKopeninference-instrumentation-google-adkPyPI Version
Guardrailsopeninference-instrumentation-guardrailsPyPI Version
VertexAIopeninference-instrumentation-vertexaiPyPI Version
CrewAIopeninference-instrumentation-crewaiPyPI Version
Haystackopeninference-instrumentation-haystackPyPI Version
LiteLLMopeninference-instrumentation-litellmPyPI Version
Groqopeninference-instrumentation-groqPyPI Version
Instructoropeninference-instrumentation-instructorPyPI Version
Anthropicopeninference-instrumentation-anthropicPyPI Version
Smolagentsopeninference-instrumentation-smolagentsPyPI Version
Agnoopeninference-instrumentation-agnoPyPI Version
MCPopeninference-instrumentation-mcpPyPI Version
Pydantic AIopeninference-instrumentation-pydantic-aiPyPI Version
Autogen AgentChatopeninference-instrumentation-autogen-agentchatPyPI Version
Portkeyopeninference-instrumentation-portkeyPyPI Version
Agent Specopeninference-instrumentation-agentspecPyPI Version
Claude Agent SDKopeninference-instrumentation-claude-agent-sdkPyPI Version

Span Processors

Normalize and convert data across other instrumentation libraries by adding span processors that unify data.

PackageDescriptionVersion
openinference-instrumentation-openlitOpenInference Span Processor for OpenLIT traces.PyPI Version
openinference-instrumentation-openllmetryOpenInference Span Processor for OpenLLMetry (Traceloop) traces.PyPI Version

JavaScript Integrations

IntegrationPackageVersion
OpenAI@arizeai/openinference-instrumentation-openaiNPM Version
LangChain.js@arizeai/openinference-instrumentation-langchainNPM Version
Vercel AI SDK@arizeai/openinference-vercelNPM Version
BeeAI@arizeai/openinference-instrumentation-beeaiNPM Version
Claude Agent SDK@arizeai/openinference-instrumentation-claude-agent-sdkNPM Version
Mastra@mastra/arizeNPM Version
MCP@arizeai/openinference-instrumentation-mcpNPM Version

Java Integrations

IntegrationPackageVersion
LangChain4jopeninference-instrumentation-langchain4jMaven Central
SpringAIopeninference-instrumentation-springAIMaven Central
Arconiaopeninference-instrumentation-springAIMaven Central

Platforms

PlatformDescriptionDocs
BeeAIAI agent framework with built-in observabilityIntegration Guide
DifyOpen-source LLM app development platformIntegration Guide
Envoy AI GatewayAI Gateway built on Envoy Proxy for AI workloadsIntegration Guide
LangFlowVisual framework for building multi-agent and RAG applicationsIntegration Guide
LiteLLM ProxyProxy server for LLMsIntegration Guide
FlowiseVisual framework for building LLM applicationsIntegration Guide
Prompt FlowMicrosoft's prompt flow orchestration toolIntegration Guide
NVIDIA NeMoNVIDIA NeMo Agent Toolkit for enterprise agentsIntegration Guide
GraphiteMulti-agent LLM workflow framework with visual builderIntegration Guide

Coding Agent Skills

This repository includes skills that teach coding agents how to work with Phoenix. They are located in .agents/skills/ and can be used with Claude Code, Cursor, and other compatible tools.

SkillDescription
phoenix-cliDebug LLM applications using the Phoenix CLI β€” fetch traces, analyze errors, review experiments, and query the GraphQL API
phoenix-evalsBuild and run evaluators for AI/LLM applications using Phoenix
phoenix-tracingOpenInference semantic conventions and instrumentation for tracing LLM applications

Security & Privacy

We take data security and privacy very seriously. For more details, see our Security and Privacy documentation.

Telemetry

By default, Phoenix collects basic web analytics (e.g., page views, UI interactions) to help us understand how Phoenix is used and improve the product. None of your trace data, evaluation results, or any sensitive information is ever collected.

You can opt-out of telemetry by setting the environment variable: PHOENIX_TELEMETRY_ENABLED=false

Community

Join our community to connect with thousands of AI builders.

  • 🌍 Join our Slack community.
  • πŸ“š Read our documentation.
  • πŸ’‘ Ask questions and provide feedback in the #phoenix-support channel.
  • 🌟 Leave a star on our GitHub.
  • 🐞 Report bugs with GitHub Issues.
  • 𝕏 Follow us on 𝕏.
  • πŸ’Ό Follow us on LinkedIn.
  • πŸ—ΊοΈ Check out our roadmap to see where we're heading next.
  • πŸ§‘β€πŸ« Deep dive into everything Agents and LLM Evaluations on Arize's Learning Hubs.

Breaking Changes

See the migration guide for a list of breaking changes.

Copyright 2025 Arize AI, Inc. All Rights Reserved.

Portions of this code are patent protected by one or more U.S. Patents. See the IP_NOTICE.

This software is licensed under the terms of the Elastic License 2.0 (ELv2). See LICENSE.