~ / directory / arize-phoenix
AP
Mixed · AI Development Tools · reviewed 2026-04

Arize Phoenix

Open-source LLM tracing, evaluation, and hallucination detection.

Visit phoenix.arize.com
01

What it does

An open-source LLM tracing, evaluation, and hallucination detection platform. Provides detailed traces of LLM interactions with built-in evaluation frameworks for measuring output quality, relevance, and factual accuracy. CB Insights positions Arize in the AI agent tech stack Oversight layer alongside Langfuse and Patronus AI. Critical for detecting ASI06 (Memory & Context Poisoning) through drift monitoring and ASI08 (Cascading Failures) through multi-agent trace analysis.

02

Security relevance

Hallucination detection is a security concern — LLM09 (Overreliance) becomes dangerous when models generate confident but incorrect information that users trust. Arize Phoenix's evaluation framework helps quantify hallucination rates and identify patterns in unreliable outputs.

03

When to use it

Use when you need to measure and monitor LLM output quality, particularly hallucination rates. Python-based with straightforward integration. Guided setup — install, instrument your application, configure evaluations.

04

OWASP coverage

Risks addressed — mapped to both OWASP Top 10 standards. 1 in LLM, 2 in Agentic.

LLM Top 10 · 2025 · 1/10 covered
01
02
03
04
05
06
07
08
09
10
Agentic Top 10 · 2026 · 2/10 covered
01
02
03
04
05
06
07
08
09
10
05

The raw record

What Yuntona stores. Single source of truth — fork it on GitHub.

name: Arize Phoenix
slug: arize-phoenix
type: Mixed
category: AI Development Tools
url: https://phoenix.arize.com

reviewed:   2026-04
added:      2026-04
updated:    2026-04

risks:
  llm:  [LLM09]
  asi:  [ASI06, ASI08]

complexity:    Guided Setup
pricing:       —
audience:      Builder
lifecycle:     [monitor]

tags: [Evals, Observability, Open Source]