Evaluations

agent evaluation, agent infrastructure, AI Agents, claude-code, claude-skills, Evaluations, large-language-models, LLM tooling, MCP, Research

MCP vs. CLI Skills for agents: what our eval found (and which you should use)

Twitter said pick a side. The eval said the question was wrong. Six months ago, MCP (model context protocol) was the hot new thing: tool usage with a built-in discovery…

The post MCP vs. CLI Skills for agents: what our eval found (and which you should use) appeared first on Arize AI.

evaluation framework, evaluation-driven development (EDD), Evaluations, llm-as-a-judge, llm-evaluation, prompt evaluation, regression testing for llms

Prompt templates as configs, not code

This post was written in April 2026. Cloud products, feature maturity, and recommended patterns change over time, so readers should treat these examples as directional guidance. For teams already using Arize, there is a natural extension of that pattern. Prompt Playground can sit upstream of the config layer as the place where prompts are edited, compared, and versioned before they are promoted into whatever config system the company already trusts in production.

The post Prompt templates as configs, not code appeared first on Arize AI.

Scroll to Top