Arize AI
banner
arize.bsky.social
Arize AI
@arize.bsky.social
Arize is an AI engineering platform focused on evaluation and observability. It helps engineers develop, evaluate, and observe AI applications and agents.
AI is woven into customer journeys at TheFork, one of Europe’s leading restaurant discovery and booking platforms.

And Arize AX is layered across TheFork's stack, with tracing helping drive tangible wins in terms of lower latency, clearer cost signals, and faster iteration.
December 10, 2025 at 1:30 PM
Metals and mining giant Rio Tinto now relies on Arize AX as it evaluates and deploys new gen-AI use cases.
December 5, 2025 at 7:28 PM
Get certified 🎓 in AI Agent Mastery in our new, free course led by Srilakshmi Chavali: courses.arize.com/l/pdp/ai-ag...

This course covers the latest in:
🟣 Agent architectures and frameworks
🟣 Tools & MCP
🟣 Agentic RAG
🟣 Agent evaluation
🟣 Post-deployment and monitoring

Each module has a lab.
December 5, 2025 at 2:30 PM
Our LLM-as-a-Judge 101 virtual workshop was so popular, we're returning for LLM-as-a-Judge 102. 🎓 RSVP: luma.com/ab78cmgo

In this session focused on meta-evaluation, you’ll learn advanced techniques -- like using high-temperature stress tests to detect prompt ambiguity or unstable reasoning.
December 2, 2025 at 3:10 PM
In case you missed it last week: we released day 0 support for Claude Opus 4.5 in Phoenix! Try it out in the prompt playground today!

Learn about the prompt playground:
arize.com/docs/phoeni...

Sign up for Phoenix Cloud:
app.phoenix.arize.com/

Release notes:
github.com/Arize-ai/ph...
December 1, 2025 at 9:56 PM
If you've been debugging agents by scrolling through spans, Arize AX can help you do better!

Agent Graph gives you a node-based visual map of your agent workflows, so you can instantly see execution paths, identify failure points, spot self-looping behavior, and more!

arize.com/docs/ax/obs...
November 28, 2025 at 5:00 PM
Arize AX Monitors: set threshold-based alerts for what matters in your LLM apps—latency, hallucination rates, eval failures, token usage, errors. One-click setup for common metrics, or fully custom. Get notified before your users notice something's wrong.

Learn more:
arize.com/docs/ax/obs...
November 27, 2025 at 5:00 PM
The team is gearing up for an epic @AWSreInvent! We have fun happenings all week (chocolate tastings, dinners, happy hours) — join us! arize.com/aws-reinven...
November 24, 2025 at 8:01 PM
Arize AX is listed as an Emerging Leader in the "Emerging Market Quadrant for Generative AI Engineering" in Gartner's latest "Innovation Guide for Generative AI Engineering" report (13 November).
November 20, 2025 at 2:00 PM
Microsoft's red teaming agent in Microsoft Foundry generates sophisticated prompts designed to simulate adversarial attacks. Arize AX can help make these vulnerabilities visible and actionable.

New blog + notebook outlines how to create self-improving agent security: arize.com/blog/how-to...
November 19, 2025 at 7:47 PM
Microsoft Foundry + Arize AX = everything you need for self-improving agents.

From the floor of #MSIgnite, a new notebook + blog walks through a concrete content safety evaluation example.

📓 Explore: arize.com/blog/evalua...
November 18, 2025 at 8:58 PM
Prompt Learning also ships with no-code optimization in Arize.
You can run experiments in the UI, track every iteration in the Prompt Hub, and test new versions in the Prompt Playground.
Great for teams who want collaboration + governance without managing a giant text blob in Git.
November 17, 2025 at 9:22 PM
GEPA requires your application to be written in DSPy to generate traces.
Prompt Learning is framework-agnostic — LangChain, CrewAI, Mastra, AutoGen, vector DBs, custom stacks, whatever.
Just add tracing (OpenInference), export the traces, and optimize.
Start tracing: arize.com/docs/ax/obs...
November 17, 2025 at 9:22 PM
Both Prompt Learning and GEPA follow the same optimization loop:
run → evaluate → improve → repeat.
They both use meta-prompting and trace-level evals so the optimizer can learn from application behavior — not just static prompts.
Under the hood, both systems are applying RL-style feedback loops.
November 17, 2025 at 9:22 PM
We benchmarked Prompt Learning (prompt optimizer) against GEPA and saw similar/better results in a fraction of the time.

Since we launched Prompt Learning in July, the most common question we get is:
“Prompt Learning or GEPA — which should I use?”
We break down the results below.
November 17, 2025 at 9:22 PM
Our thanks to Google Cloud for hosting in Sunnyvale and speaking alongside Meta AI and the Arize crew at "Agents In Action" last night!
November 13, 2025 at 8:00 PM
Quickly add valuable human insights to your sessions without breaking your flow!

With our new Session Annotations, you can now add notes directly from the Session Page, eliminating the need to switch between views or lose context.
November 13, 2025 at 5:00 PM
By popular demand, we're covering 🎓 ​LLM-as-a-judge 101 🎓 in our next workshop! RSVP: luma.com/tmipn699

Learn how to design your eval from scratch -- including what to measure, which model to use, how to prompt effectively, and how to improve your eval.
November 11, 2025 at 4:44 PM
Thanks to Mastra for organizing the first conference for TypeScript AI developers! If you missed, it, don't fret: we have an upcoming event with Mastra at GitHub HQ in SF on building and evaluating typescript agents luma.com/l3n0qg61
November 11, 2025 at 2:10 AM
We're excited to be headed to @databricksinc.bsky.social Data + AI Summit next week!

Aparna has a session on building & evaluating self-improving agents with Arize, Databricks MLFlow, & Mosaic AI.

Info here: www.databricks.com/dataaisummit...
June 4, 2025 at 8:10 PM
🚀 Get ready to learn from a powerhouse group of speakers at
Shack15 in SF on June 25. Builders, researchers, and leaders from @anthropic.com @microsoft.com @llamaindex.bsky.social (+ many more).

Get tickets: arize.com/observe-2025
June 3, 2025 at 2:34 AM
Chicago event alert: We’re bringing Chicago’s AI builders together at Google’s office for hands-on sessions focused on advancing LLM-powered agents. 🚀

Join us May 19. Space is limited!

Register: lu.ma/d6mo5zxs
April 29, 2025 at 5:23 PM
Join us 6/25 in SF for a full-day event focused on agent reliability and evaluation.

Hear from the people building the next generation of AI systems—it's conference by engineers, for engineers.

Most of our speakers on the site. 👀

Register: arize.com/observe-2025/
April 17, 2025 at 9:07 PM
Demo your app at this year's Observe! Fill out a short application by 4.30 to be considered for our Demo Den. Great opportunity to showcase your work to the AI community in SF.

Apply here: docs.google.com/forms/d/e/1F...
March 28, 2025 at 9:11 PM