Page 30 of 133 · 1,324 tools
Unify AI
NEWUnify automatically routes LLM requests to the cheapest or fastest provider based on your optimization criteria. Benchmark any…
Portkey AI
NEWPortkey is an AI gateway providing unified access to 200+ LLMs with built-in observability, caching, and fallbacks. Production-grade…
DeepEval
NEWDeepEval is an open-source LLM evaluation framework with 14+ evaluation metrics including hallucination, answer relevancy, and faithfulness. pytest-style…
TruLens
NEWTruLens is an open-source framework for evaluating and tracking LLM applications. Feedback functions assess truthfulness, harmlessness, and helpfulness…
Phoenix Arize
NEWPhoenix by Arize is an open-source AI observability library for ML engineers. Traces LLM and embedding applications, visualizes…
Braintrust
NEWBraintrust is an enterprise AI evaluation platform for measuring, improving, and shipping AI applications. Logging, evaluation datasets, prompt…
Opik
NEWOpik by Comet is an open-source LLM evaluation framework for testing AI application quality at scale. Automated evaluation…
Helicone
NEWHelicone provides one-line LLM observability — add a single line to your OpenAI calls and get full logging,…
Langfuse
NEWLangfuse is an open-source LLM engineering platform for observability, testing, and prompt management. Debug production AI issues, evaluate…
PromptLayer
NEWPromptLayer is a platform for tracking, managing, and evaluating LLM prompts in production. Log every prompt and completion,…
🔍
Don't see your tool?
We review every submission within 24–48 hours. Free listing, no strings attached.
Submit Your Tool