Page 6 of 12 · 1,324 tools
TruLens
NEWTruLens is an open-source framework for evaluating and tracking LLM applications. Feedback functions assess truthfulness, harmlessness, and helpfulness…
Phoenix Arize
NEWPhoenix by Arize is an open-source AI observability library for ML engineers. Traces LLM and embedding applications, visualizes…
Braintrust
NEWBraintrust is an enterprise AI evaluation platform for measuring, improving, and shipping AI applications. Logging, evaluation datasets, prompt…
Helicone
NEWHelicone provides one-line LLM observability — add a single line to your OpenAI calls and get full logging,…
Opik
NEWOpik by Comet is an open-source LLM evaluation framework for testing AI application quality at scale. Automated evaluation…
Langfuse
NEWLangfuse is an open-source LLM engineering platform for observability, testing, and prompt management. Debug production AI issues, evaluate…
PromptLayer
NEWPromptLayer is a platform for tracking, managing, and evaluating LLM prompts in production. Log every prompt and completion,…
Guardrails AI
NEWGuardrails AI adds input/output validation to LLM applications. Define rules for what the LLM can and cannot say,…
LiteLLM
NEWLiteLLM provides a unified API for 100+ LLM providers using the OpenAI format. Switch between GPT-4, Claude, Gemini,…
Instructor
NEWInstructor makes it easy to get structured outputs from LLMs using Python type hints. Define a Pydantic model…
🔍
Don't see your tool?
We review every submission within 24–48 hours. Free listing, no strings attached.
Submit Your Tool