Page 6 of 14 · 1,324 tools
Mirascope
NEWMirascope is a Python toolkit for building LLM applications with clean abstractions for prompts, calls, and extractions. Type-safe…
HoneyHive
NEWHoneyHive is an AI evaluation and observability platform for teams building LLM applications. Dataset management, automated evaluations, and…
Agenta
NEWAgenta is an open-source LLMOps platform for prompt management, evaluation, and deployment. Teams collaborate on prompts, run systematic…
Eden AI
NEWEden AI provides a unified API for 100+ AI models across text, image, audio, and video. Test and…
Unify AI
NEWUnify automatically routes LLM requests to the cheapest or fastest provider based on your optimization criteria. Benchmark any…
Portkey AI
NEWPortkey is an AI gateway providing unified access to 200+ LLMs with built-in observability, caching, and fallbacks. Production-grade…
DeepEval
NEWDeepEval is an open-source LLM evaluation framework with 14+ evaluation metrics including hallucination, answer relevancy, and faithfulness. pytest-style…
TruLens
NEWTruLens is an open-source framework for evaluating and tracking LLM applications. Feedback functions assess truthfulness, harmlessness, and helpfulness…
Phoenix Arize
NEWPhoenix by Arize is an open-source AI observability library for ML engineers. Traces LLM and embedding applications, visualizes…
Braintrust
NEWBraintrust is an enterprise AI evaluation platform for measuring, improving, and shipping AI applications. Logging, evaluation datasets, prompt…
🔍
Don't see your tool?
We review every submission within 24–48 hours. Free listing, no strings attached.
Submit Your Tool