We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 170921d commit 228b3ddCopy full SHA for 228b3dd
1 file changed
docs/content/docs/introduction.mdx
@@ -23,7 +23,7 @@ import {
23
- Unit test LLM outputs with Pytest-style assertions.
24
- Use 50+ ready-to-use metrics, including LLM-as-a-judge, agent, tool-use,
25
conversational, safety, RAG, and multimodal metrics.
26
-- Evaluate AI agents, MCP systems, chatbots, LLM arenas, RAG pipelines, and
+- Evaluate AI agents, conversational agents (chatbots), RAG pipelines, MCP systems, and
27
other custom workflows.
28
- Run both end-to-end evals and component-level evals with tracing.
29
- Generate synthetic datasets for edge cases that are hard to collect manually.
0 commit comments