You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We need to extend evaluation features, in particular for RAG pipelines so that users can answer questions like:
Is this pipeline good enough?
What should I focus on for optimization?
Is pipeline A better than B? (performance, costs, latency)
This includes the following components typically appearing in RAG pipelines: Retrievers, Rankers, DocumentJoiners
a) labels available => statistical metrics
b) no labels available => model based heuristics / pseudo label generator
Generators
a) labels available = model based (SAS, answer correctness ...)
b) no labels = model based (groundedness score)
The content you are editing has changed. Please copy your edits and refresh the page.
We need to extend evaluation features, in particular for RAG pipelines so that users can answer questions like:
This includes the following components typically appearing in RAG pipelines:
Retrievers, Rankers, DocumentJoiners
a) labels available => statistical metrics
b) no labels available => model based heuristics / pseudo label generator
Generators
a) labels available = model based (SAS, answer correctness ...)
b) no labels = model based (groundedness score)
Tasks
The text was updated successfully, but these errors were encountered: