← Back to all frameworks Graph RAG

Reranking + Evaluation

Cohere rerank, Ragas, TruLens — measure what you ship

What it is

Cohere Rerank for two-stage retrieval (broad ANN → precise rerank). Ragas + TruLens for measuring RAG quality (faithfulness, answer relevance, context precision) — so you actually know your system works.

How Vaaani uses it

  • Reranking top-100 ANN hits down to top-5 with much higher precision
  • Continuous eval pipelines that flag regressions before deploy
  • A/B testing different retrieval strategies with objective scores
  • Reporting accuracy metrics back to the business stakeholder

Why it makes the cut

Without eval, every prompt change is a gut feeling. With Ragas, every change has a number. Vaaani never ships RAG without it.

Sample code

from ragas import evaluate
from ragas.metrics import faithfulness, answer_relevancy

result = evaluate(
    dataset=eval_set,
    metrics=[faithfulness, answer_relevancy])
print(result)

Related in the Vaaani stack

Have a project that needs Reranking?

30-min discovery call. You describe the busywork; I map it to an AI worker and a budget.