@@ -109,13 +109,13 @@ def get_metric_description(metric_name: str) -> str:
109109 """
110110 descriptions = {
111111 'faithfulness' : "How well grounded the response is in the retrieved context" ,
112- 'answer_relevancy' : "How relevant the response is to the original query" ,
113112 'context_relevancy' : "How relevant the retrieved context is to the query" ,
113+ 'answer_relevancy' : "How relevant the response is to the original query" ,
114114 'context_precision_without_reference' : "Precision of context retrieval without reference answers" ,
115- 'answer_similarity' : "Semantic similarity between generated and reference answers" ,
116- 'answer_correctness' : "Factual accuracy against reference answers" ,
115+ 'context_recall' : "How well retrieved contexts cover the reference answer" ,
117116 'context_precision' : "More accurate precision using reference answers" ,
118- 'context_recall' : "How well retrieved contexts cover the reference answer"
117+ 'answer_similarity' : "Semantic similarity between generated and reference answers" ,
118+ 'answer_correctness' : "Factual accuracy against reference answers"
119119 }
120120
121121 return descriptions .get (metric_name , "Evaluation metric" )
@@ -154,13 +154,13 @@ def get_metric_help_text(metric_name: str) -> str:
154154 """
155155 help_texts = {
156156 'faithfulness' : "Measures how well the generated answer is supported by the retrieved context. Higher scores indicate better factual consistency." ,
157- 'answer_relevancy' : "Evaluates how well the answer addresses the original question. Higher scores indicate more relevant responses." ,
158157 'context_relevancy' : "Assesses the relevance of retrieved context to the query. Higher scores indicate better context retrieval." ,
158+ 'answer_relevancy' : "Evaluates how well the answer addresses the original question. Higher scores indicate more relevant responses." ,
159159 'context_precision_without_reference' : "Measures precision of context retrieval without requiring reference answers. Higher scores indicate more precise retrieval." ,
160- 'answer_similarity' : "Compares semantic similarity between generated and reference answers. Higher scores indicate closer alignment." ,
161- 'answer_correctness' : "Evaluates factual accuracy against reference answers. Higher scores indicate better correctness." ,
160+ 'context_recall' : "Measures how well retrieved contexts cover information in the reference answer. Higher scores indicate better coverage." ,
162161 'context_precision' : "More accurate precision measurement using reference answers for comparison. Higher scores indicate better precision." ,
163- 'context_recall' : "Measures how well retrieved contexts cover information in the reference answer. Higher scores indicate better coverage."
162+ 'answer_similarity' : "Compares semantic similarity between generated and reference answers. Higher scores indicate closer alignment." ,
163+ 'answer_correctness' : "Evaluates factual accuracy against reference answers. Higher scores indicate better correctness."
164164 }
165165
166166 return help_texts .get (metric_name , "RAGAS evaluation metric" )
0 commit comments