| |
| |
| |
|
|
| from typing import Any, Dict, List |
|
|
| from haystack import Document, component |
|
|
|
|
| @component |
| class DocumentMAPEvaluator: |
| """ |
| A Mean Average Precision (MAP) evaluator for documents. |
| |
| Evaluator that calculates the mean average precision of the retrieved documents, a metric |
| that measures how high retrieved documents are ranked. |
| Each question can have multiple ground truth documents and multiple retrieved documents. |
| |
| `DocumentMAPEvaluator` doesn't normalize its inputs, the `DocumentCleaner` component |
| should be used to clean and normalize the documents before passing them to this evaluator. |
| |
| Usage example: |
| ```python |
| from haystack import Document |
| from haystack.components.evaluators import DocumentMAPEvaluator |
| |
| evaluator = DocumentMAPEvaluator() |
| result = evaluator.run( |
| ground_truth_documents=[ |
| [Document(content="France")], |
| [Document(content="9th century"), Document(content="9th")], |
| ], |
| retrieved_documents=[ |
| [Document(content="France")], |
| [Document(content="9th century"), Document(content="10th century"), Document(content="9th")], |
| ], |
| ) |
| |
| print(result["individual_scores"]) |
| # [1.0, 0.8333333333333333] |
| print(result["score"]) |
| # 0.9166666666666666 |
| ``` |
| """ |
|
|
| |
| @component.output_types(score=float, individual_scores=List[float]) |
| def run( |
| self, ground_truth_documents: List[List[Document]], retrieved_documents: List[List[Document]] |
| ) -> Dict[str, Any]: |
| """ |
| Run the DocumentMAPEvaluator on the given inputs. |
| |
| All lists must have the same length. |
| |
| :param ground_truth_documents: |
| A list of expected documents for each question. |
| :param retrieved_documents: |
| A list of retrieved documents for each question. |
| :returns: |
| A dictionary with the following outputs: |
| - `score` - The average of calculated scores. |
| - `individual_scores` - A list of numbers from 0.0 to 1.0 that represents how high retrieved documents |
| are ranked. |
| """ |
| if len(ground_truth_documents) != len(retrieved_documents): |
| msg = "The length of ground_truth_documents and retrieved_documents must be the same." |
| raise ValueError(msg) |
|
|
| individual_scores = [] |
|
|
| for ground_truth, retrieved in zip(ground_truth_documents, retrieved_documents): |
| average_precision = 0.0 |
| average_precision_numerator = 0.0 |
| relevant_documents = 0 |
|
|
| ground_truth_contents = [doc.content for doc in ground_truth if doc.content is not None] |
| for rank, retrieved_document in enumerate(retrieved): |
| if retrieved_document.content is None: |
| continue |
|
|
| if retrieved_document.content in ground_truth_contents: |
| relevant_documents += 1 |
| average_precision_numerator += relevant_documents / (rank + 1) |
| if relevant_documents > 0: |
| average_precision = average_precision_numerator / relevant_documents |
| individual_scores.append(average_precision) |
|
|
| score = sum(individual_scores) / len(ground_truth_documents) |
| return {"score": score, "individual_scores": individual_scores} |
|
|