rageval icon indicating copy to clipboard operation
rageval copied to clipboard

Evaluation tools for Retrieval-augmented Generation (RAG) methods.

Results 27 rageval issues
Sort by recently updated
recently updated
newest added

The `rewrite` task is to convert each input question into a list of queries, which are more friendly to search systems.

enhancement

Enhance the base metric for robust evaluation: 1. add attribute: ```self.task``` to distinguish input format. 2. add function: ```def _validate_data(self, input: Dataset) -> bool``` to check the validity of the...

enhancement

subtasks: - [x] #24 - [ ] #25 - [ ] #26 - [x] #27 - [ ] #28 - [ ] #29 - [ ] #30

enhancement

@FBzzh you can list all potential metrics for the `compress` task in this issue.

enhancement

subtasks in benchmark initialization: - [ ] #14 - [ ] #9 - [ ] #10 - [ ] #11 - [ ] #12 - [ ] #13

enhancement

Here, the `rank` task includes both the retrieval stage and the re-rank stage in the searching process.

enhancement

This task is to evaluate the answer groundedness for the generated `answers`.

enhancement

standardize the api of basellm in rageval/models/base.py.

enhancement