code-rag-bench
code-rag-bench copied to clipboard
CodeRAG-Bench: Can Retrieval Augment Code Generation?
Hi there, Thank you for the excellent work and for publishing the code base. I am attempting to reproduce the retrieval performance of BGE-base as shown in Table 3 but...
Hello, We have come across your work on Code Retrieval. Can you benchmark our code embedding models, [CodeSage](https://huggingface.co/codesage)? Thank you.
I want to run benchmark with my custom search pipeline (for example add bm25 and embedding vector search with LLM reranker). Are there any simple ways to insert it into...
Your project doesn't appear to specify an open source license, is it possible to add one? Thanks.
Hi. I really appreciate your helpful contribution in this field. I have always wondered why Code RAG does not have a well-composed benchmark, and you are the one who resolved...
During the process of using your repository, I encountered some issues that I would like to bring to your attention. While attempting to build and evaluate the code, I ran...
After downloading the repositories for RepoEval (e.g., under retrieval/output/repoeval/repositories/function_level/), obtaining retrieval results (e.g., /path/to/retrieval/results/retriever-name.jsonl), **and obtaining the code generation outputs in previous steps (e.g., /path/to/generation/outputs/model-name.json),** we can run execution-based evaluation....
### **Description** Hello, I am trying to run the retrieval script to generate the index and `qrels` file for the `live_code_bench` task, following the project's documentation. The process fails because...
It seems install older vllm version lacks "PoolerConfig" class inside vllm/config.py package. Beir tries to use it and ends up crashing. I tried running the following line: python3 -u eval_beir_sbert_canonical.py...
Potential mismatch between task and test files in `repoeval_task_id2tests.json` for `deepmind_tarcr`
Hi, thank you very much for your excellent work and contributions in this domain! According to the CODERAG-BENCH paper, a reproducible execution environment is provided for RepoEval. The file `code-rag-bench/tree/main/generation/eval/tasks/custom_metrics/repoeval_task_id2tests.json`...