Official repository for the paper "Enhancing the Code Debugging Ability of LLMs via Communicative Agent Based Data Refinement".
This paper presents a benchmark, DebugEval, which is used to evaluate the code debugging ability of LLMs (Large Language Models) and proposals a framework for building training data using multiple agents, MASTER.
DebugEval designs four task scenarios: BUG Localization, BUG Identification, Code Repair, and Code Review to comprehensively evaluate the code debugging capability of LLMs.
MASTER is a framework for making use of multiple agents working together to refine training data to improve code debugging capability in LLMs.
You can clone the repository using the following command:
git clone DebugEval
cd DebugEval
Download the dataset we provide.
cd src
Please refer to src/README.md
for more details.
We use DeepSeek-Coder-6.7B-Ins and Llama3-8B-Ins as the base model, and train the models with MASTER framework.
cd neural_compiler
Please refer to neural_compiler/README.md
for more details.
cd LLaMA-Factory
Please refer to LLaMA-Factory/README.md
for more details.
We provide the trained NeuDebugger models.
Please cite the paper and star the repo if you use DebugEval and find it helpful.
Feel free to contact [email protected] or open an issue if you have any questions.
@article{DebugEval2024,
title={Enhancing the Code Debugging Ability of LLMs via Communicative Agent Based Data Refinement},
author={Weiqing Yang, Hanbin Wang, Zhenghao Liu, Xinze Li, Yukun Yan, Shuo Wang, Yu Gu, Minghe Yu, Zhiyuan Liu and Ge Yu},
year={2024},
}