
We survey, evaluate, and open source SOTA Retrieval-Augmented Generation (RAG) algorithms for LLM customization and reasoning. We offer a comprehensive evaluation on each component of GraphRAG, including graph construction (time), knowledge retrieval (time), answer generation (accuracy), and rationale generation (reasoning). We aim to provide unprecedented insights into how graph-structured knowledge enhances LLMs' reasoning capabilities compared to traditional RAG approaches.
GraphRAG-Bench: Challenging Domain-Specific Reasoning for Evaluating Graph Retrieval-Augmented Generation.
GraphRAG-Bench contains 1,018 college-level question spans 16 disciplines, e.g., computer vision, computer networks, human-computer interaction, AI ethics, etc., featuring the ability of conceptual understanding, e.g., "Given [theorem] A and B, prove [conclusion] C", complex algorithmic programming, e.g., coding with interlinked function calls, and mathematical computation, e.g., "Given [Input], [Conv1], [MaxPool], [FC], calculate the output volume dimensions."
GraphRAG-Bench contains five types of diverse questions to thoroughly evaluate different aspects of reasoning, including true-or-false (TF), multiple-choice (MC), multi-select (MS), fill-in-blank (FB), and open-ended (OE).
If you find this website helpful, welcome to cite our papers:
@article{zhang2025survey, title={A Survey of Graph Retrieval-Augmented Generation for Customized Large Language Models}, author={Zhang, Qinggang and Chen, Shengyuan and Bei, Yuanchen and Yuan, Zheng and Zhou, Huachi and Hong, Zijin and Dong, Junnan and Chen, Hao and Chang, Yi and Huang, Xiao}, journal={arXiv preprint arXiv:2501.13958}, year={2025} }
@article{xiao2025graphrag, title={GraphRAG-Bench: Challenging Domain-Specific Reasoning for Evaluating Graph Retrieval-Augmented Generation}, author={Xiao, Yilin and Dong, Junnan and Zhou, Chuang and Dong, Su and Zhang, Qianwen and Yin, Di and Sun, Xing and Huang, Xiao}, journal={arXiv preprint arXiv:2506.02404}, year={2025} }