Awesome LLMs Evaluation Papers
This repository provides a curated list of papers organized according to the survey Evaluating Large Language Models: A Comprehensive Survey.
Key Features
- Comprehensive coverage of evaluation methodologies across various aspects of LLMs.
- Categorized papers including Knowledge and Capability Evaluation, Alignment Evaluation, and Safety Evaluation.
- Includes benchmarks and leaderboards for LLM performance.
- Regular updates with new research contributions.
Benefits
- Serves as a valuable resource for researchers and practitioners in the field of AI and LLMs.
- Facilitates a better understanding of the capabilities and risks associated with large language models.
- Promotes community involvement in maintaining and expanding the paper list.
Highlights
- Authors include recognized contributors from Tianjin University and other institutions.
- Encourages citation and feedback to enhance the resource.