LogoAISecKit
icon of Awesome LLMs Evaluation Papers

Awesome LLMs Evaluation Papers

A comprehensive collection of papers focused on evaluating large language models (LLMs).

Introduction

Awesome LLMs Evaluation Papers

This repository provides a curated list of papers organized according to the survey Evaluating Large Language Models: A Comprehensive Survey.

Key Features
  • Comprehensive coverage of evaluation methodologies across various aspects of LLMs.
  • Categorized papers including Knowledge and Capability Evaluation, Alignment Evaluation, and Safety Evaluation.
  • Includes benchmarks and leaderboards for LLM performance.
  • Regular updates with new research contributions.
Benefits
  • Serves as a valuable resource for researchers and practitioners in the field of AI and LLMs.
  • Facilitates a better understanding of the capabilities and risks associated with large language models.
  • Promotes community involvement in maintaining and expanding the paper list.
Highlights
  • Authors include recognized contributors from Tianjin University and other institutions.
  • Encourages citation and feedback to enhance the resource.

Newsletter

Join the Community

Subscribe to our newsletter for the latest news and updates