A comprehensive open-source tutorial on large-scale pre-trained language models covering theory and practical applications.
Chinese safety prompts for evaluating and improving the safety of LLMs.
A research project assessing and aligning the values of Chinese large language models focusing on safety and responsibility.
This repository contains the code for generating the ToxiGen dataset for hate speech detection.
The official GitHub page for the survey paper "A Survey on Evaluation of Large Language Models".
A study evaluating geopolitical and cultural biases in large language models through dual-layered assessments.
Real-time face swap and one-click video deepfake with only a single image.
A guidebook sharing insights and knowledge about evaluating Large Language Models (LLMs).
A comprehensive collection of papers focused on evaluating large language models (LLMs).
Official GitHub repository for SafetyBench, a benchmark to evaluate the safety of large language models (LLMs).
Meridian cuts through news noise by scraping hundreds of sources, analyzing stories with AI, and delivering concise, personalized daily briefs.