A comprehensive open-source tutorial on large-scale pre-trained language models covering theory and practical applications.
EgoLife is an egocentric AI project for capturing and understanding multimodal daily activities using advanced technology.
Chinese safety prompts for evaluating and improving the safety of LLMs.
A guide for understanding and mitigating prompt attacks on large language models.
Robust Speech Recognition via Large-Scale Weak Supervision
A pytest plugin for running and analyzing LLM evaluation tests.
科技爱好者周刊,汇集每周五发布的科技内容与资源。
A comprehensive guide on prompt engineering techniques for obtaining high-quality answers from ChatGPT.
Notes for software engineers getting up to speed on new AI developments, including resources and frameworks.
The official GitHub page for the survey paper "A Survey on Evaluation of Large Language Models".
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.