Subscribe to our newsletter for the latest news and updates
一个包含213个中文大模型的评测平台,提供评分排行榜和模型输出结果。
Explore prompt injection techniques in large language models (LLMs), providing examples to improve LLM security and robustness.
The most comprehensive prompt hacking course available, focusing on prompt engineering and security.
An open-source toolkit for monitoring Large Language Models (LLMs) with features like text quality and sentiment analysis.
Chinese LLM Benchmark 是一个全面的中文大模型评测平台,旨在给用户提供关于各类大模型的详细能力评分和原始输出结果。我们目前已囊括213个大模型,覆盖包括商用模型如ChatGPT、GPT-4、Gemini等,及多个开源模型。