
A comprehensive guide on LLM applications, covering LangChain, LlamaIndex, and HuggingGPT for developers.

A knowledge-sharing platform about large language models for job interviews and general understanding.

A comprehensive open-source tutorial on large-scale pre-trained language models covering theory and practical applications.

An open-source LLM evaluation framework for testing and evaluating large language model outputs.

A research project assessing and aligning the values of Chinese large language models focusing on safety and responsibility.

LettuceDetect is a hallucination detection framework for RAG applications.

一个包含213个中文大模型的评测平台,提供评分排行榜和模型输出结果。

A pytest plugin for running and analyzing LLM evaluation tests.

Open-source platform for debugging, evaluating, and monitoring LLM applications with comprehensive tracing and automated evaluations.

A collaborative repository for sharing examples showcasing Mistral models and tools developed by the community.

A reflection agent that uses a two-agent system to validate and improve outputs in AI applications.