A comprehensive guide on LLM applications, covering LangChain, LlamaIndex, and HuggingGPT for developers.
A knowledge-sharing platform about large language models for job interviews and general understanding.
A comprehensive open-source tutorial on large-scale pre-trained language models covering theory and practical applications.
An open-source LLM evaluation framework for testing and evaluating large language model outputs.
A research project assessing and aligning the values of Chinese large language models focusing on safety and responsibility.
LettuceDetect is a hallucination detection framework for RAG applications.
一个包含213个中文大模型的评测平台,提供评分排行榜和模型输出结果。
A pytest plugin for running and analyzing LLM evaluation tests.
Open-source platform for debugging, evaluating, and monitoring LLM applications with comprehensive tracing and automated evaluations.
A collaborative repository for sharing examples showcasing Mistral models and tools developed by the community.
A reflection agent that uses a two-agent system to validate and improve outputs in AI applications.
Open source LLM engineering platform for observability, metrics, evals, and prompt management.