
An open-source LLM evaluation framework for testing and evaluating large language model outputs.

LLM Arena by KCORES is a benchmarking platform for programming skills using large models.

A CLI tool to convert your codebase into a single LLM prompt with source tree, prompt templating, and token counting.

An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.

OpenManus is an open-source framework for building general AI agents without an Invite Code.

Toolkit for linearizing PDFs for LLM datasets/training.

A package for parsing PDFs and analyzing their content using LLMs.

Convert PDF to markdown + JSON quickly with high accuracy.

A GitHub repository for practicing large language models (LLMs) with various resources and projects.

Open-Source Chrome extension for AI-powered web automation. Run multi-agent workflows using your own LLM API key.

一个包含213个中文大模型的评测平台,提供评分排行榜和模型输出结果。

A LLM CTF Challenge designed to teach prompt injection in multi-chain LLM applications.