Specification and documentation for the Model Context Protocol.
A Datacenter Scale Distributed Inference Serving Framework for generative AI and reasoning models.
Run your own AI cluster at home with everyday devices.
Genspark2API is a deployment tool for AI applications with various integration and configuration options.
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
Phoenix is an open-source AI observability platform for experimentation, evaluation, and troubleshooting.
A unified toolkit for automatic evaluations of large language models (LLMs).
An open-source project for comparing two LLMs head-to-head with a given prompt, focusing on backend integration.