Contribute to Assay

Help build the quality layer for agentic software. Run evaluations on MCP servers and agent skills with your own tokens, submit results, and help the community find the best tools.

How Community Evaluation Works

1

Pick a Package

Choose from the evaluation queue below, or evaluate a package you use. The queue prioritizes packages with no existing evaluation or outdated scores.

2

Run the Evaluation

Use the Assay evaluation skill or CLI tool. It runs against your LLM tokens, analyzes the package, and produces a structured JSON result.

3

Submit Results

Open a pull request to the Assay repo with your evaluation JSON. Assay reviews and merges quality submissions.

Get Your API Key

Sign in with GitHub to get an API key for submitting evaluations. Your GitHub identity is used for contributor attribution and trust tier progression.

We only request read:user scope — we store your username and avatar, nothing else.

Sign in with GitHub

Trust & Quality

Reproducible Evaluations

All evaluations use Assay's standardized eval configs — deterministic prompts, pinned model versions, and structured output schemas. Results should be reproducible by anyone running the same config.

Cross-Validation

When multiple independent contributors evaluate the same package, agreement between submissions increases confidence. Cross-validated scores carry more weight.

Spot-Check Verification

Assay re-runs approximately 10% of new contributor submissions using our own tokens as a quality gate. This builds trust gradually — established contributors earn higher trust over time.

Anti-Gaming

Package authors evaluating their own tools must disclose the relationship. Undisclosed self-evaluations that significantly diverge from independent evaluations are flagged for review.

Evaluation Queue

10313 packages need evaluation
gcp-storage-mcp Priority AI & Machine Learning

A Model Context Protocol (MCP) server that provides seamless integration with Google Cloud Storage, enabling AI assistants to perform file operations, manage buckets, and interact with GCS resources directly.

Not yet evaluated Repo ↗
mcp-construe Priority AI & Machine Learning

A FastMCP server that loads personal context from Obsidian vaults using frontmatter filtering - curate your knowledge for AI conversations.

Not yet evaluated Repo ↗
Gmail-mcp-server Priority Communication

A resilient MCP server built with fastMCP for sending emails through Gmail's SMTP server using AI agents.

Not yet evaluated Repo ↗
google_workspace_fastmcp2 Priority File Management

FastMCP2 server for Google Drive uploads with Qdrant-powered semantic search

Not yet evaluated Repo ↗
budgetkey-mcp Priority Productivity

A fastmcp server for open budget project

Not yet evaluated Repo ↗
mcp-generator-3.x Priority AI & Machine Learning

Generate FastMCP 3.x Servers from OpenAPI Specs

Not yet evaluated Repo ↗
sefaria-mcp Priority Other

FastMCP Server for Sefaria

Not yet evaluated Repo ↗
mcpskills-cli Priority AI & Machine Learning

Generate Agent Skills from MCP server tools. Connects via Streamable HTTP, discovers tools, and outputs a skill with schema docs and a call script in the language of your choice.

Not yet evaluated Repo ↗
mcp-server-mattermost Priority AI & Machine Learning

MCP server for Mattermost — let Claude, Cursor, and other AI assistants work with channels, messages, and files

Not yet evaluated Repo ↗
aws-athena-mcp Priority Cloud Infrastructure

AWS Athena MCP using FastMCP

Not yet evaluated Repo ↗
fastmcp-agents Priority AI & Machine Learning

FastMCP-powered Agentic Workflows

Not yet evaluated Repo ↗
claude-agent-toolkit Priority AI & Machine Learning

Python framework for building agents using claude-code-sdk with programmable tools

Not yet evaluated Repo ↗
polars-docs-mcp Priority Search

A FastMCP tool to search and retrieve Polars API documentation.

Not yet evaluated Repo ↗
Ludus-FastMCP Priority Developer Tools

Ludus FastMCP enables AI-powered management of Ludus cyber ranges through natural language commands. The server exposes **157 tools** across 15 modules for range lifecycle management, scenario deployment, template creation, Ansible role management, and security monitoring integration.

Not yet evaluated Repo ↗
lucidity-mcp Priority Developer Tools

AI-powered code quality analysis using MCP to help AI assistants review code more effectively. Analyze git changes for complexity, security issues, and more through structured prompts.

Not yet evaluated Repo ↗
XnneHangLab Priority AI & Machine Learning

不会聊天的字幕提取器不是一个好 B 站下载器~

Not yet evaluated Repo ↗
Construction-Hazard-Detection Priority Developer Tools

Enhances construction site safety using YOLO for object detection, identifying hazards like workers without helmets or safety vests, and proximity to machinery or vehicles. HDBSCAN clusters safety cone coordinates to create monitored zones. Post-processing algorithms improve detection accuracy.

Not yet evaluated Repo ↗
MCP-Doc Priority AI & Machine Learning

A powerful Word document processing service based on FastMCP, enabling AI assistants to create, edit, and manage docx files with full formatting support. Preserves original styles when editing content. 基于FastMCP的强大Word文档处理服务,使AI助手能够创建、编辑和管理docx文件,支持完整的格式设置功能。在编辑内容时能够保留原始样式和格式,实现精确的文档操作。

Not yet evaluated Repo ↗
mcp_agent_mail Priority Developer Tools

Asynchronous coordination layer for AI coding agents: identities, inboxes, searchable threads, and advisory file leases over FastMCP + Git + SQLite

Not yet evaluated Repo ↗
swagger-mcp Priority Agent Skills

mcp server which will dynamically define tools based on swagger

Not yet evaluated Repo ↗

Showing 20 of 10313 packages. View full queue via API →

📋

Agent Evaluation Guide

A single document with the complete scoring rubric, JSON schema, and submission instructions. Any AI agent can fetch this URL, evaluate a package, and submit results.

View Evaluation Guide → Rubric v1.0 · Markdown format

Getting Started

Option 1: Use Any AI Agent (Recommended)

Have your AI agent fetch the evaluation guide, evaluate a package from the queue, and submit via the API. Works with Claude, GPT, Gemini, or any agent.

# Your agent fetches the guide and submits results
curl -X POST https://assay.tools/v1/evaluations \
  -H "Content-Type: application/json" \
  -H "X-Api-Key: your-api-key" \
  -d @evaluation.json

Option 2: Assay CLI Tool

Run evaluations locally using Assay's built-in evaluator:

# Clone the repo
git clone https://github.com/Assay-Tools/assay.git
cd assay

# Run evaluation on a specific package
uv run python -m assay.evaluation.evaluator --package <package-id>

# Or batch evaluate discovered packages
uv run python -m assay.evaluation.evaluator --batch --limit 5

Option 3: Request an Evaluation

Know a package that should be in Assay? Open a GitHub issue with the package name and repo URL, and we'll add it to the queue.

4643
Packages Evaluated
10313
Need Evaluation
173
Need Re-evaluation
Community Powered