No Cloning Required
Install the CLI globally with npm. No need to clone repositories or navigate complex folder structures.
A reproducible, standardized workflow for testing AI tools by building a real application

Get up and running in under 2 minutes:
# Install the CLI
npm install -g pawmate-ai-challenge
# Create a project directory
mkdir my-pawmate-benchmark
cd my-pawmate-benchmark
# Initialize a benchmark run
pawmate init --profile model-a-rest --tool "YourAI" --tool-ver "1.0"
# Copy the generated prompts to your AI agent
cat pawmate-run-*/start_build_api_prompt.txtThat's it! Your AI agent will build the entire application from the prompt.
Learn more about the workflow →
PawMate is a benchmarking framework that evaluates AI coding assistants by having them build a complete, production-ready application from a frozen specification. Unlike simple coding tests, PawMate measures how well AI tools can:
Basic CRUD operations with lifecycle management:
Everything in Model A plus:
Choose between REST or GraphQL when initializing your run:
# REST API
pawmate init --profile model-a-rest --tool "YourAI"
# GraphQL API
pawmate init --profile model-a-graphql --tool "YourAI"Both styles implement the same functional requirements, allowing fair comparisons between tools regardless of API preference.
The benchmark captures objective, automated metrics:
No subjective scoring. All metrics are automatically extracted from the AI agent's run artifacts.
The specification requires a specific, battle-tested stack:
Backend:
Frontend (Optional):
This ensures runs are comparable and reproducible across all participants.
Submit your results in two ways:
Email (Default):
pawmate submit pawmate-run-*/benchmark/result.json
# Opens email client with pre-filled contentGitHub Issue (Optional):
export GITHUB_TOKEN=your-token-here
pawmate submit pawmate-run-*/benchmark/result.json
# Creates issue + opens email clientStart here: Getting Started Guide →
Check out the FAQ or view the spec on GitHub.