The AI Tool Watch API is a set of static JSON files generated from the same canonical data as the site. There is no authentication, no rate limiting, and no registration required.
Base URL: https://aitool.watch/api/v1/
Start at index.json for a manifest of all available files.
| File | Description |
|---|---|
capabilities.json |
Capabilities with search terms, definitions, what-counts/what-doesn't, and cross-links to implementations |
products.json |
Products (hosted and runtime) with provider links, pricing pages, and implementation lists |
implementations.json |
Implementations with full plan/surface availability, gating, and evidence cross-links |
providers.json |
Providers with logo, website, and status page URLs |
model-access.json |
Open-model records with deployment modes, constraints, and runtime information |
evidence.json |
Evidence records with sources, changelog, and verification dates |
Check index.json for live counts — these change with each verification cycle.
| File | Description |
|---|---|
capability-matrix.json |
Capability x product availability grid with best-gating per cell |
product-comparisons.json |
Pairwise capability overlap for all hosted product pairs |
plan-entitlements.json |
Per-product breakdown of what each subscription plan unlocks |
Every JSON file includes a meta.generated timestamp showing when it was built. Entity records include their own freshness fields:
verified— date when a human last confirmed the data is accuratechecked— date when the automated verification cascade last reviewed itlaunched— date when the feature originally launched
Consumers should check these dates before presenting data as current. The AI landscape changes rapidly. Data older than 30 days should be treated as potentially stale.
- File paths under
/api/v1/are stable. New files may be added but existing files will not be removed or renamed withinv1. - Field names within JSON files are stable. New fields may be added but existing fields will not be removed or renamed within
v1. - If a breaking change is needed, it will ship under
/api/v2/withv1preserved for a transition period. - IDs (capability IDs, product IDs, implementation IDs) are stable and safe to use as foreign keys.
The MCP server (scripts/mcp-server.js) exposes 15 read-only tools over stdio JSON-RPC. Configure it via mcp.json in the project root.
Note: The MCP server was renamed from
ai-capability-referencetoai-tool-watchin v4.0.0. Update yourmcp.jsonconfiguration if upgrading.
| Tool | Description |
|---|---|
list_capabilities |
List all capabilities with IDs, names, and groups |
get_capability |
Full details for a capability by ID |
list_products |
List products, optionally filter by kind (hosted/runtime) |
get_product |
Full details for a product including plan names |
compare_products |
Pairwise capability overlap between two hosted products |
check_availability |
Whether a product implements a capability, with gating and plan details |
search |
Keyword search across capabilities, products, implementations, and open models |
get_plan |
Capabilities and implementations included in a specific product plan (now with pricing) |
list_providers |
All providers with websites, status pages, and product lists |
get_provider |
Full details for a provider by ID |
find_products_by_capabilities |
Find products that support ALL of the specified capabilities |
get_evidence |
Verification sources, changelog, and freshness dates for an implementation |
list_model_access |
All open/self-hostable models with deployment modes and related capabilities |
get_model_access |
Full details for an open model by ID |
get_staleness_report |
Evidence records whose verified date exceeds a threshold (default 30 days) |
Error responses follow the Graceful Boundaries pattern. Every error includes four fields:
error— machine-parseable category (e.g.not_found,invalid_input)detail— human-readable explanationwhy— reason the constraint existsguidance— constructive next step: which tool to call,did_you_meansuggestions, or valid IDs
This means an agent that passes an incorrect ID will receive a suggestion of the correct ID and a pointer to the list tool, rather than a bare error string.
If you are building an agent or tool that reads this data:
-
Always include freshness context. When presenting data to users, include the
verifieddate so they know how current it is. Example: "As of March 2026, ChatGPT Search is available on the Free plan." -
Respect gating and constraints. A capability being "available" on a product does not mean it's available to everyone. Check the
gatingfield (free/paid/limited) and theplansarray for specifics. -
Link back to sources. Evidence records include
sourcesarrays with URLs. When citing specific claims, include these links so users can verify. -
Don't strip caveats. Implementation records include
talking_pointfields that contain important context (regional restrictions, plan requirements, deprecation notices). Omitting these can mislead users. -
Cache responsibly. The data updates roughly weekly. Caching for 24 hours is reasonable. Caching for months defeats the purpose of freshness tracking.
This data is published under the MIT license. You may use it freely, but we appreciate attribution:
Data from AI Tool Watch by PAICE.work.
This API is served as static files from GitHub Pages. There is no rate limit enforced, but please be reasonable:
- Don't poll for changes more than once per hour
- Cache responses locally rather than re-fetching on every request
- If you need real-time change notification, watch the GitHub repository for commits instead
- This is read-only data, not a live service. There is no query API — download the JSON files and filter locally.
- The data covers a curated set of subscription AI products and major open models. It is not exhaustive.
- Plan pricing, surface availability, and feature status change frequently. Always check the
verifieddate. - Regional availability data is incomplete for most products. Absence of regional restrictions does not guarantee global availability.