Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.autousers.ai/llms.txt

Use this file to discover all available pages before exploring further.

The Autousers MCP server lets you run UX evaluations without leaving your AI coding tool. Connect once and you can create evaluations, queue autousers against live URLs, read results, and calibrate personas — all from Claude, Cursor, VS Code, ChatGPT, or any MCP-aware client. The server is hosted at https://mcp.autousers.ai/mcp; no package installation is required for most clients.

Connect

Authentication

OAuth is recommended for interactive sessions (Claude.ai, Claude Desktop, Cursor, VS Code, ChatGPT). API keys are the right choice for CI pipelines and headless scripts.

OAuth 2.1

When you add the server URL to a supported client, an OAuth 2.1 PKCE flow launches in the browser. After you approve, the client receives a short-lived access token (~15 minutes) that rotates automatically via a refresh token. No static credentials are stored on the client.

API keys

API keys are long-lived bearer tokens for headless environments. Pass them as Authorization: Bearer ak_live_... — either as an HTTP header in direct-URL clients or as the AUTOUSERS_API_KEY environment variable for stdio. Mint keys at app.autousers.ai/settings/api-keys. Keys are shown once at creation — store them in a secrets manager immediately.

Scopes

ScopeGrants
templates:readList and fetch templates
templates:writeCreate, update, delete, duplicate templates
evaluations:readList, fetch, export, view results
evaluations:writeCreate, update, delete, share evaluations
autousers:readList autousers, fetch runs and telemetry
autousers:writeCreate, update, delete, run, calibrate autousers
ratings:readList ratings for an evaluation
webhooks:readList endpoints, deliveries, and inspect signatures
webhooks:writeCreate, rotate, retry, and delete webhook endpoints
events:readList the event log feeding webhook deliveries

Tools

The server exposes 39 tools across four categories. Read-only tools carry readOnlyHint=true and are safe to call without side effects.

Templates (6)

ToolDescription
templates_listList question templates available to your team
templates_getFetch a single template by ID
templates_createCreate a new team-scoped template
templates_updatePatch a template (only supplied fields change)
templates_deleteHard-delete a template
templates_duplicateDeep-clone a template into a destination team

Evaluations (14)

ToolDescription
evaluations_listList evaluations the caller can see
evaluations_getFetch one evaluation including config and comparisons
evaluations_createCreate an SSE or SxS evaluation; optionally queue autouser runs
evaluations_updatePatch fields on an evaluation
evaluations_deleteDelete an evaluation and its dependent rows
evaluations_save_draftMerge wizard fields into a draft evaluation
evaluations_ratings_listList human and autouser ratings
evaluations_results_getAggregate stats and per-rater summaries
evaluations_agreement_getPairwise Cohen’s Kappa inter-rater agreement
evaluations_ai_insights_getAI-authored summary, key findings, and recommendations
evaluations_export_getDownload results as JSON or CSV
evaluations_share_createGrant a per-user VIEWER / EDITOR / OWNER share
evaluations_shares_listList explicit per-user shares
evaluations_transferTransfer evaluation ownership to another user

Autousers (15)

ToolDescription
autousers_listList autousers (built-in + custom)
autousers_getFetch a single autouser by ID
autousers_createCreate a team-scoped custom autouser
autousers_updatePatch a custom autouser
autousers_deleteSoft-delete a custom autouser
autousers_duplicateDeep-clone an autouser into a team
autousers_runQueue autouser runs against an evaluation
autousers_run_stopCancel pending or running autouser runs
autouser_status_getRun statuses and summary counts for an evaluation
autouser_run_getFetch one autouser run with full context
autouser_run_turns_listPer-turn token and cost telemetry for a run
autousers_calibration_startCompute Cohen’s Kappa vs human ratings
autousers_calibration_status_getGet calibration status
autousers_calibration_freezeFreeze a rubric version and set it as active
autousers_calibration_optimizeSend disagreements to AI for rubric suggestions

Settings (4)

ToolDescription
teams_listList every team the caller belongs to
teams_getFetch one team’s detail
team_members_listList team members with roles and profiles
usage_getUsage rollup: free-run pool remaining and token spend

Prompts

Five canned workflows are registered as MCP prompts. Invoke them from your client’s slash menu — the server pre-fetches relevant data and guides the AI through each workflow end-to-end.
PromptDescription
evaluate-urlCreate an SSE evaluation against a URL and surface results
compare-designsCreate an SxS evaluation between two URLs
analyze-resultsSummarise evaluation results with key findings
calibrate-autouserRun calibration and freeze the rubric when stable
triage-low-agreementSurface autouser/human disagreements and suggest fixes

Resources

Three readable resource URIs let your AI client fetch structured data directly:
URIDescription
autousers://evaluation/{id}Full evaluation object including results
autousers://template/{id}Template definition with all dimensions
autousers://autouser/{id}Autouser persona with rubric

Configuration

VariableDefaultDescription
AUTOUSERS_API_KEYBearer token for headless/stdio auth (ak_live_...)
AUTOUSERS_BASE_URLhttps://app.autousers.aiOverride the API host (for self-hosted or local dev)

Diagnostics

Run the bundled doctor command to verify your configuration:
npx -y -p @autousers/mcp autousers-mcp-doctor
It checks that AUTOUSERS_API_KEY is set (for stdio auth), that the API host is reachable, and that at least one tool call succeeds. The command exits non-zero on any misconfiguration, so it integrates cleanly into CI pre-flight checks.