The Oncrawl MCP Server exposes a set of tools that map to Oncrawl capabilities.
These tools allow AI clients to retrieve data, run analyses, and perform operations.
Context and account discovery
get_context_user
Returns the current user context, including:
workspaces
the last active project for each workspace
the last active crawl for each project
the crawl configuration for the last active crawl
This is useful when a user wants a quick overview of their Oncrawl environment before querying deeper.
Workspace and project management
workspace_operations
Supports:
listget
Use this tool to retrieve workspace information.
project_operations
Supports:
listgetcreatedelete
Use this tool to manage Oncrawl projects.
Project creation requires:
namestart_url
A project does not automatically create a crawl configuration or launch a crawl. To start crawling, you must create a crawl configuration and then run a crawl.
Crawl configuration and crawl management
crawl_config_operations
Supports:
listgetcreateupdatedelete
Use this tool to manage crawl configurations for a project.
A crawl configuration requires these mandatory fields:
namestart_urluser_agentmax_speedmax_depthmax_url
crawl_operations
Supports:
listgetrunstopprogressget_commandscommand
Use this tool to list crawls, start a crawl from a project and config, stop a crawl, inspect progress, or retrieve available crawl commands.
To run a crawl, you need:
project_idconfig_id
If a project does not yet have a crawl configuration, you must create one first.
scheduled_crawl_operations
Supports:
listgetcreateupdatedelete
Use this tool to manage recurring crawls for a project.
A scheduled crawl requires:
config_idstart_daterecurrence
Optional:
allow_concurrent_crawls
Crawl-over-crawl management
crawl_over_crawl_operations
Supports:
listgetcreatedelete
Use this tool to manage crawl-over-crawl comparisons.
Creating a comparison requires:
project_idcrawl1crawl2
Optional configuration supports URL transformation rules, such as prefix replacement.
Crawl, log, and comparison data retrieval
The MCP server distinguishes between search tools for retrieving rows and aggregate tools for grouped statistics.
Search tools
get_data_search_crawl
Searches crawl data for:
pageslinksclustersstructured_data
Use this when you want raw matching results rather than summaries.
get_data_search_coc
Searches crawl-over-crawl data.
Use this when you want matching URLs or pages from a specific comparison.
get_data_search_logs
Searches project log data for:
pagesevents
When querying pages, a granularity is required:
daysweeksmonths
Use this when you need raw log rows rather than aggregated metrics.
Aggregation tools
get_data_aggregate_crawl
Runs aggregations on crawl datasets.
Supported crawl data types:
pageslinksclustersstructured_data
Use this for grouped counts, averages, sums, cardinality, and similar statistics.
get_data_aggregate_coc
Runs aggregations on crawl-over-crawl page data.
Use this for questions such as:
how many pages are new, removed, or matching
how many pages changed status code
average delta metrics across matching pages
get_data_aggregate_logs
Runs aggregations on log datasets.
Supported log data types:
pagesevents
For page-level logs, granularity is required:
daysweeksmonths
Use this for grouped counts, summaries, and time-based log analysis.
Field discovery and OQL support
get_fields_crawl
Returns available fields for a crawl data type.
get_fields_coc
Returns available fields for a crawl-over-crawl dataset.
get_fields_logs
Returns available fields for log data.
These tools are especially useful before writing advanced filters or selecting
response fields.
get_oql_documentation
Returns documentation for OQL (Oncrawl Query Language).
Supported topics include:
syntax
examples
field filters
logical operators
data types
data-type-specific examples
Use this tool when building or explaining OQL filters.
URL inspection
url_details_operations
Supports:
get_url_detailsget_url_headersget_url_body
Use this tool to inspect a URL from a crawl, including:
fetch date
response headers
response body
status code
This is useful for troubleshooting page-level crawl behavior.
Exports, secrets, and related operations
data_export_operations
Supports:
listgetcreate
Use this tool to manage workspace data exports.
Supported export targets include:
S3
GCS
Supported output formats include:
JSON
CSV
Parquet
secret_operations
Supports:
listcreatedelete
Use this tool to manage workspace secrets used by export targets.
Supported secret types include:
gcs_credentialss3_credentials
Ranking performance
ranking_performance_operations
Supports:
fieldsaggregations
Use this tool to retrieve available ranking fields or run aggregate queries on ranking performance data for a project.
Search vs. aggregation tools
When working with Oncrawl data through MCP, it helps to choose the right type of tool.
Use a search tool when you want:
a list of matching pages
a list of links
raw crawl-over-crawl results
raw log rows
Use an aggregation tool when you want:
counts by status code
average load time
grouped summaries
distributions by template, URL type, or time period
Limits and query behavior
Some data search tools return paginated results and enforce limits.
For these tools:
get_data_search_crawlget_data_search_cocget_data_search_logs
the maximum result limit is 20 per request.
If you need more results, use pagination with offsets.
