Reusable XCrawl skill definitions for multi-agent runtimes, focused on API-first web data workflows.
Canonical repository: https://github.com/xcrawl-api/xcrawl-skills
English | 中文
XCrawl is a web data infrastructure product for search, scraping, URL mapping, and site crawling. This repository provides production-oriented skill definitions that help agents call XCrawl APIs consistently.
- Ready-to-use skills for common XCrawl workflows
- API-oriented instructions with request/response parameter documentation
- cURL and Node examples suitable for runtime execution
- A consistent contract for multi-agent orchestration
xcrawl: Default XCrawl entry skill for direct lookup and single-URL extractionxcrawl-scrape: Single-URL extraction and structured data workflowsxcrawl-map: Site URL discovery and scope planning workflowsxcrawl-crawl: Bulk site crawling and async result handling workflowsxcrawl-search: Query-based discovery with location/language controls
- An XCrawl API key
- Register at
https://dash.xcrawl.com/and activate the free1000credits plan - Runtime binaries:
curlandnode - Access to this repository
Create local config file:
Path: ~/.xcrawl/config.json
{
"XCRAWL_API_KEY": "<your_api_key>"
}Skills in this repo are designed to read XCRAWL_API_KEY from this local file.
Open one of:
skills/xcrawl/SKILL.mdskills/xcrawl-scrape/SKILL.mdskills/xcrawl-map/SKILL.mdskills/xcrawl-crawl/SKILL.mdskills/xcrawl-search/SKILL.md
Each skill includes:
- Applicable scenarios
- Request parameters
- Response parameters
- cURL / Node examples
Use the examples in each SKILL.md directly, then adapt request payloads for your business scenario.
- "Scrape this page in sync mode and return markdown + links."
- "Map only
/docs/URLs under this domain with a limit of 2000." - "Start a bounded crawl (depth 2, limit 100) and poll until completed."
- "Search in US English and return top 20 results for this query."
Each skill can be executed through a runtime adapter layer.
- Input normalization:
goal,inputs,constraints,credentials_ref,runtime_context - Output normalization:
status,request_payload,raw_responseor async pair,task_ids,error
Default behavior is raw passthrough: return upstream API response bodies as-is.
- Base URL:
https://run.xcrawl.com - Local config file:
~/.xcrawl/config.json - API key field in local config:
XCRAWL_API_KEY - Auth header:
Authorization: Bearer <XCRAWL_API_KEY> - Main endpoints:
POST /v1/scrapeandGET /v1/scrape/{scrape_id}POST /v1/mapPOST /v1/crawlandGET /v1/crawl/{crawl_id}POST /v1/search