GET/api/crawl/:id

Get Job Status

Retrieve the current status and metadata of a crawl job. For completed jobs with 100 pages or fewer, the response also includes all crawled page data inline.

Request

curl
curl https://webextract.mabai.tech/api/crawl/{jobId} \
  -H "Authorization: Bearer cai_sk_••••"

Response

200 OK — running job
{
  "success": true,
  "job": {
    "id": "550e8400-…",
    "url": "https://docs.acme.com",
    "status": "running",
    "pagesTotal": 42,
    "pagesFinished": 17,
    "creditsUsed": 0,
    "createdAt": "2026-03-15T10:00:00Z"
  },
  "pages": null
}

When status is completed and the job has ≤100 pages, the pages array is included inline. For larger jobs use GET /api/crawl/:id/pages.

Status values

pendingJob created, not yet submitted to crawler.
runningCrawl is actively in progress.
completedAll pages crawled and stored successfully.
erroredCrawl failed. Check errorMessage field.
cancelled_by_userCancelled via DELETE /api/crawl/:id.
cancelled_due_to_timeoutJob exceeded the maximum allowed duration.
cancelled_due_to_limitsJob exceeded page or credit limits.

Related