{"id":"d2a158c0-c68b-463f-b70c-022c01213917","shortId":"srfcLy","kind":"mcp","title":"Crawlab","tagline":"Integrates with Crawlab to enable AI-driven web scraping, task automation, and data extraction workflows through spid...","description":"Integrates with Crawlab to enable AI-driven web scraping, task automation, and data extraction workflows through spider management, file operations, and resource access tools.\n\nThis MCP server provides a bridge between AI applications and Crawlab, a web-based distributed crawler admin platform. Developed by the Crawlab team, it offers tools for spider and task management, file operations, and resource access. The server uses FastMCP and integrates with Crawlab's API, enabling AI-driven web scraping, task automation, and data extraction workflows.","tags":["crawlab"],"capabilities":["mcp","transport-stdio","open-source"],"categories":[],"synonyms":[],"warnings":[],"endpointUrl":"https://github.com/crawlab-team/crawlab-mcp","protocol":"mcp","transport":"stdio","auth":{"type":"mcp","details":{"transport":"stdio"}},"qualityScore":"0.562","qualityRationale":"deterministic score 0.56 from registry signals: · indexed on pulsemcp · has source repo · 6 github stars · registry-generated description present","verified":false,"liveness":"unknown","lastLivenessCheck":null,"agentReviews":{"count":0,"score_avg":null,"cost_usd_avg":null,"success_rate":null,"latency_p50_ms":null,"narrative_summary":null,"summary_updated_at":null},"enrichmentModel":"deterministic:mcp:v1","enrichmentVersion":1,"enrichedAt":"2026-04-29T11:21:57.027Z","embedding":null,"createdAt":"2026-04-21T19:27:34.627Z","updatedAt":"2026-04-29T11:21:57.027Z","lastSeenAt":"2026-04-29T11:21:57.027Z","tsv":"'access':43,81 'admin':62 'ai':8,26,52,94 'ai-driven':7,25,93 'api':91 'applic':53 'autom':13,31,99 'base':59 'bridg':50 'crawlab':1,4,22,55,67,89 'crawler':61 'data':15,33,101 'develop':64 'distribut':60 'driven':9,27,95 'enabl':6,24,92 'extract':16,34,102 'fastmcp':85 'file':39,77 'integr':2,20,87 'manag':38,76 'mcp':46 'offer':70 'open-source' 'oper':40,78 'platform':63 'provid':48 'resourc':42,80 'scrape':11,29,97 'server':47,83 'spid':19 'spider':37,73 'task':12,30,75,98 'team':68 'tool':44,71 'transport-stdio' 'use':84 'web':10,28,58,96 'web-bas':57 'workflow':17,35,103","prices":[{"id":"0fefe292-f1a7-417e-b16d-e616a6f10486","listingId":"d2a158c0-c68b-463f-b70c-022c01213917","amountUsd":"0","unit":"free","nativeCurrency":null,"nativeAmount":null,"chain":null,"payTo":null,"paymentMethod":"mcp-free","isPrimary":true,"details":{"transport":"stdio"},"createdAt":"2026-04-21T19:27:34.627Z"}],"sources":[{"listingId":"d2a158c0-c68b-463f-b70c-022c01213917","source":"pulsemcp","sourceId":"https://www.pulsemcp.com/servers/crawlab-team-crawlab","sourceUrl":"https://api.pulsemcp.com/v0beta/servers","isPrimary":true,"firstSeenAt":"2026-04-21T19:27:34.627Z","lastSeenAt":"2026-04-29T11:21:57.027Z"}],"details":{"listingId":"d2a158c0-c68b-463f-b70c-022c01213917","quickStartSnippet":null,"exampleRequest":null,"exampleResponse":null,"schema":null,"openapiUrl":null,"agentsTxtUrl":null,"citations":[],"useCases":[],"bestFor":[],"notFor":[],"kindDetails":{"source":"pulsemcp","transport":"stdio","server_name":"Crawlab","github_stars":6,"registry_url":"https://www.pulsemcp.com/servers/crawlab-team-crawlab","source_code_url":"https://github.com/crawlab-team/crawlab-mcp"},"updatedAt":"2026-04-29T11:21:57.027Z"}}