
Puppeteer Scraper
Pricing
Pay per usage

Puppeteer Scraper
Crawls websites with the headless Chrome and Puppeteer library using a provided server-side Node.js code. This crawler is an alternative to apify/web-scraper that gives you finer control over the process. Supports both recursive crawling and list of URLs. Supports login to website.
5.0 (5)
Pricing
Pay per usage
154
Total users
7.3K
Monthly users
897
Runs succeeded
>99%
Issues response
41 days
Last modified
5 days ago
You can access the Puppeteer Scraper programmatically from your own applications by using the Apify API. You can also choose the language preference from below. To use the Apify API, you’ll need an Apify account and your API token, found in Integrations settings in Apify Console.
{ "mcpServers": { "local-actors-mcp-server": { "command": "npx", "args": [ "-y", "@apify/actors-mcp-server", "--actors", "apify/puppeteer-scraper" ], "env": { "APIFY_TOKEN": "<YOUR_API_TOKEN>" } } }}
Configure MCP server with Puppeteer Scraper for headless Chrome
You can interact with the MCP server via standard input/output - stdio (as shown above), which is ideal for local integrations and command-line tools such as the Claude desktop client, or you can interact with the server through Server-Sent Events (SSE) to send messages and receive responses, which looks as follows:
{ "mcpServers": { "remote-actors-mcp-server": { "type": "sse", "url": "https://mcp.apify.com/sse?actors=apify/puppeteer-scraper", "headers": { "Authorization": "Bearer <YOUR_API_TOKEN>" } } }}
You can connect to the Apify MCP Server using clients like Tester MCP Client, or any other supported MCP client of your choice.
If you want to learn more about our Apify MCP implementation, check out our MCP documentation. To learn more about the Model Context Protocol in general, refer to the official MCP documentation or read our blog post.