Web Scraper avatar

Web Scraper

Try for free

No credit card required

Go to Store
Web Scraper

Web Scraper

apify/web-scraper
Try for free

No credit card required

Crawls arbitrary websites using the Chrome browser and extracts structured data from web pages using a provided JavaScript function. The Actor supports both recursive crawling and lists of URLs, and automatically manages concurrency for maximum performance.

You can access the Web Scraper programmatically from your own applications by using the Apify API. You can choose the language preference from below. To use the Apify API, you’ll need an Apify account and your API token, found in Integrations settings in Apify Console.

1{
2  "openapi": "3.0.1",
3  "info": {
4    "version": "3.0",
5    "x-build-id": "tfHAuSQ2x4KpkucDF"
6  },
7  "servers": [
8    {
9      "url": "https://api.apify.com/v2"
10    }
11  ],
12  "paths": {
13    "/acts/apify~web-scraper/run-sync-get-dataset-items": {
14      "post": {
15        "operationId": "run-sync-get-dataset-items-apify-web-scraper",
16        "x-openai-isConsequential": false,
17        "summary": "Executes an Actor, waits for its completion, and returns Actor's dataset items in response.",
18        "tags": [
19          "Run Actor"
20        ],
21        "requestBody": {
22          "required": true,
23          "content": {
24            "application/json": {
25              "schema": {
26                "$ref": "#/components/schemas/inputSchema"
27              }
28            }
29          }
30        },
31        "parameters": [
32          {
33            "name": "token",
34            "in": "query",
35            "required": true,
36            "schema": {
37              "type": "string"
38            },
39            "description": "Enter your Apify token here"
40          }
41        ],
42        "responses": {
43          "200": {
44            "description": "OK"
45          }
46        }
47      }
48    },
49    "/acts/apify~web-scraper/runs": {
50      "post": {
51        "operationId": "runs-sync-apify-web-scraper",
52        "x-openai-isConsequential": false,
53        "summary": "Executes an Actor and returns information about the initiated run in response.",
54        "tags": [
55          "Run Actor"
56        ],
57        "requestBody": {
58          "required": true,
59          "content": {
60            "application/json": {
61              "schema": {
62                "$ref": "#/components/schemas/inputSchema"
63              }
64            }
65          }
66        },
67        "parameters": [
68          {
69            "name": "token",
70            "in": "query",
71            "required": true,
72            "schema": {
73              "type": "string"
74            },
75            "description": "Enter your Apify token here"
76          }
77        ],
78        "responses": {
79          "200": {
80            "description": "OK",
81            "content": {
82              "application/json": {
83                "schema": {
84                  "$ref": "#/components/schemas/runsResponseSchema"
85                }
86              }
87            }
88          }
89        }
90      }
91    },
92    "/acts/apify~web-scraper/run-sync": {
93      "post": {
94        "operationId": "run-sync-apify-web-scraper",
95        "x-openai-isConsequential": false,
96        "summary": "Executes an Actor, waits for completion, and returns the OUTPUT from Key-value store in response.",
97        "tags": [
98          "Run Actor"
99        ],
100        "requestBody": {
101          "required": true,
102          "content": {
103            "application/json": {
104              "schema": {
105                "$ref": "#/components/schemas/inputSchema"
106              }
107            }
108          }
109        },
110        "parameters": [
111          {
112            "name": "token",
113            "in": "query",
114            "required": true,
115            "schema": {
116              "type": "string"
117            },
118            "description": "Enter your Apify token here"
119          }
120        ],
121        "responses": {
122          "200": {
123            "description": "OK"
124          }
125        }
126      }
127    }
128  },
129  "components": {
130    "schemas": {
131      "inputSchema": {
132        "type": "object",
133        "required": [
134          "startUrls",
135          "pageFunction",
136          "proxyConfiguration"
137        ],
138        "properties": {
139          "runMode": {
140            "title": "Run mode",
141            "enum": [
142              "PRODUCTION",
143              "DEVELOPMENT"
144            ],
145            "type": "string",
146            "description": "This property indicates the scraper's mode of operation. In DEVELOPMENT mode, the scraper ignores page timeouts, doesn't use sessionPool, opens pages one by one and enables debugging via Chrome DevTools.  Open the live view tab or the container URL to access the debugger. Further debugging options can be configured in the Advanced configuration section. PRODUCTION mode disables debugging and enables timeouts and concurrency. <br><br>For details, see <a href='https://apify.com/apify/web-scraper#run-mode' target='_blank' rel='noopener'>Run mode</a> in README.",
147            "default": "PRODUCTION"
148          },
149          "startUrls": {
150            "title": "Start URLs",
151            "type": "array",
152            "description": "A static list of URLs to scrape. <br><br>For details, see <a href='https://apify.com/apify/web-scraper#start-urls' target='_blank' rel='noopener'>Start URLs</a> in README.",
153            "items": {
154              "type": "object",
155              "required": [
156                "url"
157              ],
158              "properties": {
159                "url": {
160                  "type": "string",
161                  "title": "URL of a web page",
162                  "format": "uri"
163                }
164              }
165            }
166          },
167          "keepUrlFragments": {
168            "title": "URL #fragments identify unique pages",
169            "type": "boolean",
170            "description": "Indicates that URL fragments (e.g. <code>http://example.com<b>#fragment</b></code>) should be included when checking whether a URL has already been visited or not. Typically, URL fragments are used for page navigation only and therefore they should be ignored, as they don't identify separate pages. However, some single-page websites use URL fragments to display different pages; in such a case, this option should be enabled.",
171            "default": false
172          },
173          "linkSelector": {
174            "title": "Link selector",
175            "type": "string",
176            "description": "A CSS selector saying which links on the page (<code>&lt;a&gt;</code> elements with <code>href</code> attribute) shall be followed and added to the request queue. To filter the links added to the queue, use the <b>Pseudo-URLs</b> and/or <b>Glob patterns</b> setting.<br><br>If <b>Link selector</b> is empty, the page links are ignored.<br><br>For details, see <a href='https://apify.com/apify/web-scraper#link-selector' target='_blank' rel='noopener'>Link selector</a> in README."
177          },
178          "globs": {
179            "title": "Glob Patterns",
180            "type": "array",
181            "description": "Glob patterns to match links in the page that you want to enqueue. Combine with Link selector to tell the scraper where to find links. Omitting the Glob patterns will cause the scraper to enqueue all links matched by the Link selector.",
182            "default": [],
183            "items": {
184              "type": "object",
185              "required": [
186                "glob"
187              ],
188              "properties": {
189                "glob": {
190                  "type": "string",
191                  "title": "Glob of a web page"
192                }
193              }
194            }
195          },
196          "pseudoUrls": {
197            "title": "Pseudo-URLs",
198            "type": "array",
199            "description": "Specifies what kind of URLs found by <b>Link selector</b> should be added to the request queue. A pseudo-URL is a URL with regular expressions enclosed in <code>[]</code> brackets, e.g. <code>http://www.example.com/[.*]</code>. <br><br>If <b>Pseudo-URLs</b> are omitted, the actor enqueues all links matched by the <b>Link selector</b>.<br><br>For details, see <a href='https://apify.com/apify/web-scraper#pseudo-urls' target='_blank' rel='noopener'>Pseudo-URLs</a> in README.",
200            "default": [],
201            "items": {
202              "type": "object",
203              "required": [
204                "purl"
205              ],
206              "properties": {
207                "purl": {
208                  "type": "string",
209                  "title": "Pseudo-URL of a web page"
210                }
211              }
212            }
213          },
214          "excludes": {
215            "title": "Exclude Glob Patterns",
216            "type": "array",
217            "description": "Glob patterns to match links in the page that you want to exclude from being enqueued.",
218            "default": [],
219            "items": {
220              "type": "object",
221              "required": [
222                "glob"
223              ],
224              "properties": {
225                "glob": {
226                  "type": "string",
227                  "title": "Glob of a web page"
228                }
229              }
230            }
231          },
232          "pageFunction": {
233            "title": "Page function",
234            "type": "string",
235            "description": "JavaScript (ES6) function that is executed in the context of every page loaded in the Chrome browser. Use it to scrape data from the page, perform actions or add new URLs to the request queue.<br><br>For details, see <a href='https://apify.com/apify/web-scraper#page-function' target='_blank' rel='noopener'>Page function</a> in README."
236          },
237          "injectJQuery": {
238            "title": "Inject jQuery",
239            "type": "boolean",
240            "description": "If enabled, the scraper will inject the <a href='http://jquery.com' target='_blank' rel='noopener'>jQuery</a> library into every web page loaded, before <b>Page function</b> is invoked. Note that the jQuery object (<code>$</code>) will not be registered into global namespace in order to avoid conflicts with libraries used by the web page. It can only be accessed through <code>context.jQuery</code> in <b>Page function</b>.",
241            "default": true
242          },
243          "proxyConfiguration": {
244            "title": "Proxy configuration",
245            "type": "object",
246            "description": "Specifies proxy servers that will be used by the scraper in order to hide its origin.<br><br>For details, see <a href='https://apify.com/apify/web-scraper#proxy-configuration' target='_blank' rel='noopener'>Proxy configuration</a> in README.",
247            "default": {
248              "useApifyProxy": true
249            }
250          },
251          "proxyRotation": {
252            "title": "Proxy rotation",
253            "enum": [
254              "RECOMMENDED",
255              "PER_REQUEST",
256              "UNTIL_FAILURE"
257            ],
258            "type": "string",
259            "description": "This property indicates the strategy of proxy rotation and can only be used in conjunction with Apify Proxy. The recommended setting automatically picks the best proxies from your available pool and rotates them evenly, discarding proxies that become blocked or unresponsive. If this strategy does not work for you for any reason, you may configure the scraper to either use a new proxy for each request, or to use one proxy as long as possible, until the proxy fails. IMPORTANT: This setting will only use your available Apify Proxy pool, so if you don't have enough proxies for a given task, no rotation setting will produce satisfactory results.",
260            "default": "RECOMMENDED"
261          },
262          "sessionPoolName": {
263            "title": "Session pool name",
264            "pattern": "[0-9A-z-]",
265            "minLength": 3,
266            "maxLength": 200,
267            "type": "string",
268            "description": "<b>Use only english alphanumeric characters dashes and underscores.</b> A session is a representation of a user. It has it's own IP and cookies which are then used together to emulate a real user. Usage of the sessions is controlled by the Proxy rotation option. By providing a session pool name, you enable sharing of those sessions across multiple actor runs. This is very useful when you need specific cookies for accessing the websites or when a lot of your proxies are already blocked. Instead of trying randomly, a list of working sessions will be saved and a new actor run can reuse those sessions. Note that the IP lock on sessions expires after 24 hours, unless the session is used again in that window."
269          },
270          "initialCookies": {
271            "title": "Initial cookies",
272            "type": "array",
273            "description": "A JSON array with cookies that will be set to every Chrome browser tab opened before loading the page, in the format accepted by Puppeteer's <a href='https://pptr.dev/#?product=Puppeteer&show=api-pagesetcookiecookies' target='_blank' rel='noopener'><code>Page.setCookie()</code></a> function. This option is useful for transferring a logged-in session from an external web browser. For details how to do this, read this <a href='https://help.apify.com/en/articles/1444249-log-in-to-website-by-transferring-cookies-from-web-browser-legacy' target='_blank' rel='noopener'>help article</a>.",
274            "default": []
275          },
276          "useChrome": {
277            "title": "Use Chrome",
278            "type": "boolean",
279            "description": "If enabled, the scraper will use a real Chrome browser instead of Chromium bundled with Puppeteer. This option may help bypass certain anti-scraping protections, but might make the scraper unstable. Use at your own risk 🙂",
280            "default": false
281          },
282          "headless": {
283            "title": "Run browsers in headless mode",
284            "type": "boolean",
285            "description": "By default, browsers run in headless mode. You can toggle this off to run them in headful mode, which can help with certain rare anti-scraping protections but is slower and more costly.",
286            "default": true
287          },
288          "ignoreSslErrors": {
289            "title": "Ignore SSL errors",
290            "type": "boolean",
291            "description": "If enabled, the scraper will ignore SSL/TLS certificate errors. Use at your own risk.",
292            "default": false
293          },
294          "ignoreCorsAndCsp": {
295            "title": "Ignore CORS and CSP",
296            "type": "boolean",
297            "description": "If enabled, the scraper will ignore Content Security Policy (CSP) and Cross-Origin Resource Sharing (CORS) settings of visited pages and requested domains. This enables you to freely use XHR/Fetch to make HTTP requests from <b>Page function</b>.",
298            "default": false
299          },
300          "downloadMedia": {
301            "title": "Download media files",
302            "type": "boolean",
303            "description": "If enabled, the scraper will download media such as images, fonts, videos and sound files, as usual. Disabling this option might speed up the scrape, but certain websites could stop working correctly.",
304            "default": true
305          },
306          "downloadCss": {
307            "title": "Download CSS files",
308            "type": "boolean",
309            "description": "If enabled, the scraper will download CSS files with stylesheets, as usual. Disabling this option may speed up the scrape, but certain websites could stop working correctly, and the live view will not look as cool.",
310            "default": true
311          },
312          "maxRequestRetries": {
313            "title": "Max page retries",
314            "minimum": 0,
315            "type": "integer",
316            "description": "The maximum number of times the scraper will retry to load each web page on error, in case of a page load error or an exception thrown by <b>Page function</b>.<br><br>If set to <code>0</code>, the page will be considered failed right after the first error.",
317            "default": 3
318          },
319          "maxPagesPerCrawl": {
320            "title": "Max pages per run",
321            "minimum": 0,
322            "type": "integer",
323            "description": "The maximum number of pages that the scraper will load. The scraper will stop when this limit is reached. It's always a good idea to set this limit in order to prevent excess platform usage for misconfigured scrapers. Note that the actual number of pages loaded might be slightly higher than this value.<br><br>If set to <code>0</code>, there is no limit.",
324            "default": 0
325          },
326          "maxResultsPerCrawl": {
327            "title": "Max result records",
328            "minimum": 0,
329            "type": "integer",
330            "description": "The maximum number of records that will be saved to the resulting dataset. The scraper will stop when this limit is reached. <br><br>If set to <code>0</code>, there is no limit.",
331            "default": 0
332          },
333          "maxCrawlingDepth": {
334            "title": "Max crawling depth",
335            "minimum": 0,
336            "type": "integer",
337            "description": "Specifies how many links away from <b>Start URLs</b> the scraper will descend. This value is a safeguard against infinite crawling depths for misconfigured scrapers. Note that pages added using <code>context.enqueuePage()</code> in <b>Page function</b> are not subject to the maximum depth constraint. <br><br>If set to <code>0</code>, there is no limit. To crawl only the pages specified by the Start URLs, set <a href=\"#linkSelector\"><code>linkSelector</code></a> empty instead.",
338            "default": 0
339          },
340          "maxConcurrency": {
341            "title": "Max concurrency",
342            "minimum": 1,
343            "type": "integer",
344            "description": "Specified the maximum number of pages that can be processed by the scraper in parallel. The scraper automatically increases and decreases concurrency based on available system resources. This option enables you to set an upper limit, for example to reduce the load on a target web server.",
345            "default": 50
346          },
347          "pageLoadTimeoutSecs": {
348            "title": "Page load timeout",
349            "minimum": 1,
350            "type": "integer",
351            "description": "The maximum amount of time the scraper will wait for a web page to load, in seconds. If the web page does not load in this timeframe, it is considered to have failed and will be retried (subject to <b>Max page retries</b>), similarly as with other page load errors.",
352            "default": 60
353          },
354          "pageFunctionTimeoutSecs": {
355            "title": "Page function timeout",
356            "minimum": 1,
357            "type": "integer",
358            "description": "The maximum amount of time the scraper will wait for <b>Page function</b> to execute, in seconds. It's a good idea to set this limit, to ensure that unexpected behavior in page function will not get the scraper stuck.",
359            "default": 60
360          },
361          "waitUntil": {
362            "title": "Navigation waits until",
363            "type": "array",
364            "description": "Contains a JSON array with names of page events to wait, before considering a web page fully loaded. The scraper will wait until <b>all</b> of the events are triggered in the web page before executing <b>Page function</b>. Available events are <code>domcontentloaded</code>, <code>load</code>, <code>networkidle2</code> and <code>networkidle0</code>.<br><br>For details, see <a href='https://pptr.dev/#?product=Puppeteer&show=api-pagegotourl-options' target='_blank' rel='noopener'><code>waitUntil</code> option</a> in Puppeteer's <code>Page.goto()</code> function documentation.",
365            "default": [
366              "networkidle2"
367            ]
368          },
369          "preNavigationHooks": {
370            "title": "Pre-navigation hooks",
371            "type": "string",
372            "description": "Async functions that are sequentially evaluated before the navigation. Good for setting additional cookies or browser properties before navigation. The function accepts two parameters, `crawlingContext` and `gotoOptions`, which are passed to the `page.goto()` function the crawler calls to navigate."
373          },
374          "postNavigationHooks": {
375            "title": "Post-navigation hooks",
376            "type": "string",
377            "description": "Async functions that are sequentially evaluated after the navigation. Good for checking if the navigation was successful. The function accepts `crawlingContext` as the only parameter."
378          },
379          "breakpointLocation": {
380            "title": "Insert breakpoint",
381            "enum": [
382              "NONE",
383              "BEFORE_GOTO",
384              "BEFORE_PAGE_FUNCTION",
385              "AFTER_PAGE_FUNCTION"
386            ],
387            "type": "string",
388            "description": "This property has no effect if Run mode is set to PRODUCTION. When set to DEVELOPMENT it inserts a breakpoint at the selected location in every page the scraper visits. Execution of code stops at the breakpoint until manually resumed in the DevTools window accessible via Live View tab or Container URL. Additional breakpoints can be added by adding <code>debugger;</code> statements within your Page function. <br><br>See <a href='https://apify.com/apify/web-scraper#run-mode' target='_blank' rel='noopener'>Run mode</a> in README for details.",
389            "default": "NONE"
390          },
391          "closeCookieModals": {
392            "title": "Dismiss cookie modals",
393            "type": "boolean",
394            "description": "Using the [I don't care about cookies](https://addons.mozilla.org/en-US/firefox/addon/i-dont-care-about-cookies/) browser extension. When on, the crawler will automatically try to dismiss cookie consent modals. This can be useful when crawling European websites that show cookie consent modals.",
395            "default": false
396          },
397          "maxScrollHeightPixels": {
398            "title": "Maximum scrolling distance in pixels",
399            "type": "integer",
400            "description": "The crawler will scroll down the page until all content is loaded or the maximum scrolling distance is reached. Setting this to `0` disables scrolling altogether.",
401            "default": 5000
402          },
403          "debugLog": {
404            "title": "Debug log",
405            "type": "boolean",
406            "description": "If enabled, the actor log will include debug messages. Beware that this can be quite verbose. Use <code>context.log.debug('message')</code> to log your own debug messages from <b>Page function</b>.",
407            "default": false
408          },
409          "browserLog": {
410            "title": "Browser log",
411            "type": "boolean",
412            "description": "If enabled, the actor log will include console messages produced by JavaScript executed by the web pages (e.g. using <code>console.log()</code>). Beware that this may result in the log being flooded by error messages, warnings and other messages of little value, especially with high concurrency.",
413            "default": false
414          },
415          "customData": {
416            "title": "Custom data",
417            "type": "object",
418            "description": "A custom JSON object that is passed to <b>Page function</b> as <code>context.customData</code>. This setting is useful when invoking the scraper via API, in order to pass some arbitrary parameters to your code.",
419            "default": {}
420          },
421          "datasetName": {
422            "title": "Dataset name",
423            "type": "string",
424            "description": "Name or ID of the dataset that will be used for storing results. If left empty, the default dataset of the run will be used."
425          },
426          "keyValueStoreName": {
427            "title": "Key-value store name",
428            "type": "string",
429            "description": "Name or ID of the key-value store that will be used for storing records. If left empty, the default key-value store of the run will be used."
430          },
431          "requestQueueName": {
432            "title": "Request queue name",
433            "type": "string",
434            "description": "Name of the request queue that will be used for storing requests. If left empty, the default request queue of the run will be used."
435          }
436        }
437      },
438      "runsResponseSchema": {
439        "type": "object",
440        "properties": {
441          "data": {
442            "type": "object",
443            "properties": {
444              "id": {
445                "type": "string"
446              },
447              "actId": {
448                "type": "string"
449              },
450              "userId": {
451                "type": "string"
452              },
453              "startedAt": {
454                "type": "string",
455                "format": "date-time",
456                "example": "2025-01-08T00:00:00.000Z"
457              },
458              "finishedAt": {
459                "type": "string",
460                "format": "date-time",
461                "example": "2025-01-08T00:00:00.000Z"
462              },
463              "status": {
464                "type": "string",
465                "example": "READY"
466              },
467              "meta": {
468                "type": "object",
469                "properties": {
470                  "origin": {
471                    "type": "string",
472                    "example": "API"
473                  },
474                  "userAgent": {
475                    "type": "string"
476                  }
477                }
478              },
479              "stats": {
480                "type": "object",
481                "properties": {
482                  "inputBodyLen": {
483                    "type": "integer",
484                    "example": 2000
485                  },
486                  "rebootCount": {
487                    "type": "integer",
488                    "example": 0
489                  },
490                  "restartCount": {
491                    "type": "integer",
492                    "example": 0
493                  },
494                  "resurrectCount": {
495                    "type": "integer",
496                    "example": 0
497                  },
498                  "computeUnits": {
499                    "type": "integer",
500                    "example": 0
501                  }
502                }
503              },
504              "options": {
505                "type": "object",
506                "properties": {
507                  "build": {
508                    "type": "string",
509                    "example": "latest"
510                  },
511                  "timeoutSecs": {
512                    "type": "integer",
513                    "example": 300
514                  },
515                  "memoryMbytes": {
516                    "type": "integer",
517                    "example": 1024
518                  },
519                  "diskMbytes": {
520                    "type": "integer",
521                    "example": 2048
522                  }
523                }
524              },
525              "buildId": {
526                "type": "string"
527              },
528              "defaultKeyValueStoreId": {
529                "type": "string"
530              },
531              "defaultDatasetId": {
532                "type": "string"
533              },
534              "defaultRequestQueueId": {
535                "type": "string"
536              },
537              "buildNumber": {
538                "type": "string",
539                "example": "1.0.0"
540              },
541              "containerUrl": {
542                "type": "string"
543              },
544              "usage": {
545                "type": "object",
546                "properties": {
547                  "ACTOR_COMPUTE_UNITS": {
548                    "type": "integer",
549                    "example": 0
550                  },
551                  "DATASET_READS": {
552                    "type": "integer",
553                    "example": 0
554                  },
555                  "DATASET_WRITES": {
556                    "type": "integer",
557                    "example": 0
558                  },
559                  "KEY_VALUE_STORE_READS": {
560                    "type": "integer",
561                    "example": 0
562                  },
563                  "KEY_VALUE_STORE_WRITES": {
564                    "type": "integer",
565                    "example": 1
566                  },
567                  "KEY_VALUE_STORE_LISTS": {
568                    "type": "integer",
569                    "example": 0
570                  },
571                  "REQUEST_QUEUE_READS": {
572                    "type": "integer",
573                    "example": 0
574                  },
575                  "REQUEST_QUEUE_WRITES": {
576                    "type": "integer",
577                    "example": 0
578                  },
579                  "DATA_TRANSFER_INTERNAL_GBYTES": {
580                    "type": "integer",
581                    "example": 0
582                  },
583                  "DATA_TRANSFER_EXTERNAL_GBYTES": {
584                    "type": "integer",
585                    "example": 0
586                  },
587                  "PROXY_RESIDENTIAL_TRANSFER_GBYTES": {
588                    "type": "integer",
589                    "example": 0
590                  },
591                  "PROXY_SERPS": {
592                    "type": "integer",
593                    "example": 0
594                  }
595                }
596              },
597              "usageTotalUsd": {
598                "type": "number",
599                "example": 0.00005
600              },
601              "usageUsd": {
602                "type": "object",
603                "properties": {
604                  "ACTOR_COMPUTE_UNITS": {
605                    "type": "integer",
606                    "example": 0
607                  },
608                  "DATASET_READS": {
609                    "type": "integer",
610                    "example": 0
611                  },
612                  "DATASET_WRITES": {
613                    "type": "integer",
614                    "example": 0
615                  },
616                  "KEY_VALUE_STORE_READS": {
617                    "type": "integer",
618                    "example": 0
619                  },
620                  "KEY_VALUE_STORE_WRITES": {
621                    "type": "number",
622                    "example": 0.00005
623                  },
624                  "KEY_VALUE_STORE_LISTS": {
625                    "type": "integer",
626                    "example": 0
627                  },
628                  "REQUEST_QUEUE_READS": {
629                    "type": "integer",
630                    "example": 0
631                  },
632                  "REQUEST_QUEUE_WRITES": {
633                    "type": "integer",
634                    "example": 0
635                  },
636                  "DATA_TRANSFER_INTERNAL_GBYTES": {
637                    "type": "integer",
638                    "example": 0
639                  },
640                  "DATA_TRANSFER_EXTERNAL_GBYTES": {
641                    "type": "integer",
642                    "example": 0
643                  },
644                  "PROXY_RESIDENTIAL_TRANSFER_GBYTES": {
645                    "type": "integer",
646                    "example": 0
647                  },
648                  "PROXY_SERPS": {
649                    "type": "integer",
650                    "example": 0
651                  }
652                }
653              }
654            }
655          }
656        }
657      }
658    }
659  }
660}

Web Scraper OpenAPI definition

OpenAPI is a standard for designing and describing RESTful APIs, allowing developers to define API structure, endpoints, and data formats in a machine-readable way. It simplifies API development, integration, and documentation.

OpenAPI is effective when used with AI agents and GPTs by standardizing how these systems interact with various APIs, for reliable integrations and efficient communication.

By defining machine-readable API specifications, OpenAPI allows AI models like GPTs to understand and use varied data sources, improving accuracy. This accelerates development, reduces errors, and provides context-aware responses, making OpenAPI a core component for AI applications.

You can download the OpenAPI definitions for Web Scraper from the options below:

If you’d like to learn more about how OpenAPI powers GPTs, read our blog post.

You can also check out our other API clients:

Developer
Maintained by Apify

Actor Metrics

  • 3.3k monthly users

  • 456 bookmarks

  • >99% runs succeeded

  • 4.8 days response time

  • Created in Mar 2019

  • Modified a month ago