Legacy PhantomJS Crawler
No credit card required
Legacy PhantomJS Crawler
No credit card required
Replacement for the legacy Apify Crawler product with a backward-compatible interface. The actor uses PhantomJS headless browser to recursively crawl websites and extract data from them using a piece of front-end JavaScript code.
Do you want to learn more about this Actor?
Get a demoYou can access the Legacy PhantomJS Crawler programmatically from your own applications by using the Apify API. You can choose the language preference from below. To use the Apify API, you’ll need an Apify account and your API token, found in Integrations settings in Apify Console.
1# Set API token
2API_TOKEN=<YOUR_API_TOKEN>
3
4# Prepare Actor input
5cat > input.json << 'EOF'
6{
7 "startUrls": [
8 {
9 "key": "START",
10 "value": "https://www.example.com/"
11 }
12 ],
13 "crawlPurls": [
14 {
15 "key": "MY_LABEL",
16 "value": "https://www.example.com/[.*]"
17 }
18 ],
19 "clickableElementsSelector": "a:not([rel=nofollow])",
20 "pageFunction": "function pageFunction(context) {\n // called on every page the crawler visits, use it to extract data from it\n var $ = context.jQuery;\n var result = {\n title: $('title').text(),\n myValue: $('TODO').text()\n };\n return result;\n}\n",
21 "interceptRequest": "function interceptRequest(context, newRequest) {\n // called whenever the crawler finds a link to a new page,\n // use it to override default behavior\n return newRequest;\n}\n"
22}
23EOF
24
25# Run the Actor using an HTTP API
26# See the full API reference at https://docs.apify.com/api/v2
27curl "https://api.apify.com/v2/acts/apify~legacy-phantomjs-crawler/runs?token=$API_TOKEN" \
28 -X POST \
29 -d @input.json \
30 -H 'Content-Type: application/json'
Legacy PhantomJS Crawler - Crawl websites, extract data API
Below, you can find a list of relevant HTTP API endpoints for calling the Legacy PhantomJS Crawler Actor. For this, you’ll need an Apify account. Replace <YOUR_API_TOKEN> in the URLs with your Apify API token, which you can find under Integrations in Apify Console. For details, see the API reference .
Run Actor
https://api.apify.com/v2/acts/apify~legacy-phantomjs-crawler/runs?token=<YOUR_API_TOKEN>
Note: By adding the method=POST
query parameter, this API endpoint can be called using a GET request and thus used in third-party webhooks. Please refer to our Run Actor API documentation .
Run Actor synchronously and get dataset items
https://api.apify.com/v2/acts/apify~legacy-phantomjs-crawler/run-sync-get-dataset-items?token=<YOUR_API_TOKEN>
Note: This endpoint supports both POST and GET request methods. However, only the POST method allows you to pass input data. For more information, please refer to our Run Actor synchronously and get dataset items API documentation .
Get Actor
https://api.apify.com/v2/acts/apify~legacy-phantomjs-crawler?token=<YOUR_API_TOKEN>
For more information, please refer to our Get Actor API documentation .
Actors can be used to scrape web pages, extract data, or automate browser tasks. Use the Legacy PhantomJS Crawler API programmatically via the Apify API.
You can choose from:
You can start Legacy PhantomJS Crawler with the Apify API by sending an HTTP POST request to the Run Actor endpoint. An Actor’s input and its content type can be passed as a payload of the POST request, and additional options can be specified using URL query parameters. The Legacy PhantomJS Crawler is identified within the API by its ID, which is the creator’s username and the name of the Actor.
When the Legacy PhantomJS Crawler run finishes you can list the data from its default dataset (storage) via the API or you can preview the data directly on Apify Console .
- 106 monthly users
- 21 stars
- 100.0% runs succeeded
- 61 days response time
- Created in Mar 2019
- Modified 4 months ago