Extended GPT Scraper

  • drobnikj/extended-gpt-scraper
  • Modified
  • Users 434
  • Runs 35.4k
  • Created by Author's avatarJakub Drobn铆k

Extract data from any website and feed it into GPT via the OpenAI API. Use ChatGPT to proofread content, analyze sentiment, summarize reviews, extract contact details, and much more.

To run the code examples, you need to have an Apify account. Replace <YOUR_API_TOKEN> in the code with your API token. For a more detailed explanation, please read about running Actors via the API in Apify Docs.

import { ApifyClient } from 'apify-client';

// Initialize the ApifyClient with API token
const client = new ApifyClient({
    token: '<YOUR_API_TOKEN>',

// Prepare Actor input
const input = {
    "startUrls": [
            "url": "https://news.ycombinator.com/"
    "globs": [],
    "linkSelector": "a[href]",
    "instructions": "Gets the post with the most points from the page and returns it as JSON in this format: \npostTitle\npostUrl\npointsCount",
    "model": "gpt-3.5-turbo",
    "targetSelector": "",
    "schema": {
        "type": "object",
        "properties": {
            "title": {
                "type": "string",
                "description": "Page title"
            "description": {
                "type": "string",
                "description": "Page description"
        "required": [
    "proxyConfiguration": {
        "useApifyProxy": true

(async () => {
    // Run the Actor and wait for it to finish
    const run = await client.actor("drobnikj/extended-gpt-scraper").call(input);

    // Fetch and print Actor results from the run's dataset (if any)
    console.log('Results from dataset');
    const { items } = await client.dataset(run.defaultDatasetId).listItems();
    items.forEach((item) => {