Public Actors Lister avatar
Public Actors Lister

Deprecated

Pricing

Pay per usage

Go to Store
Public Actors Lister

Public Actors Lister

Deprecated

Developed by

Jan Čurn

Jan Čurn

Maintained by Community

Downloads a list of all Actors published in Apify Store, with all properties such as URL, title, description, etc. This is useful to create a knowledge file for a GPT, so that it knows which Actors can it use.

5.0 (1)

Pricing

Pay per usage

7

Total users

46

Monthly users

6

Runs succeeded

>99%

Last modified

a day ago

.actor/Dockerfile

# Specify the base Docker image. You can read more about
# the available images at https://docs.apify.com/sdk/js/docs/guides/docker-images
# You can also use any other image from Docker Hub.
FROM apify/actor-node:18
# Copy just package.json and package-lock.json
# to speed up the build using Docker layer cache.
COPY package*.json ./
# Install NPM packages, skip optional and development dependencies to
# keep the image small. Avoid logging too much and print the dependency
# tree for debugging
RUN npm --quiet set progress=false \
&& npm install --omit=dev --omit=optional \
&& echo "Installed NPM packages:" \
&& (npm list --omit=dev --all || true) \
&& echo "Node.js version:" \
&& node --version \
&& echo "NPM version:" \
&& npm --version \
&& rm -r ~/.npm
# Next, copy the remaining files and directories with the source code.
# Since we do this after NPM install, quick build will be really fast
# for most source file changes.
COPY . ./
# Run the image.
CMD npm start --silent

.actor/actor.json

{
"actorSpecification": 1,
"name": "my-actor-23",
"title": "Scrape single page in JavaScript",
"description": "Scrape data from single page with provided URL.",
"version": "0.0",
"meta": {
"templateId": "js-start"
},
"input": "./input_schema.json",
"storages": {
"dataset": {
"actorSpecification": 1,
"views": {
"overview": {
"title": "Overview",
"transformation": {
"fields": [
"pictureUrl",
"url",
"title",
"description",
"categories",
"authorFullName",
"authorPictureUrl"
]
},
"display": {
"component": "table",
"properties": {
"pictureUrl": {
"label": "Image",
"format": "image"
},
"url": {
"label": "URL",
"format": "link"
},
"title": {
"label": "Title",
"format": "text"
},
"description": {
"label": "Description",
"format": "text"
},
"authorFullName": {
"label": "Author",
"format": "text"
},
"authorPictureUrl": {
"label": "Author picture",
"format": "image"
}
}
}
}
}
}
},
"dockerfile": "./Dockerfile"
}

.actor/input_schema.json

{
"title": "This Actor has no input",
"type": "object",
"schemaVersion": 1,
"properties": {
"search": {
"title": "Search term",
"type": "string",
"description": "Filter Actors by string in title, name, description, username and readme.",
"editor": "textfield"
}
},
"required": []
}

src/main.js

1import { Actor } from 'apify';
2import { ApifyClient } from 'apify-client';
3import escapeStringRegexp from 'escape-string-regexp';
4
5await Actor.init();
6
7const apifyClient = new ApifyClient();
8
9const { search } = await Actor.getInput();
10
11// Fetch all Actors
12let actors = [];
13const fetchNextChunk = async (offset = 0) => {
14 const limit = 1000;
15 const value = await apifyClient.store({ search }).list({ offset, limit });
16
17 if (value.count === 0) {
18 return;
19 }
20
21 actors.push(...value.items);
22
23 if (value.total > offset + value.count) {
24 return fetchNextChunk(offset + value.count);
25 }
26};
27await fetchNextChunk();
28
29const usernames = {};
30
31// Tranform the records
32const results = [];
33for (const actor of actors) {
34 results.push({
35 id: actor.id,
36 url: `https://apify.com/${actor.username}/${actor.name}`,
37 title: actor.title,
38 titleWithUrl: `[(${actor.title})](https://apify.com/${actor.username}/${actor.name})`,
39 pictureUrl: actor.pictureUrl,
40 description: actor.description,
41 categories: actor.categories,
42 authorPictureUrl: actor.userPictureUrl,
43 authorFullName: actor.userFullName,
44 usersTotal: actor.stats.totalUsers,
45 usersMonthly: actor.stats.totalUsers30Days,
46 });
47 // Limit this to reduce the length of regex
48 if (actor.stats.totalUsers > 35) {
49 usernames[actor.username] = true;
50 }
51}
52
53// An extra: generate a regex for Google Search Console
54let regexAll = `^https:\\/\\/apify\\.com\\/(${Object.keys(usernames).map(key => escapeStringRegexp(key)).join('|')})(\\/|$)`;
55
56await Actor.setValue('gsc_regex_all_user_generated_content.txt', regexAll, { contentType: 'text/plain' });
57
58await Actor.pushData(results);
59
60await Actor.exit();

.dockerignore

# configurations
.idea
# crawlee and apify storage folders
apify_storage
crawlee_storage
storage
# installed files
node_modules
# git folder
.git

.gitignore

# This file tells Git which files shouldn't be added to source control
.DS_Store
.idea
dist
node_modules
apify_storage
storage/*
!storage/key_value_stores
storage/key_value_stores/*
!storage/key_value_stores/default
storage/key_value_stores/default/*
!storage/key_value_stores/default/INPUT.json

package.json

{
"name": "js-scrape-single-page",
"version": "0.0.1",
"type": "module",
"description": "This is an example of an Apify actor.",
"engines": {
"node": ">=18.0.0"
},
"dependencies": {
"apify": "^3.1.10",
"escape-string-regexp": "^5.0.0"
},
"scripts": {
"start": "node ./src/main.js",
"test": "echo \"Error: oops, the actor has no tests yet, sad!\" && exit 1"
},
"author": "It's not you it's me",
"license": "ISC"
}