Public Actors Lister avatar
Public Actors Lister

Pricing

Pay per usage

Go to Store
Public Actors Lister

Public Actors Lister

Developed by

Jan Čurn

Jan Čurn

Maintained by Community

Downloads a list of all Actors published in Apify Store, with all properties such as URL, title, description, etc. This is useful to create a knowledge file for a GPT, so that it knows which Actors can it use.

0.0 (0)

Pricing

Pay per usage

6

Total users

39

Monthly users

6

Runs succeeded

>99%

Last modified

a month ago

.actor/Dockerfile

# Specify the base Docker image. You can read more about
# the available images at https://docs.apify.com/sdk/js/docs/guides/docker-images
# You can also use any other image from Docker Hub.
FROM apify/actor-node:18
# Copy just package.json and package-lock.json
# to speed up the build using Docker layer cache.
COPY package*.json ./
# Install NPM packages, skip optional and development dependencies to
# keep the image small. Avoid logging too much and print the dependency
# tree for debugging
RUN npm --quiet set progress=false \
&& npm install --omit=dev --omit=optional \
&& echo "Installed NPM packages:" \
&& (npm list --omit=dev --all || true) \
&& echo "Node.js version:" \
&& node --version \
&& echo "NPM version:" \
&& npm --version \
&& rm -r ~/.npm
# Next, copy the remaining files and directories with the source code.
# Since we do this after NPM install, quick build will be really fast
# for most source file changes.
COPY . ./
# Run the image.
CMD npm start --silent

.actor/actor.json

{
"actorSpecification": 1,
"name": "my-actor-23",
"title": "Scrape single page in JavaScript",
"description": "Scrape data from single page with provided URL.",
"version": "0.0",
"meta": {
"templateId": "js-start"
},
"input": "./input_schema.json",
"storages": {
"dataset": {
"actorSpecification": 1,
"views": {
"overview": {
"title": "Overview",
"transformation": {
"fields": [
"pictureUrl",
"url",
"title",
"description",
"categories",
"authorFullName",
"authorPictureUrl"
]
},
"display": {
"component": "table",
"properties": {
"pictureUrl": {
"label": "Image",
"format": "image"
},
"url": {
"label": "URL",
"format": "link"
},
"title": {
"label": "Title",
"format": "text"
},
"description": {
"label": "Description",
"format": "text"
},
"authorFullName": {
"label": "Author",
"format": "text"
},
"authorPictureUrl": {
"label": "Author picture",
"format": "image"
}
}
}
}
}
}
},
"dockerfile": "./Dockerfile"
}

.actor/input_schema.json

{
"title": "This Actor has no input",
"type": "object",
"schemaVersion": 1,
"properties": {
"search": {
"title": "Search term",
"type": "string",
"description": "Filter Actors by string in title, name, description, username and readme.",
"editor": "textfield"
}
},
"required": []
}

src/main.js

1import { Actor } from 'apify';
2import { ApifyClient } from 'apify-client';
3import escapeStringRegexp from 'escape-string-regexp';
4
5await Actor.init();
6
7const apifyClient = new ApifyClient();
8
9const { search } = await Actor.getInput();
10
11// Fetch all Actors
12let actors = [];
13const fetchNextChunk = async (offset = 0) => {
14 const limit = 1000;
15 const value = await apifyClient.store({ search }).list({ offset, limit });
16
17 if (value.count === 0) {
18 return;
19 }
20
21 actors.push(...value.items);
22
23 if (value.total > offset + value.count) {
24 return fetchNextChunk(offset + value.count);
25 }
26};
27await fetchNextChunk();
28
29const usernames = {};
30
31// Tranform the records
32const results = [];
33for (const actor of actors) {
34 results.push({
35 id: actor.id,
36 url: `https://apify.com/${actor.username}/${actor.name}`,
37 title: actor.title,
38 titleWithUrl: `[(${actor.title})](https://apify.com/${actor.username}/${actor.name})`,
39 pictureUrl: actor.pictureUrl,
40 description: actor.description,
41 categories: actor.categories,
42 authorPictureUrl: actor.userPictureUrl,
43 authorFullName: actor.userFullName,
44 usersTotal: actor.stats.totalUsers,
45 usersMonthly: actor.stats.totalUsers30Days,
46 });
47 // Limit this to reduce the length of regex
48 if (actor.stats.totalUsers > 35) {
49 usernames[actor.username] = true;
50 }
51}
52
53// An extra: generate a regex for Google Search Console
54let regexAll = `^https:\\/\\/apify\\.com\\/(${Object.keys(usernames).map(key => escapeStringRegexp(key)).join('|')})(\\/|$)`;
55
56await Actor.setValue('gsc_regex_all_user_generated_content.txt', regexAll, { contentType: 'text/plain' });
57
58await Actor.pushData(results);
59
60await Actor.exit();

.dockerignore

# configurations
.idea
# crawlee and apify storage folders
apify_storage
crawlee_storage
storage
# installed files
node_modules
# git folder
.git

.gitignore

# This file tells Git which files shouldn't be added to source control
.DS_Store
.idea
dist
node_modules
apify_storage
storage/*
!storage/key_value_stores
storage/key_value_stores/*
!storage/key_value_stores/default
storage/key_value_stores/default/*
!storage/key_value_stores/default/INPUT.json

package.json

{
"name": "js-scrape-single-page",
"version": "0.0.1",
"type": "module",
"description": "This is an example of an Apify actor.",
"engines": {
"node": ">=18.0.0"
},
"dependencies": {
"apify": "^3.1.10",
"escape-string-regexp": "^5.0.0"
},
"scripts": {
"start": "node ./src/main.js",
"test": "echo \"Error: oops, the actor has no tests yet, sad!\" && exit 1"
},
"author": "It's not you it's me",
"license": "ISC"
}