My Actor
Try for free
No credit card required
View all Actors
My Actor
oh_shit_here_we_go_again/my-actor
Try for free
No credit card required
.actor/Dockerfile
1# First, specify the base Docker image.
2# You can see the Docker images from Apify at https://hub.docker.com/r/apify/.
3# You can also use any other image from Docker Hub.
4FROM apify/actor-python-selenium:3.11
5
6# Second, copy just requirements.txt into the Actor image,
7# since it should be the only file that affects the dependency install in the next step,
8# in order to speed up the build
9COPY requirements.txt ./
10
11# Install the packages specified in requirements.txt,
12# Print the installed Python version, pip version
13# and all installed packages with their versions for debugging
14RUN echo "Python version:" \
15 && python --version \
16 && echo "Pip version:" \
17 && pip --version \
18 && echo "Installing dependencies:" \
19 && pip install -r requirements.txt \
20 && echo "All installed Python packages:" \
21 && pip freeze
22
23# Next, copy the remaining files and directories with the source code.
24# Since we do this after installing the dependencies, quick build will be really fast
25# for most source file changes.
26COPY . ./
27
28# Use compileall to ensure the runnability of the Actor Python code.
29RUN python3 -m compileall -q .
30
31# Specify how to launch the source code of your Actor.
32# By default, the "python3 -m src" command is run
33CMD ["python3", "-m", "src"]
.actor/actor.json
1{
2 "actorSpecification": 1,
3 "name": "my-actor",
4 "title": "Getting started with Python and Selenium",
5 "description": "Scrapes titles of websites using Selenium.",
6 "version": "0.0",
7 "meta": {
8 "templateId": "python-selenium"
9 },
10 "input": "./input_schema.json",
11 "dockerfile": "./Dockerfile",
12 "storages": {
13 "dataset": {
14 "actorSpecification": 1,
15 "title": "URLs and their titles",
16 "views": {
17 "titles": {
18 "title": "URLs and their titles",
19 "transformation": {
20 "fields": [
21 "url",
22 "title"
23 ]
24 },
25 "display": {
26 "component": "table",
27 "properties": {
28 "url": {
29 "label": "URL",
30 "format": "text"
31 },
32 "title": {
33 "label": "Title",
34 "format": "text"
35 }
36 }
37 }
38 }
39 }
40 }
41 }
42}
.actor/input_schema.json
1{
2 "title": "Python Selenium Scraper",
3 "type": "object",
4 "schemaVersion": 1,
5 "properties": {
6 "start_urls": {
7 "title": "Start URLs",
8 "type": "array",
9 "description": "URLs to start with",
10 "prefill": [
11 { "url": "https://apify.com" }
12 ],
13 "editor": "requestListSources"
14 },
15 "max_depth": {
16 "title": "Maximum depth",
17 "type": "integer",
18 "description": "Depth to which to scrape to",
19 "default": 1
20 }
21 },
22 "required": ["start_urls"]
23}
.dockerignore
1# configurations
2.idea
3
4# crawlee and apify storage folders
5apify_storage
6crawlee_storage
7storage
8
9# installed files
10.venv
11
12# git folder
13.git
.editorconfig
1root = true
2
3[*]
4indent_style = space
5indent_size = 4
6charset = utf-8
7trim_trailing_whitespace = true
8insert_final_newline = true
9end_of_line = lf
.gitignore
1# This file tells Git which files shouldn't be added to source control
2
3.idea
4.DS_Store
5
6apify_storage
7storage
8
9.venv/
10.env/
11__pypackages__
12dist/
13build/
14*.egg-info/
15*.egg
16
17__pycache__
18
19.mypy_cache
20.dmypy.json
21dmypy.json
22.pytest_cache
23.ruff_cache
24
25.scrapy
26*.log
requirements.txt
1# Feel free to add your Python dependencies below. For formatting guidelines, see:
2# https://pip.pypa.io/en/latest/reference/requirements-file-format/
3
4apify ~= 1.7.0
5selenium ~= 4.14.0
src/__main__.py
1"""
2This module serves as the entry point for executing the Apify Actor. It handles the configuration of logging
3settings. The `main()` coroutine is then executed using `asyncio.run()`.
4
5Feel free to modify this file to suit your specific needs.
6"""
7
8import asyncio
9import logging
10
11from apify.log import ActorLogFormatter
12
13from .main import main
14
15# Configure loggers
16handler = logging.StreamHandler()
17handler.setFormatter(ActorLogFormatter())
18
19apify_client_logger = logging.getLogger('apify_client')
20apify_client_logger.setLevel(logging.INFO)
21apify_client_logger.addHandler(handler)
22
23apify_logger = logging.getLogger('apify')
24apify_logger.setLevel(logging.DEBUG)
25apify_logger.addHandler(handler)
26
27# Execute the Actor main coroutine
28asyncio.run(main())
src/main.py
1"""
2This module defines the `main()` coroutine for the Apify Actor, executed from the `__main__.py` file.
3
4Feel free to modify this file to suit your specific needs.
5
6To build Apify Actors, utilize the Apify SDK toolkit, read more at the official documentation:
7https://docs.apify.com/sdk/python
8"""
9
10from urllib.parse import urljoin
11
12from selenium import webdriver
13from selenium.webdriver.chrome.options import Options as ChromeOptions
14from selenium.webdriver.common.by import By
15
16from apify import Actor
17
18# To run this Actor locally, you need to have the Selenium Chromedriver installed.
19# https://www.selenium.dev/documentation/webdriver/getting_started/install_drivers/
20# When running on the Apify platform, it is already included in the Actor's Docker image.
21
22
23async def main() -> None:
24 """
25 The main coroutine is being executed using `asyncio.run()`, so do not attempt to make a normal function
26 out of it, it will not work. Asynchronous execution is required for communication with Apify platform,
27 and it also enhances performance in the field of web scraping significantly.
28 """
29 async with Actor:
30 # Read the Actor input
31 actor_input = await Actor.get_input() or {}
32 start_urls = actor_input.get('start_urls', [{'url': 'https://apify.com'}])
33 max_depth = actor_input.get('max_depth', 1)
34
35 if not start_urls:
36 Actor.log.info('No start URLs specified in actor input, exiting...')
37 await Actor.exit()
38
39 # Enqueue the starting URLs in the default request queue
40 default_queue = await Actor.open_request_queue()
41 for start_url in start_urls:
42 url = start_url.get('url')
43 Actor.log.info(f'Enqueuing {url} ...')
44 await default_queue.add_request({'url': url, 'userData': {'depth': 0}})
45
46 # Launch a new Selenium Chrome WebDriver
47 Actor.log.info('Launching Chrome WebDriver...')
48 chrome_options = ChromeOptions()
49 if Actor.config.headless:
50 chrome_options.add_argument('--headless')
51 chrome_options.add_argument('--no-sandbox')
52 chrome_options.add_argument('--disable-dev-shm-usage')
53 driver = webdriver.Chrome(options=chrome_options)
54
55 driver.get('http://www.example.com')
56 assert driver.title == 'Example Domain'
57
58 # Process the requests in the queue one by one
59 while request := await default_queue.fetch_next_request():
60 url = request['url']
61 depth = request['userData']['depth']
62 Actor.log.info(f'Scraping {url} ...')
63
64 try:
65 # Open the URL in the Selenium WebDriver
66 driver.get(url)
67
68 # If we haven't reached the max depth,
69 # look for nested links and enqueue their targets
70 if depth < max_depth:
71 for link in driver.find_elements(By.TAG_NAME, 'a'):
72 link_href = link.get_attribute('href')
73 link_url = urljoin(url, link_href)
74 if link_url.startswith(('http://', 'https://')):
75 Actor.log.info(f'Enqueuing {link_url} ...')
76 await default_queue.add_request({
77 'url': link_url,
78 'userData': {'depth': depth + 1},
79 })
80
81 # Push the title of the page into the default dataset
82 title = driver.title
83 await Actor.push_data({'url': url, 'title': title})
84 except Exception:
85 Actor.log.exception(f'Cannot extract data from {url}.')
86 finally:
87 await default_queue.mark_request_as_handled(request)
88
89 driver.quit()