1from apify_client import ApifyClient
2
3
4
5client = ApifyClient("<YOUR_API_TOKEN>")
6
7
8run_input = {
9 "startUrl": "https://crawlee.dev/docs/introduction",
10 "maxDepth": 10,
11 "maxPages": 100,
12 "includeGlobs": [],
13 "excludeGlobs": [
14 "**/blog/**",
15 "**/changelog/**",
16 "**/api-reference/**",
17 ],
18 "excludeElements": "nav, footer, .sidebar, script, style, .ads, header, .navigation, .menu, .toc, .breadcrumb, .edit-page, .feedback, .newsletter, aside",
19 "contentSelector": "main, article, .content, .documentation",
20 "mergeOutput": True,
21 "includeMetadata": True,
22}
23
24
25run = client.actor("direct_duty/llm-ready-documentation-scraper").call(run_input=run_input)
26
27
28print("💾 Check your data here: https://console.apify.com/storage/datasets/" + run["defaultDatasetId"])
29for item in client.dataset(run["defaultDatasetId"]).iterate_items():
30 print(item)
31
32