Website Content Crawler avatar
Website Content Crawler

Pricing

Pay per usage

Go to Store
Website Content Crawler

Website Content Crawler

Developed by

Apify

Apify

Maintained by Apify

Crawl websites and extract text content to feed AI models, LLM applications, vector databases, or RAG pipelines. The Actor supports rich formatting using Markdown, cleans the HTML, downloads files, and integrates well with 🦜🔗 LangChain, LlamaIndex, and the wider LLM ecosystem.

4.0 (41)

Pricing

Pay per usage

1596

Total users

62K

Monthly users

8.2K

Runs succeeded

>99%

Issues response

8.2 days

Last modified

15 hours ago

NC

Crawler does not extract content, with no useful logs to debug

Closed

nimble_caretaker opened this issue
8 months ago

We are attempting to scan our client's website. They have whitelisted US proxies. Unfortunately, most of the time, the results are not extracted. With the same settings, it worked once, but now it completely doesn't. The logs are not insightful enough to resolve or understand the issue the crawler has encountered.

We need to provide an answer to our client on whether the problem can be solved by us or if it's something on their end (e.g., CDN blocking).

Using: CrawlerType - playwright:firefox Proxy - US proxy (3 IP address that our clients have whitelisted)

jiri.spilka avatar

Hi, thank you for using Website Content Crawler.

I checked the logs of your run and adjusted a couple of settings, and it’s now working. Here are the key changes: In the logs, I noticed Navigation timed out after 20 seconds. This was due to the Request timeout setting. Increasing this value to 60 seconds (the default) resolves the issue.

Other settings to change:

"maxCrawlDepth": 0 -> 20,
"maxCrawlPages": 0 -> 9999999
  • Max crawl depth: Setting this to 0 means the Actor will only scrape the startUrls without further crawling. You’ll need to set it to a higher value; the default is 20.
  • Max crawl pages: This controls the total number of pages to scrape. Remove this setting to use the default value.

With these updated settings:

"requestTimeoutSecs": 60,
"maxCrawlDepth": 20,
"maxCrawlPages": 9999999

I was able to successfully crawl the website. Here’s a log snippet:

2024-11-12T12:02:44.268Z INFO Crawling will be started using 1 start URLs and 0 sitemap URLs
2024-11-12T12:02:44.907Z INFO PlaywrightCrawler: Starting the crawler.
2024-11-12T12:03:22.368Z INFO Enqueued 66 new links on https://****il/.

Please let me know if it works for you too.

NC

nimble_caretaker

8 months ago

Works well, Thank you for your time.