Website Content Crawler avatar
Website Content Crawler

Pricing

Pay per usage

Go to Store
Website Content Crawler

Website Content Crawler

apify/website-content-crawler

Developed by

Apify

Maintained by Apify

Crawl websites and extract text content to feed AI models, LLM applications, vector databases, or RAG pipelines. The Actor supports rich formatting using Markdown, cleans the HTML, downloads files, and integrates well with 🦜🔗 LangChain, LlamaIndex, and the wider LLM ecosystem.

4.6 (38)

Pricing

Pay per usage

1.1k

Monthly users

6k

Runs succeeded

>99%

Response time

2.3 days

Last modified

7 days ago

JF

Crawling fails when URLs respond with application/pdf

Closed
jfmatt opened this issue
a month ago

I have a number of use cases where links go to PDF documents - which should use the raw HTML downloader - but where the URLs do not end in .pdf. In the linked example run, the URLs end in "pdf", but don't have an apparent file extension. When this happens, the crawler fails - it looks like it's attempting to use the browser-based crawl, and rejecting these pages.

Ideally, the decision of which crawler to use would be based on the MIME type of the response after fetching. Failing that - is there any way to configure different heuristics for how to handle these URLs?

jakub.kopecky avatar

Hi,

This behavior is expected when the crawler downloads a file (it throws an exception for that). If you want to access the file, you need to set "saveFiles": true or (in UI "Output settings" -> "Save files") and the files will be accessible in Key Value storage. Please see this run: https://console.apify.com/view/runs/QLJaWGZRHSb20u0X3

Let me know if this does not work for you or if you have any questions.

Jakub

Pricing

Pricing model

Pay per usage

This Actor is paid per platform usage. The Actor is free to use, and you only pay for the Apify platform usage.