Turn your website into an AI chatbot
Automate content updates for customer support and beyond with Website Content Crawler. Convert your website, blog, or FAQ into a chatbot-ready format. Keep your data current and relevant with fresh web data without worrying about scraping challenges or infrastructure.
Website Content Crawler
apify/website-content-crawler
Crawl websites and extract text content to feed AI models, LLM applications, vector databases, or RAG pipelines. The Actor supports rich formatting using Markdown, cleans the HTML, downloads files, and integrates well with 🦜🔗 LangChain, LlamaIndex, and the wider LLM ecosystem.
29.1k
734
Google Search Results Scraper
apify/google-search-scraper
Scrape Google Search Engine Results Pages (SERPs). Select the country or language and extract organic and paid results, AI overviews, ads, queries, People Also Ask, prices, reviews, like a Google SERP API. Export scraped data, run the scraper via API, schedule runs, or integrate with other tools.
51.1k
261
Extended GPT Scraper
drobnikj/extended-gpt-scraper
Extract data from any website and feed it into GPT via the OpenAI API. Use ChatGPT to proofread content, analyze sentiment, summarize reviews, extract contact details, and much more.
1.2k
54
RAG Web Browser
apify/rag-web-browser
Web browser for OpenAI Assistants API and RAG pipelines, similar to a web browser in ChatGPT. It queries Google Search, scrapes the top N pages from the results, and returns their cleaned content as Markdown for further processing by an LLM.
270
41
Pinecone Integration
apify/pinecone-integration
This integration transfers data from Apify Actors to a Pinecone and is a good starting point for a question-answering, search, or RAG use case.
119
19
Qdrant Integration
apify/qdrant-integration
Transfer data from Apify Actors to a Qdrant vector database.
19
3
Integrate with RAG pipelines for smart solutions
Use the data for RAG pipelines to create customer support chatbots that can answer questions directly from your site’s content, agent Q&A systems to connect your data with vector databases for retrieval, and current documentation hubs for developers working with specific libraries.