Rss Feed Parser avatar

Rss Feed Parser

Pricing

from $10.00 / 1,000 results

Go to Apify Store
Rss Feed Parser

Rss Feed Parser

Parses and extracts structured data from RSS and Atom feeds. Provide one or more feed URLs and the actor will fetch each feed, parse all items, and output a normalized dataset with titles, links, dates, authors, content snippets, and categories. Supports both RSS 2.0 and Atom feed formats. Ideal ...

Pricing

from $10.00 / 1,000 results

Rating

0.0

(0)

Developer

Donny

Donny

Maintained by Community

Actor stats

0

Bookmarked

1

Total users

0

Monthly active users

7 hours ago

Last modified

Categories

Share

RSS Feed Parser & Scraper

What it does

Parses and extracts structured data from RSS and Atom feeds. Provide one or more feed URLs and the actor will fetch each feed, parse all items, and output a normalized dataset with titles, links, dates, authors, content snippets, and categories. Supports both RSS 2.0 and Atom feed formats. Ideal for monitoring news sources, blog aggregation, content curation, and media tracking workflows.

This Apify actor automates the collection of data from a public API or website, extracting structured information and saving it directly into an Apify dataset. It handles pagination automatically where applicable, supports configurable result limits, and includes robust error handling with timeouts on all HTTP requests. The actor is designed for reliability: it validates inputs, applies sensible defaults, and produces a fallback record when no results are found, so your downstream workflows never receive an empty dataset. Built on the Apify SDK with native Node.js 20 fetch for lightweight, fast execution without browser overhead.

Why use it

Manually collecting data from web APIs and websites is tedious and error-prone. This actor eliminates that burden by running in the cloud on the Apify platform, where it can be scheduled, integrated with webhooks, or chained with other actors. Whether you are conducting research, building a knowledge base, monitoring data sources, or feeding data into an analytics pipeline, this actor gives you structured, ready-to-use JSON output with zero browser overhead. It uses lightweight HTTP requests instead of a full browser, which makes it fast and cost-effective. Every request includes a 120-second timeout to prevent hanging, and all string fields are null-checked for data consistency.

Input parameters

  • feedUrls (array, required): List of RSS or Atom feed URLs to parse. Default: ["https://feeds.bbci.co.uk/news/rss.xml"].
  • maxResults (integer, optional): Maximum number of total items to return across all feeds. Default: 100. Range: 1-5000.

All inputs are validated at startup with sensible defaults applied when values are missing. The actor will log warnings for any misconfigured options and continue with safe defaults rather than failing outright.

Output data

Each item in the output dataset contains the following fields:

  • feedTitle: Title of the parent feed
  • feedUrl: URL of the feed
  • itemTitle: Title of the feed item
  • itemLink: URL link to the full article
  • itemDate: Publication date of the item
  • itemAuthor: Author or creator of the item
  • itemContent: Content snippet (truncated to 500 characters)
  • categories: Comma-separated list of item categories

All string fields are null-checked; missing values are stored as null rather than undefined.

Example output

{
"feedTitle": "BBC News",
"feedUrl": "https://feeds.bbci.co.uk/news/rss.xml",
"itemTitle": "Breaking news headline",
"itemLink": "https://www.bbc.co.uk/news/article-123",
"itemDate": "2024-01-15T12:00:00.000Z",
"itemAuthor": "BBC Correspondent",
"itemContent": "Article summary text goes here...",
"categories": "World, Politics"
}

Pricing

This actor is priced on a usage basis:

  • $0.01 per result returned in the dataset.
  • $0.005 per actor start (fixed platform fee).

For example, scraping 500 results would cost approximately $5.005. Apify provides free monthly credits for new users, so you can try the actor at no charge. Actual costs depend on the number of results, API response times, and memory allocation. You can control costs by setting the maxResults parameter to limit the number of results collected per run. For high-volume use cases, consider running the actor on a schedule during off-peak hours to optimize platform resource usage.

More scrapers from brave_paradise

Check out other useful scrapers built by brave_paradise:

Visit the brave_paradise profile on Apify to see the full catalogue of actors.