US Energy.gov Data Scraper
Pricing
Pay per event
US Energy.gov Data Scraper
Scrape energy-related content from Energy.gov, including articles, press releases, documents, titles, dates, offices, and types. Automate collection of structured data from the U.S. Department of Energy, ideal for researchers, journalists, and professionals needing accurate, up-to-date information.
Pricing
Pay per event
Rating
0.0
(0)
Developer
ParseForge
Maintained by CommunityActor stats
0
Bookmarked
2
Total users
0
Monthly active users
5 days ago
Last modified
Categories
Share

🔬 Energy.gov Scraper
🚀 Collect U.S. Department of Energy articles, press releases, and documents in seconds. Filter by keyword, office, article type, and language. No coding, no API keys required.
🕒 Last updated: 2026-04-16 · 📊 8 fields · 🏛️ Official DOE data · 🌐 English and Spanish
The Energy.gov Scraper automates the collection of official content from the U.S. Department of Energy website. It pulls articles, press releases, congressional testimonies, blog posts, success stories, and multimedia content directly from the DOE search system. Each record includes the headline, publication date, source office, content category, direct link, and unique identifier. You can filter by keyword, article type, DOE office, and language to zero in on exactly the content you need. Free users can collect up to 10 items per run, while paid users can retrieve up to 1,000,000 results.
Whether you are tracking renewable energy policy shifts, monitoring nuclear research announcements, or building a dataset of DOE press releases for media analysis, this tool replaces hours of manual browsing with a single automated run. Results export to JSON, CSV, or Excel for immediate use in spreadsheets, dashboards, or data pipelines. Schedule recurring runs to stay current with the latest DOE publications without lifting a finger. The scraper handles pagination, deduplication, and rate limiting automatically so you can focus on analysis instead of data collection.
| Target Audience | Use Cases |
|---|---|
| Policy Analysts | Monitor federal energy policy announcements and congressional testimonies |
| Academic Researchers | Build literature databases from DOE research publications |
| Energy Industry Professionals | Track regulatory changes and press releases by office |
| Journalists | Follow DOE news across topics like renewables, nuclear, and fossil fuels |
| Data Analysts | Export structured DOE content for trend analysis and reporting |
| Government Affairs Teams | Stay current on DOE initiatives and funding announcements |
📋 What the Energy.gov Scraper does
- 📝 Article headlines - capture the title of every article, press release, blog post, or document published on energy.gov
- 🔗 Direct URLs - collect working links to each piece of content for quick reference or archival
- 📅 Publication dates - track when content was published to build timelines and spot trends
- 👤 Source offices - identify which DOE office or organization published the content (e.g., Office of Energy Efficiency and Renewable Energy)
- 🎯 Content categories - classify each item by type: blog, press release, document, success story, congressional testimony, or multimedia
- 🆔 Unique identifiers - get UUIDs for each article to manage deduplication and data integrity
The scraper connects to the DOE search system and iterates through results using your specified filters. It collects structured data from each listing, normalizes timestamps, and removes duplicate entries using unique article IDs. All results are pushed to an Apify dataset in real time, so you can preview data as the run progresses.
💡 Why it matters: Energy.gov publishes thousands of articles annually across dozens of offices. Manually tracking this content is impractical. This scraper gives you structured, filterable access to the entire catalog in minutes.
🎬 Full Demo
🚧 Coming soon...
⚙️ Input
| Field | Type | Required | Description |
|---|---|---|---|
maxItems | integer | No | Maximum articles to collect. Free users limited to 10. Paid users up to 1,000,000. |
keywords | string | No | Search terms to filter articles (e.g., "renewable energy", "nuclear"). |
articleType | string | No | Content type filter: Blog, Press Releases, Documents, Success Stories, Multimedia, Congressional Testimony. |
language | string | No | Language filter: "en" for English, "es" for Spanish. |
office | string | No | Filter by DOE office or organization name. |
sort | string | No | Sort order: "date" or "relevance". |
Example 1: Renewable energy press releases
{"maxItems": 50,"keywords": "renewable energy","articleType": "7","language": "en","sort": "date"}
Example 2: All recent content from a specific office
{"maxItems": 100,"office": "Office of Energy Efficiency & Renewable Energy","sort": "date"}
⚠️ Good to Know: Free users are automatically limited to 10 items per run. Leave
keywordsempty to browse all available content. ThearticleTypefield uses numeric codes internally, but you can also use descriptive names.
📊 Output
🧾 Schema
| Emoji | Field | Type | Description |
|---|---|---|---|
| 📝 | title | string | Article headline or document title |
| 🔗 | url | string | Direct link to the full article on energy.gov |
| 📅 | publicationDate | string | Date the article was published |
| 👤 | office | string | DOE office or organization that published the content |
| 🎯 | contentType | string | Category: blog, press release, document, testimony, etc. |
| 🆔 | uuid | string | Unique article identifier for deduplication |
| 🕐 | scrapedAt | string | Timestamp of when the data was collected |
| ⚠️ | error | string | Error message if the item failed to process |
📦 Sample records
✨ Why choose this Actor
| Feature | Energy.gov Scraper | Alternatives |
|---|---|---|
| Filter by DOE office | Yes | Rarely available |
| Content type filtering (blog, press, testimony) | Yes | Limited |
| English and Spanish language support | Yes | English only |
| Sort by date or relevance | Yes | Date only |
| Automatic pagination and deduplication | Yes | Manual |
| Up to 1,000,000 results per run | Yes | Capped lower |
| Export to JSON, CSV, and Excel | Yes | JSON only |
📊 The DOE publishes content from 20+ offices. This scraper lets you filter by any combination of office, topic, and content type in a single run.
📈 How it compares to alternatives
| Capability | This Actor | Manual Browsing | Generic Web Scrapers |
|---|---|---|---|
| DOE-specific filters (office, type, language) | Yes | Yes | No |
| Automatic pagination | Yes | No | Partial |
| Structured output with deduplication | Yes | No | Varies |
| Scheduled recurring runs | Yes | No | Varies |
| No coding required | Yes | Yes | No |
| Bulk export (CSV, Excel, JSON) | Yes | No | Partial |
This scraper is purpose-built for energy.gov, so it handles the site's search system, pagination, and data structure out of the box.
🚀 How to use
- Sign up - Create a free Apify account with $5 credit
- Find the Actor - Search for "Energy.gov Scraper" in the Apify Store
- Configure your filters - Set keywords, article type, office, language, and max items
- Start the run - Click "Start" and watch results appear in real time
- Export your data - Download as JSON, CSV, or Excel from the dataset tab
🕒 Typical run time: 30 seconds to 2 minutes for up to 100 items. Larger runs with 500+ items may take 5 to 10 minutes.
💼 Business use cases
|
Policy Analysis
|
Academic Research
|
|
Media and Journalism
|
Energy Industry Intelligence
|
🌟 Beyond business use cases
Data like this powers more than commercial workflows. The same structured records support research, education, civic projects, and personal initiatives.
🤖 Ask an AI assistant about this scraper
Open a ready-to-send prompt about this ParseForge actor in the AI of your choice:
- 💬 ChatGPT
- 🧠 Claude
- 🔍 Perplexity
- 🅒 Copilot
❓ Frequently Asked Questions
💳 Do I need a paid Apify plan to run this actor?
No. You can start right now on the free Apify plan, which includes $5 in free monthly credit. That is enough to run this actor several times and explore the output before committing to anything. Paid plans unlock higher limits, more concurrent runs, and larger datasets. Create a free Apify account here to get started.
🚨 What happens if my run fails or returns no results?
Failed runs are not charged. If the source site changes, proxies get rate-limited, or a specific input matches nothing, re-run the actor or open our contact form and we will investigate. You can also check the run log in the Apify console to see why the run stopped.
📏 How many items can I scrape per run?
Free users are limited to 10 items per run so you can preview the output and confirm the actor works for your use case. Paid users can raise maxItems up to 1,000,000 per run. Upgrade here if you need full scale.
🕒 How fresh is the data?
Every run fetches live data at the moment of execution. There is no cache or delay: the records you get reflect what the source returned at that moment. Schedule the actor to maintain a rolling snapshot of the data you need.
🧑💻 Can I call this actor from my own code?
Yes. Apify exposes every actor as a REST endpoint and ships first-class SDKs for Node.js and Python. You can start a run, read the dataset, and handle webhooks from your own app in a few lines. All you need is your Apify API token.
📤 How do I export the data?
Every Apify dataset can be downloaded in one click from the console as CSV, JSON, JSONL, Excel, HTML, XML, or RSS. You can also pull results programmatically via the Apify API or stream them into BigQuery, S3, and other destinations through built-in integrations.
📅 Can I schedule the actor to run automatically?
Yes. Use the Apify scheduler to run the actor on any cadence, from hourly to monthly. Results are saved to your dataset and can be delivered to webhooks, email, Slack, cloud storage, or automation tools such as Zapier and Make.
🔌 Automating Energy.gov Scraper
Node.js example:
import { ApifyClient } from 'apify-client';const client = new ApifyClient({ token: 'YOUR_API_TOKEN' });const run = await client.actor('parseforge/energy-gov-scraper').call({maxItems: 50,keywords: 'renewable energy',sort: 'date'});const { items } = await client.dataset(run.defaultDatasetId).listItems();console.log(items);
Python example:
from apify_client import ApifyClientclient = ApifyClient('YOUR_API_TOKEN')run = client.actor('parseforge/energy-gov-scraper').call(run_input={'maxItems': 50,'keywords': 'renewable energy','sort': 'date'})items = list(client.dataset(run['defaultDatasetId']).iterate_items())print(items)
Schedules: Set up recurring runs on daily, weekly, or monthly intervals using the Apify scheduler. Combine with Slack or email integrations to get notified whenever new DOE content matches your filters.
🔌 Integrate with any app
- Make - Automate energy data workflows and route articles to your team
- Zapier - Connect to 5,000+ apps and trigger actions on new DOE content
- Slack - Get notifications when new articles match your filters
- Airbyte - Stream DOE data into your data warehouse
- GitHub - Version control your scraper configurations
- Google Drive - Export results directly to Google Sheets
🔗 Recommended Actors
| Actor | Description |
|---|---|
| USAspending Scraper | Extract federal spending data and contract information from USAspending.gov |
| GSA eLibrary Scraper | Collect government contractor and vendor data from the GSA eLibrary |
| PR Newswire Scraper | Collect press releases and news articles from PR Newswire |
| FINRA BrokerCheck Scraper | Search broker and firm registration data from the FINRA registry |
| FAA Aircraft Registry Scraper | Look up aircraft registration records by N-number from the FAA |
💡 Pro Tip: Combine the Energy.gov Scraper with the USAspending Scraper to cross-reference DOE announcements with actual federal spending data.
🆘 Need Help? Open our contact form to request a new scraper, propose a custom data project, or report an issue. We typically respond within 24 hours.
Disclaimer: This Actor is an independent tool and is not affiliated with, endorsed by, or sponsored by the U.S. Department of Energy or Energy.gov. All trademarks mentioned are the property of their respective owners.