Wikipedia Pageviews Scraper
Pricing
from $8.25 / 1,000 items
Wikipedia Pageviews Scraper
Pull Wikipedia pageview metrics for any article in any language edition. Daily or monthly granularity, filter by access type (desktop, mobile, app) and agent type (user, spider, automated). Pick a date range. Export to JSON, CSV, or Excel for SEO research and content benchmarking.
Pricing
from $8.25 / 1,000 items
Rating
0.0
(0)
Developer
ParseForge
Actor stats
0
Bookmarked
2
Total users
1
Monthly active users
a day ago
Last modified
Share

📚 Wikipedia Pageviews Scraper
🚀 Pull daily and monthly Wikipedia pageviews for any article in any language. Filter by date range, access type, and agent type. No API key, no registration, no quota negotiation.
🕒 Last updated: 2026-05-01 · 📊 8 fields per row · 📚 300+ language editions · 📅 daily and monthly granularity · 🗓️ data from July 2015 onward
The Wikipedia Pageviews Scraper queries the official Wikimedia REST API and returns the number of times any Wikipedia article was viewed during a date range. Each row reports the language project, article title, timestamp, access type, agent type, and view count. The endpoint covers every Wikipedia language edition, and the underlying dataset goes back to July 2015, giving you nearly a decade of continuous traffic history per article.
Wikipedia is the eighth most visited website in the world with billions of pageviews per month. Pageview trends are a leading indicator for cultural moments, search demand, breaking news, and product launches. Building your own pipeline against the Wikimedia API means handling URL encoding, paginated date ranges, and per-language hosts. This Actor handles all of that and lets you focus on the analysis.
| 🎯 Target Audience | 💡 Primary Use Cases |
|---|---|
| SEO teams, journalists, trend researchers, market analysts, academics, dashboard builders | Search demand forecasting, cultural research, content benchmarking, trend tracking, comparative analysis |
📋 What the Wikipedia Pageviews Scraper does
Five filtering workflows in a single run:
- 📚 Per-article views. Submit any Wikipedia article title and pull its full traffic history for the date range you choose.
- 🌍 Any language edition. Pick from 20+ supported language projects including English, Spanish, German, French, Japanese, Russian, and Chinese Wikipedia.
- 📅 Daily or monthly granularity. Daily rollups give you weekday seasonality. Monthly rollups give you long-term trend lines.
- 📱 Access type filter. Slice traffic by desktop, mobile web, mobile app, or all-access combined.
- 🤖 Agent type filter. Separate human (
user) traffic from spiders and automated agents to clean up trend lines.
Each row in the dataset reports the project (e.g. en.wikipedia), URL-encoded article title, granularity, timestamp in YYYYMMDD00 format, access slice, agent slice, and view count. Dataset entries go back to July 2015.
💡 Why it matters: pageview data is one of the cleanest free signals for tracking real-world attention. When a celebrity dies, a film trailer drops, or a country votes, the matching Wikipedia article spikes within hours. SEO teams use the pageview series as a free proxy for search demand. Researchers cite it in studies of collective attention. Dashboard builders embed it as a public-interest gauge.
🎬 Full Demo
🚧 Coming soon: a 3-minute walkthrough showing how to go from sign-up to a downloaded dataset.
⚙️ Input
| Input | Type | Default | Behavior |
|---|---|---|---|
maxItems | integer | 10 | Rows to return. Free plan caps at 10, paid plan at 1,000,000. |
articles | array of strings | ["Albert_Einstein"] | Article titles with underscores in place of spaces. One title per array entry. |
project | string | "en.wikipedia.org" | Wikipedia language project. Pick from the enum of 20 supported language editions. |
granularity | string | "daily" | Either daily or monthly. |
startDate | string | 30 days ago | ISO date YYYY-MM-DD. Earliest supported is 2015-07-01. |
endDate | string | yesterday | ISO date YYYY-MM-DD. |
access | string | "all-access" | all-access, desktop, mobile-app, or mobile-web. |
agent | string | "all-agents" | all-agents, user, spider, or automated. |
Example: daily English-Wikipedia views for three articles in April 2026.
{"maxItems": 100,"articles": ["Albert_Einstein", "ChatGPT", "Taylor_Swift"],"project": "en.wikipedia.org","granularity": "daily","startDate": "2026-04-01","endDate": "2026-04-30","access": "all-access","agent": "user"}
Example: monthly Spanish-Wikipedia views since 2020.
{"maxItems": 1000,"articles": ["Lionel_Messi", "Real_Madrid_CF"],"project": "es.wikipedia.org","granularity": "monthly","startDate": "2020-01-01","endDate": "2026-04-01"}
⚠️ Good to Know: Wikipedia article titles are case sensitive and use underscores, not spaces. Submit
Albert_Einstein, notalbert einstein. Articles that have been moved or deleted return zero rows. The Wikimedia API is unauthenticated but expects a descriptive User-Agent string, which the Actor sends automatically.
📊 Output
Each row contains 8 fields. Download the dataset as CSV, Excel, JSON, or XML.
🧾 Schema
| Field | Type | Example |
|---|---|---|
🌐 project | string | "en.wikipedia" |
📄 article | string | "Albert_Einstein" |
⏱️ granularity | string | "daily" |
📅 timestamp | string | "2026040100" |
📱 access | string | "all-access" |
🤖 agent | string | "all-agents" |
👁️ views | integer | 15626 |
🕒 scrapedAt | ISO 8601 | "2026-05-01T02:00:11.931Z" |
📦 Sample records
✨ Why choose this Actor
| Capability | |
|---|---|
| 🆓 | Free official source. Pulls directly from the public Wikimedia REST API, no scraping of HTML pages. |
| 🌍 | All Wikipedia languages. Pick from 20+ enum-listed projects, request more if you need them. |
| 📅 | Decade of history. Data goes back to July 2015, with daily and monthly rollups. |
| 🧪 | Clean filter slices. Separate desktop from mobile, separate human traffic from spiders. |
| 🚀 | Sub-10-second runs. A typical 100-row pull finishes in under 10 seconds. |
| 🛠️ | Bulk article support. Submit dozens of articles in a single run, results pushed in order. |
| 🔄 | Export anywhere. Output ships as CSV, Excel, JSON, or XML through the Apify dataset endpoints. |
📊 The Wikimedia Foundation reports more than 18 billion pageviews per month across all editions.
📈 How it compares to alternatives
| Approach | Cost | Coverage | Refresh | Filters | Setup |
|---|---|---|---|---|---|
| Manual queries to the Wikimedia REST API | Free | Full | Live | Manual | Engineer hours |
| Third-party paid SEO suites | $$$ subscription | Partial | Daily | Built-in | Account setup |
| Generic web traffic estimators | $$ subscription | Estimated | Weekly | Limited | Account setup |
| ⭐ Wikipedia Pageviews Scraper (this Actor) | Pay-per-event | Full | Live | Granularity, access, agent | None |
The same data the Wikimedia Foundation publishes, exposed as clean structured records you can pipe into anything.
🚀 How to use
- 🆓 Create a free Apify account. Sign up here and get $5 in free credit.
- 🔍 Open the Actor. Search for "Wikipedia Pageviews" in the Apify Store.
- ⚙️ Set your inputs. Pick articles, project, date range, granularity, and any filters.
- ▶️ Click Start. Most runs finish in under 10 seconds.
- 📥 Download. Export as CSV, Excel, JSON, or XML, or wire it into a Make / Zapier flow.
⏱️ Total time from sign-up to first dataset: under five minutes.
💼 Business use cases
🌟 Beyond business use cases
Data like this powers more than commercial workflows. The same structured records support research, education, civic projects, and personal initiatives.
🔌 Automating Wikipedia Pageviews Scraper
Run this Actor on a schedule, from your codebase, or inside another tool:
- Node.js SDK: see Apify JavaScript client for programmatic runs and dataset exports.
- Python SDK: see Apify Python client for the same flow in Python.
- HTTP API: see Apify API docs for raw REST integration.
Schedule daily, weekly, or monthly runs from the Apify Console. Export results to Google Sheets, S3, or your own webhook with the built-in integrations.
❓ Frequently Asked Questions
🔌 Integrate with any app
- Make - drop run results into 1,800+ apps with a no-code visual builder.
- Zapier - trigger automations off completed runs.
- Slack - post run summaries to a channel.
- Google Sheets - sync each run into a spreadsheet.
- Webhooks - notify your own services on run finish.
- Airbyte - load runs into Snowflake, BigQuery, or Postgres.
🔗 Recommended Actors
- 🅱️ Bing Search Scraper - track organic search demand alongside Wikipedia traffic.
- 🦆 DuckDuckGo Search Scraper - alternative SERP signal for the same topic.
- 📰 Substack Publication Scraper - pair Wikipedia trends with newsletter cadence.
- 🐙 GitHub Trending Repos Scraper - capture developer attention next to public attention.
- 🌐 Common Crawl Index Scraper - cross-reference web archive captures with traffic data.
💡 Pro Tip: browse the complete ParseForge collection for more pre-built scrapers and data tools.
🆘 Need Help? Open our contact form and we'll route the question to the right person.
Wikipedia is a registered trademark of the Wikimedia Foundation. This Actor is not affiliated with or endorsed by the Wikimedia Foundation. It is built on the public Wikimedia REST API and respects all published rate limits.