Github Issues Scraper
Pricing
Pay per usage
Github Issues Scraper
Pricing
Pay per usage
Rating
0.0
(0)
Developer

Donny Nguyen
Actor stats
0
Bookmarked
2
Total users
1
Monthly active users
2 days ago
Last modified
Categories
Share
What does it do?
GitHub Issues Scraper extracts issues and pull requests from GitHub repositories using the GitHub public API. It collects issue titles, descriptions, author information, labels, assignees, comment counts, state (open/closed), and timestamps. You can scrape multiple repositories in a single run and filter by issue state.
Why use this actor?
Monitoring GitHub repositories is essential for open source project management, competitive intelligence, developer tools research, and bug tracking analysis. This scraper automates the collection of issue data at scale, enabling you to analyze trends, track bug reports, monitor feature requests, and build dashboards across multiple repositories.
How to use it
- Go to the actor's page on the Apify platform.
- Click Start to open the input configuration.
- Enter one or more repository names in owner/repo format.
- Select the issue state filter (open, closed, or all).
- Set the maximum number of issues per repository.
- Click Start and wait for the results.
- Download your data from the Dataset tab.
The actor uses the GitHub public API, which has rate limits. For larger scrapes, consider running during off-peak hours.
Input configuration
| Field | Type | Description | Default |
|---|---|---|---|
| repos | array | GitHub repos (owner/repo format) | ["apify/crawlee", "microsoft/vscode"] |
| state | string | Issue state: open, closed, or all | "open" |
| maxIssues | integer | Max issues per repository | 500 |
| proxyConfiguration | object | Proxy settings | Apify Proxy |
Output data
Each item in the dataset contains:
{"repo": "apify/crawlee","number": 1234,"title": "CheerioCrawler fails with certain MIME types","state": "open","author": "developer123","labels": "bug, crawler","assignees": "maintainer1","comments": 5,"body": "When using CheerioCrawler with application/xml...","isPullRequest": false,"createdAt": "2026-01-15T10:30:00Z","updatedAt": "2026-02-19T08:15:00Z","closedAt": null,"url": "https://github.com/apify/crawlee/issues/1234","scrapedAt": "2026-02-19T14:30:00.000Z"}
Cost of usage
This actor uses the GitHub public API and CheerioCrawler, making it extremely lightweight. A typical run scraping 2 repositories with up to 500 issues each takes 1-3 minutes and costs under $0.01 in platform credits. The actor is priced at $0.30 per 1,000 results with pay-per-event pricing.
Tips
- GitHub's unauthenticated API allows 60 requests per hour; each page returns up to 100 issues.
- The isPullRequest field lets you filter between actual issues and pull requests.
- Use the labels field to categorize and analyze issues by type (bug, feature, etc.).
- The body field is truncated to 3,000 characters; most issue descriptions fit within this limit.
- Schedule regular runs to track how issue counts and types change over time.
- Combine data from multiple repositories to benchmark project health across competitors.
- Use the closedAt field to calculate average issue resolution time.
Built with Crawlee and Apify SDK. See more scrapers by consummate_mandala on Apify Store.