N8N Template Scraper avatar
N8N Template Scraper

Pricing

Pay per event

Go to Apify Store
N8N Template Scraper

N8N Template Scraper

Scrape every public n8n workflow and extract metadata, categories, node usage, and full import-ready JSON files. Outputs cleaned descriptions, timestamps, slugs, and simplified node lists. Perfect for automation development, workflow libraries, analytics, and AI-driven analysis.

Pricing

Pay per event

Rating

5.0

(2)

Developer

Gavin Campbell

Gavin Campbell

Maintained by Community

Actor stats

5

Bookmarked

29

Total users

13

Monthly active users

5 days ago

Last modified

Share

n8n Template Scraper – Workflow JSON, Nodes & Metadata

Fast, lightweight scraper for the public n8n.io workflow template library.

This Actor talks directly to the official api.n8n.io templates API to fetch:

  • Workflow metadata (name, description, categories, views, author, timestamps)
  • A normalised summary of node types used in the workflow
  • A clean array of nodes with simplified parameters
  • The complete raw workflow JSON
  • An importable .json file for each template (saved to Key-Value Store)

Use it to build your own n8n template library, analyse node usage across templates, or feed workflows into your own AI/automation tools.


🚀 Key Features

  • Scrape all templates or a subset

    • Toggle “Scrape All Workflows” to crawl the entire n8n template library.
    • Or pass a list of specific workflow IDs to fetch only what you need.
  • Full workflow JSON export

    • Each template’s importable JSON is saved as a separate file in the run’s Key-Value Store.
    • Filenames use a stable slug: {{workflow_slug}}.json (e.g. build-your-first-ai-agent.json).
  • Rich, normalised dataset

    • For every workflow the dataset includes:
      • IDs, slugs, URLs
      • Description & categories
      • Author info
      • Views & timestamps
      • Node summary (counts by type/family)
      • A clean list of nodes (with human-friendly pretty_type)
      • The full raw workflow JSON as a string field
  • Node type intelligence

    • Automatically normalises node types like:
      • n8n-nodes-base.httpRequestToolHTTP Request Tool
      • @n8n/n8n-nodes-langchain.lmChatGoogleGeminiLM Chat – Google Gemini
    • Classifies nodes into families: core, langchain, community, llmTool, llmModel, ui.
  • Efficient & robust

    • Uses BasicCrawler + got (no browsers), so it’s fast and compute-efficient.
    • Request queue is pre-filled from template IDs, then fetched in parallel respecting your maxConcurrency.

📂 Where is my data?

This Actor writes data to:

1. Dataset (structured table)

Location: Run → Dataset

Each item in the dataset corresponds to one n8n template and includes (fields abbreviated for clarity):

  • workflow_id – numeric ID (e.g. 6270)
  • workflow_name – template title
  • workflow_slug – URL-friendly slug
  • description – markdown-stripped description
  • categories – JSON string of category names
    (e.g. ["Personal Productivity","AI Chatbot"])
  • complexity_levelbeginner | simple | intermediate | advanced | complex
  • created_at, updated_at – ISO timestamps
  • total_views, recent_views

Author object

  • author.name
  • author.username
  • author.verified (boolean)
  • author.social_links – JSON string of links

URLs & file key

  • template_url – public n8n page, e.g.
    https://n8n.io/workflows/6270-build-your-first-ai-agent
  • api_url – internal template API endpoint
  • file_key – filename of the JSON in key-value store (e.g. build-your-first-ai-agent.json)

Node summary (for analytics)

  • node_summary.total_nodes
  • node_summary.core_nodes
  • node_summary.langchain_nodes
  • node_summary.community_nodes
  • node_summary.llm_model_nodes
  • node_summary.llm_tool_nodes
  • node_summary.unique_node_types – JSON string of raw type IDs
  • node_summary.pretty_node_types – JSON string of human-friendly node names
    (e.g. ["Sticky Note","RSS Feed Read Tool","HTTP Request Tool","LangChain Agent","LangChain Chat Trigger","Memory Buffer Window","LM Chat – Google Gemini"])

Nodes array (per workflow)

  • nodes – JSON string of an array like:
[
{
"id": "3808de8d-ef18-47f5-9621-b08ba961ae01",
"name": "Introduction Note",
"type": "n8n-nodes-base.stickyNote",
"pretty_type": "Sticky Note",
"family": "ui",
"position": [-752, -256],
"parameters": {
"content": "## Try It Out! ..."
}
}
]
  • workflow_json_raw – JSON string of the full raw workflow JSON as returned by the n8n API (nodes, connections, settings, meta, etc.)

You can download the dataset as CSV, JSON, or Excel from the Dataset tab.

2. Key-Value Store (importable JSON files)

For each template, an importable workflow JSON file is stored in the run’s Key-Value Store.

  • Key: file_key from the dataset
    (e.g. build-your-first-ai-agent.json)
  • Value: object with:
{
"name": "Build Your First AI Agent",
"nodes": [...],
"connections": {...},
"settings": {},
"versionId": ""
}

You can download these files and import them directly into your own n8n instance.


🔧 Input Parameters

These appear in the Input tab as toggles/fields.

FieldTypeDescription
scrapeAllWorkflowsBooleanIf true, the Actor attempts to discover and scrape all available workflows from the template API.
maxItemsNumberApproximate maximum number of workflows to scrape when not scraping all workflows. Default: 100.
idListArrayOptional list of specific workflow IDs to fetch (e.g. [6270, 3521, 1200]). Only these IDs are scraped.
maxConcurrencyNumberMaximum parallel HTTP requests for fetching individual templates. Default: 5.

Tip

  • Use ID list mode when you know exactly which templates you want.
  • Use Auto-discovery mode (with scrapeAllWorkflows or maxItems) to crawl the template index.

▶️ Example Inputs

  1. Scrape the first 100 workflows (auto-discovery)
{
"maxItems": 100
}
  1. Scrape all public templates
{
"scrapeAllWorkflows": true
}
  1. Scrape only specific workflow IDs
{
"idList": [6270, 3521, 1200],
"maxConcurrency": 10
}

🤖 API & Automation

You can trigger this Actor programmatically using Apify’s REST API and plug it into:

  • n8n (meta!)
  • Make.com / Zapier
  • Custom back-end scripts or cron jobs

Typical use cases:

  • Run weekly to collect new AI/LLM-related templates.
  • Mirror n8n’s template library into your own internal catalogue.
  • Feed workflow structures into AI agents or documentation generators.
  • Analyse which nodes and tools are most popular over time.

Check the API tab on the Actor’s Apify page for copy-paste examples in:

  • Node.js
  • Python
  • Curl
  • PHP
  • Browser fetch

⚡ Performance & Cost

  • Uses BasicCrawler + got – no headless browsers.
  • Very low compute usage even for hundreds of workflows.
  • Concurrency is configurable via maxConcurrency.

🧑‍💻 Development Notes

Built with:

Run locally:

npm install
apify run

📞 Customisation & Support

If you’d like to:

  • Add GitHub repository scraping for related assets
  • Enrich templates with extra metadata
  • Push results into your own database or CRM
  • Build bespoke scrapers or automation workflows

…feel free to contact the author via the Apify profile.
They are available for custom automation, n8n integration, and data-extraction projects.


title: "n8n Template Scraper – Workflow JSON, Nodes & Metadata" slug: "n8n-template-scraper-workflow-json-nodes-metadata" description: "Fast, lightweight scraper for the public n8n.io workflow template library on Apify." tags:

  • n8n
  • web-scraping
  • apify
  • automation
  • workflows
  • templates date: "2025-11-26"