YouTube Insight & Summarisation avatar

YouTube Insight & Summarisation

Pricing

Pay per usage

Go to Apify Store
YouTube Insight & Summarisation

YouTube Insight & Summarisation

Extract structured insights from YouTube videos — including summaries,and key moments — even when transcripts are missing. Built for: - Product research - Content consumption - Content intelligence

Pricing

Pay per usage

Rating

0.0

(0)

Developer

Shubham Deokar

Shubham Deokar

Maintained by Community

Actor stats

0

Bookmarked

0

Total users

0

Monthly active users

a day ago

Last modified

Categories

Share

🎬 YouTube Insight & Summarisation

Extract deep, structured insights from any YouTube video — without watching it. Summaries, timestamped key moments, and full transcripts. Works even when captions are missing. Supports Groq (free), Gemini (free), and OpenAI.

💰 Free to use — you only pay standard Apify compute costs. No per-result fees unlike most competitors.


Why YouTube Insight & Summarisation?

Most YouTube summarizers break when a video has no captions. This Actor doesn't.

It uses a two-step transcript pipeline:

  1. Caption extraction first — fast, free, works on most videos
  2. ASR audio transcription fallback — automatically kicks in when captions are unavailable, using AI speech recognition directly on the audio

This means it works on Hindi podcasts, regional content, older videos, and any video where the creator has disabled subtitles — content that other summarizers simply fail on.


Who Is This For?

  • 🎓 Students & researchers — extract key insights from long lectures without watching in full
  • 📰 Content creators — repurpose YouTube content into blogs, newsletters, or social posts
  • 💼 Professionals — catch up on industry podcasts and interviews in minutes
  • 🌍 Non-English speakers — get summaries in your native language using outputLanguage
  • 🔁 Automation builders — integrate YouTube summarization into n8n, Zapier, or custom pipelines via API
  • 🤖 Developers — call via REST API and plug structured JSON output into your product

Quick Start

Paste this into the Actor input and run:

{
"urls": ["https://www.youtube.com/watch?v=UF8uR6Z6KLc"],
"llmProvider": "groq",
"openaiApiKey": "YOUR_GROQ_KEY_HERE",
"includeTimestamps": true
}

💡 Get a free Groq key at console.groq.com — no credit card needed. Get a free Gemini key at aistudio.google.com — also free, higher token limits.


What This Actor Does

  • Accepts one or more YouTube URLs (batch processing supported)
  • Fetches video metadata: title, channel, description, views, duration, thumbnail
  • Fetches transcript using the caption-first → ASR fallback pipeline
  • Summarizes with your chosen LLM provider (Gemini, Groq, or OpenAI)
  • Optionally returns:
    • ⏱ Timestamped key points with clickable deep-linked YouTube URLs
    • 📄 Full transcript text in timestamped format

Features

FeatureDetails
✅ Works without captionsASR fallback via agentx/video-transcript
⏱ Timestamped key pointsMM:SS / HH:MM:SS with deep links (&t=Xs)
🌍 Multilingual outputSummarize in any language via outputLanguage
📦 Batch processingProcess multiple URLs in a single run
🔌 LLM flexibilityGroq (free), Gemini (free), or OpenAI
📄 Full transcriptOptional raw transcript output

Transcript Fallback Logic

For each video, the Actor follows this order:

Step 1 → Call pintostudio/youtube-transcript-scraper (caption-based, fast & free)
↓ captions found?
YES → use captions, transcriptSource: "captions"
NO
Step 2 → Call agentx/video-transcript (ASR audio transcription, ~$0.43/video)
ASR succeeded?
YES → use ASR transcript, transcriptSource: "asr"
NO
Step 3 → Push error item, skip video, continue to next URL

The transcriptSource field in the output tells you exactly which method was used.


Input Schema

FieldTypeRequiredDefaultDescription
urlsstring[]✅ YesList of YouTube video URLs to process
llmProvider"gemini" | "groq" | "openai"NoopenaiLLM provider for summarization
openaiApiKeystring✅ YesAPI key for the selected LLM provider
modelNamestringNoprovider defaultOverride the default model
outputLanguagestringNosame as videoOutput language (e.g. Spanish, Hindi)
includeTimestampsbooleanNotrueGenerate timestamped key points
includeCommentAnalysisbooleanNotrueAnalyze top comments for sentiment
includeFullTranscriptbooleanNofalseInclude full transcript in output
maxCommentsToAnalyzenumberNo50Max comments to include in sentiment analysis

Example Input

{
"urls": [
"https://www.youtube.com/watch?v=UF8uR6Z6KLc",
"https://www.youtube.com/watch?v=dQw4w9WgXcQ"
],
"llmProvider": "groq",
"openaiApiKey": "YOUR_GROQ_KEY_HERE",
"outputLanguage": "",
"includeTimestamps": true,
"includeCommentAnalysis": true,
"includeFullTranscript": false,
"maxCommentsToAnalyze": 50
}

ProviderDefault ModelFree TierGet Key
groqllama-3.3-70b-versatile✅ Yes (rate limited)console.groq.com
geminigemini-2.0-flash✅ Yes (generous limits)aistudio.google.com
openaigpt-4o-mini❌ Paidplatform.openai.com

💡 Recommended for most users: Start with Groq for short videos (< 30 min). Use Gemini for longer content — it handles larger contexts on the free tier.


Output Schema

Successful result

{
"videoUrl": "https://www.youtube.com/watch?v=UF8uR6Z6KLc",
"videoId": "UF8uR6Z6KLc",
"videoTitle": "Steve Jobs' 2005 Stanford Commencement Address",
"channelName": "Stanford",
"description": "...",
"publishedAt": "2008-03-17",
"viewCount": "45000000",
"likeCount": "900000",
"duration": "15m 4s",
"thumbnailUrl": "https://img.youtube.com/vi/UF8uR6Z6KLc/maxresdefault.jpg",
"summary": "Steve Jobs shares three personal stories...",
"keyPoints": [
{
"timestamp": "1:00",
"seconds": 60,
"timestampedUrl": "https://www.youtube.com/watch?v=UF8uR6Z6KLc&t=60s",
"topic": "Connecting the Dots",
"summary": "Jobs dropped out of Reed College but audited classes that later shaped the Mac's typography."
}
],
"keyPointsCount": 8,
"commentAnalysis": {
"overallSentiment": "Positive",
"consensusPoints": ["Highly motivational", "Timeless advice"],
"controversialPoints": [],
"topThemes": ["Career", "Purpose", "Death"],
"summary": "Audience overwhelmingly finds the speech life-changing..."
},
"fullTranscript": null,
"outputLanguage": "same as video",
"modelUsed": "groq/llama-3.3-70b-versatile",
"transcriptSource": "captions",
"processedAt": "2026-04-24T12:00:00.000Z"
}

Failed result (both caption and ASR failed)

{
"videoUrl": "https://www.youtube.com/watch?v=VIDEO_ID",
"videoId": "VIDEO_ID",
"error": "Transcription failed: video is private or region-locked",
"processedAt": "2026-04-24T12:00:00.000Z"
}

Cost & Billing

  • This Actor is free — no per-result charges
  • You only pay standard Apify compute costs (typically a few cents per run)
  • LLM token usage is billed directly by your provider (Groq and Gemini have free tiers)
  • The ASR fallback (agentx/video-transcript) costs approximately $0.43 per video and is only triggered when captions are unavailable — videos with captions incur no ASR cost

Local Development

$npm install

Create or edit storage/key_value_stores/default/INPUT.json:

{
"urls": ["https://www.youtube.com/watch?v=UF8uR6Z6KLc"],
"llmProvider": "groq",
"openaiApiKey": "YOUR_KEY_HERE",
"includeTimestamps": true,
"includeCommentAnalysis": true,
"includeFullTranscript": false,
"maxCommentsToAnalyze": 50
}

Run locally (Apify token required for sub-Actor calls):

$APIFY_TOKEN=your_apify_token npm run dev

Build check:

$npm run build

Deploy to Apify Console

npm install -g apify-cli
apify login
apify push

After deploying, run a smoke test with two videos:

  • One with captions → expect transcriptSource: "captions"
  • One without captions → expect transcriptSource: "asr"

Security

  • Never commit your APIFY_TOKEN or LLM API keys to version control
  • Use placeholder values in example inputs (e.g. "YOUR_GROQ_KEY_HERE")
  • Rotate any keys that were accidentally shared or logged

License

ISC