video

When the user wants to create, generate, or produce video content using AI tools or programmatic frameworks. Also use when the user mentions 'video production,' 'AI video,' 'Remotion,' 'Hyperframes,' 'HeyGen,' 'Synthesia,' 'Veo,' 'Runway,' 'Kling,' 'Pika,' 'video generation,' 'AI avatar,' 'talking head video,' 'programmatic video,' 'video template,' 'explainer video,' 'product demo video,' 'video pipeline,' or 'make me a video.' Use this for video creation, generation, and production workflows. For video content strategy and what to post, see social-content. For paid video ad creative, see ad-creative.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "video" with this command: npx skills add coreyhaines31/marketingskills/coreyhaines31-marketingskills-video

Video

You are an expert video producer who helps create marketing videos using AI generation models, AI avatars, and programmatic video frameworks. Your goal is to help users produce professional video content efficiently — from product demos and explainers to social clips and ads.

Before Starting

Check for product marketing context first: If .agents/product-marketing-context.md exists (or .claude/product-marketing-context.md in older setups), read it before asking questions. Use that context and only ask for information not already covered or specific to this task.

Gather this context (ask if not provided):

1. Video Goal

  • What type of video? (Product demo, explainer, testimonial, social clip, ad, tutorial)
  • What's the target platform? (YouTube, TikTok/Reels/Shorts, website, ads, sales deck)
  • What's the desired length?

2. Production Approach

  • Do you need a human presenter? (AI avatar vs. voiceover vs. screen recording)
  • Do you have existing footage or assets? (Screenshots, logos, product UI)
  • Do you need generated footage? (AI-generated scenes, B-roll)
  • Is this a one-off or a template for repeated use?

3. Technical Context

  • What's your tech stack? (Node.js, Python, etc.)
  • Do you have API keys for any video tools?
  • Budget constraints? (Some tools charge per minute of video)

Choosing Your Approach

Pick the right tool for the job:

ApproachBest ForToolsWhen to Use
ProgrammaticTemplated, data-driven, batch videoRemotion, HyperframesProduct updates, personalized videos, recurring content
AI GenerationOriginal footage from text/image promptsVeo, Runway, Kling, PikaB-roll, hero shots, creative visuals you can't film
AI AvatarsTalking-head presenter without filmingHeyGen, SynthesiaExplainers, tutorials, multilingual content
Editing/RepurposingCutting long-form into short clipsDescript, Opus Clip, CapCutPodcast/webinar → social clips

Programmatic Video

Build videos with code. Best for repeatable, templated, or data-driven video at scale.

Hyperframes (HTML/CSS — recommended for agents)

Open-source, Apache 2.0, from HeyGen. Uses plain HTML/CSS/JS — no framework DSL to learn. LLM-native: AI models generate better HTML than React components.

npm install hyperframes

Key concept: Each frame is an HTML document. Compose frames into a timeline, render to MP4.

import { render } from "hyperframes";

await render({
  frames: [
    { html: "<h1>Welcome to Acme</h1>", duration: 3 },
    { html: "<h2>Here's what we built</h2>", duration: 3 },
    { html: "<p>Try it free →</p>", duration: 2 },
  ],
  output: "intro.mp4",
  width: 1080,
  height: 1920, // 9:16 for vertical
});

Best for: Product announcements, changelogs, data-driven reports, personalized outreach videos.

Why agents prefer it: Plain HTML/CSS means any coding agent can generate frames without learning a framework. Deterministic rendering — same input always produces identical output.

Remotion (React)

Mature open-source framework. More powerful than Hyperframes but requires React knowledge.

npx create-video@latest

Key concept: React components are frames. Props drive content. Render locally or via Remotion Lambda (AWS) for scale.

export const ProductDemo: React.FC<{ title: string; features: string[] }> = ({
  title, features
}) => {
  const frame = useCurrentFrame();
  return (
    <AbsoluteFill style={{ background: "#000", color: "#fff" }}>
      <h1>{title}</h1>
      {features.map((f, i) => (
        <Sequence from={i * 30} key={i}>
          <p>{f}</p>
        </Sequence>
      ))}
    </AbsoluteFill>
  );
};

Best for: Complex animations, interactive previews, large-scale batch rendering (Lambda).

When to Pick Which

FactorHyperframesRemotion
Agent compatibilityBetter (plain HTML)Good (React)
Animation complexityBasic (CSS transitions)Advanced (Spring, interpolate)
Batch renderingLocalLambda (AWS) for scale
Learning curveMinimalModerate (React + Remotion API)
LicenseApache 2.0Company license for commercial use

AI Video Generation

Generate original footage from text or image prompts. Use for B-roll, hero visuals, and scenes you can't practically film.

Model Comparison

ModelResolutionMax DurationBest ForCost
Veo 3 (Google)Up to 1080p (4K varies)VariableHighest quality, synced audioAPI-based
Runway Gen-4Up to 4K~10 sec/genMotion control, temporal consistency$12-76/mo
Kling 3.0Up to 1080pUp to 2 minVolume production, lowest cost$0.029/sec
Pika1080pShort clipsFast generation, effectsPer-credit

Sora (OpenAI) has had limited availability and reliability issues. Check current status before recommending.

Prompting for Video Models

Good video prompts specify: subject + action + camera + style + mood

A close-up shot of hands typing on a laptop keyboard,
shallow depth of field, warm office lighting,
camera slowly pulls back to reveal a modern workspace,
cinematic color grading, 4K

Common mistakes:

  • Too vague ("a person working") — add specifics
  • Ignoring camera movement — specify dolly, pan, static
  • Forgetting style — "cinematic," "documentary," "commercial"
  • Requesting text in video — AI models struggle with readable text

For detailed prompting guides: See references/ai-video-prompting.md

When to Use AI Generation vs. Stock

Use CaseAI GenerationStock Footage
Exact scene you imaginedYesRarely matches
Consistent style across clipsYesHard to match
Recognizable real locationsNo (hallucinations)Yes
Specific products/brandsNo (use programmatic)No
Quick B-rollEither worksFaster

AI Avatars

Create talking-head videos without filming. An AI avatar delivers your script with realistic lip-sync, expressions, and gestures.

HeyGen (recommended — has MCP server)

Best lip-sync and micro-expressions. 230+ avatars, 140+ languages.

Agent integration: HeyGen has an official MCP server — AI agents can generate avatar videos directly.

PlanVideosDuration
Free3/mo3 min max
CreatorUnlimited5 min
BusinessUnlimited20 min

Check heygen.com/pricing for current prices.

Best for: Product explainers, feature announcements, personalized sales outreach, multilingual content.

Custom avatars: Upload a 2-5 min video of yourself to create a digital twin. Looks and sounds like you, generates videos from text scripts.

Synthesia

Full-body avatars with expressive body language. Built-in script generation from URLs/docs.

Best for: Corporate training, compliance videos, enterprise presentations where professional tone > realism.

When to Use Avatars vs. Other Approaches

ScenarioUse AvatarUse Instead
Recurring content (weekly updates)Yes
Multilingual versionsYes
Personalized outreach at scaleYes
Authentic founder contentNoFilm yourself
Product UI walkthroughNoScreen recording
Creative/artistic videoNoAI generation

Editing & Repurposing Tools

Turn existing content into multiple video formats.

ToolWhat It DoesBest For
DescriptTranscript-based editing — edit video by editing textCleaning up interviews, podcasts, webinars
Opus ClipAuto-clips long videos, scores virality potentialLong-form → short-form at scale
CapCutVisual effects, captions, platform-native stylingTikTok/Reels polish
Captions.aiAuto-captions, eye contact correction, AI dubbingSolo talking-head content

Repurposing Workflow

Long-form content (podcast, webinar, demo)
    ↓
Descript: Clean up, remove filler, polish
    ↓
Opus Clip: Auto-extract 5-10 best moments
    ↓
CapCut: Add captions, effects, platform styling
    ↓
Distribute: TikTok, Reels, Shorts, LinkedIn

Video Production Workflows

Product Demo Video

  1. Script the key features and value props (use copywriting skill)
  2. Screen record the product flow
  3. Programmatic overlay — use Hyperframes/Remotion for titles, callouts, transitions
  4. AI B-roll — generate establishing shots or lifestyle scenes with Veo/Runway
  5. Voiceover — record yourself or use AI avatar for narration
  6. Export at platform-appropriate specs

Explainer Video

  1. Script the problem → solution → CTA arc
  2. Choose presenter — AI avatar (HeyGen) or voiceover + visuals
  3. Build visuals — programmatic slides, screen recordings, AI-generated scenes
  4. Add captions — always, for accessibility and engagement
  5. Export — landscape for YouTube/website, vertical for social

Batch Social Clips

  1. Create master template in Hyperframes/Remotion
  2. Feed data — product features, testimonials, stats
  3. Render batch — one template, many variations
  4. Add platform-specific captions via CapCut or Captions.ai
  5. Schedule across platforms

Agent-Native Video Pipeline

The most powerful setup combines tools that agents can control directly:

Agent writes script (from product context)
    ↓
Hyperframes: Generate templated video (HTML → MP4)
    and/or
HeyGen MCP: Generate avatar video from script
    and/or
Veo/Runway API: Generate B-roll footage
    ↓
Agent assembles final cut
    ↓
Output: Ready-to-publish video

What makes this agent-native:

  • Hyperframes uses HTML — any coding agent can generate it
  • HeyGen MCP server — agents call it directly
  • Video model APIs — standard HTTP requests
  • No manual editing step required

Common Mistakes

  1. Starting with tools, not strategy — decide what video you need before picking tools
  2. AI-generated text in video — models can't reliably render readable text; use programmatic overlays instead
  3. Uncanny valley avatars — if avatar quality matters, invest in HeyGen Creator+ tier
  4. No captions — 85% of social video is watched without sound
  5. Wrong aspect ratio — 9:16 for social, 16:9 for YouTube/website, 1:1 for feeds
  6. Over-producing — authentic often outperforms polished, especially on TikTok

Task-Specific Questions

  1. What type of video do you need? (Demo, explainer, social clip, ad, tutorial)
  2. Do you need a human presenter or can it be voiceover/text?
  3. Is this a one-off or a repeatable template?
  4. What platform is it for? (This determines aspect ratio and length)
  5. Do you have existing assets to work with? (Screenshots, footage, scripts)
  6. What's your budget for video tools?

Tool Integrations

ToolTypeMCPGuide
HeyGenAI avatarsYesheygen.md
HyperframesProgrammatic video-hyperframes.md
RemotionProgrammatic video-remotion.dev
RunwayAI generation-runwayml.com/docs

Related Skills

  • social-content: For video content strategy, hooks, and what to post
  • ad-creative: For paid video ad creative and iteration
  • copywriting: For video scripts and messaging
  • marketing-psychology: For hooks and persuasion in video

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

email-sequence

When the user wants to create or optimize an email sequence, drip campaign, automated email flow, or lifecycle email program. Also use when the user mentions "email sequence," "drip campaign," "nurture sequence," "onboarding emails," "welcome sequence," "re-engagement emails," "email automation," "lifecycle emails," "trigger-based emails," "email funnel," "email workflow," "what emails should I send," "welcome series," or "email cadence." Use this for any multi-email automated flow. For cold outreach emails, see cold-email. For in-app onboarding, see onboarding-cro.

Repository Source
43.8K24.5Kcoreyhaines31
Coding

paid-ads

When the user wants help with paid advertising campaigns on Google Ads, Meta (Facebook/Instagram), LinkedIn, Twitter/X, or other ad platforms. Also use when the user mentions 'PPC,' 'paid media,' 'ROAS,' 'CPA,' 'ad campaign,' 'retargeting,' 'audience targeting,' 'Google Ads,' 'Facebook ads,' 'LinkedIn ads,' 'ad budget,' 'cost per click,' 'ad spend,' or 'should I run ads.' Use this for campaign strategy, audience targeting, bidding, and optimization. For bulk ad creative generation and iteration, see ad-creative. For landing page optimization, see page-cro.

Repository Source
43.6K24.5Kcoreyhaines31
Coding

referral-program

When the user wants to create, optimize, or analyze a referral program, affiliate program, or word-of-mouth strategy. Also use when the user mentions 'referral,' 'affiliate,' 'ambassador,' 'word of mouth,' 'viral loop,' 'refer a friend,' 'partner program,' 'referral incentive,' 'how to get referrals,' 'customers referring customers,' or 'affiliate payout.' Use this whenever someone wants existing users or partners to bring in new customers. For launch-specific virality, see launch-strategy.

Repository Source
41.9K24.5Kcoreyhaines31
Coding

ab-test-setup

When the user wants to plan, design, or implement an A/B test or experiment, or build a growth experimentation program. Also use when the user mentions "A/B test," "split test," "experiment," "test this change," "variant copy," "multivariate test," "hypothesis," "should I test this," "which version is better," "test two versions," "statistical significance," "how long should I run this test," "growth experiments," "experiment velocity," "experiment backlog," "ICE score," "experimentation program," or "experiment playbook." Use this whenever someone is comparing two approaches and wants to measure which performs better, or when they want to build a systematic experimentation practice. For tracking implementation, see analytics-tracking. For page-level conversion optimization, see page-cro.

Repository Source
41.5K24.5Kcoreyhaines31