Skip to main content
Give your AI agent a URL and a creative direction. It captures the site, extracts the brand identity, writes a script and storyboard, generates voiceover, builds animated compositions, and delivers a renderable video.
"Create a 20-second product launch video from https://linear.app.
 Make it feel like an Apple keynote announcement."

Getting Started

1

Install skills

Skills teach your AI agent how to capture websites and create HyperFrames compositions. Install once — they persist across sessions.
npx skills add heygen-com/hyperframes
Works with Claude Code, Cursor, Gemini CLI, and Codex CLI.
2

Prompt your agent

Open your agent in any directory and describe the video you want:
Create a 25-second product launch video from https://example.com. Bold, cinematic, dark theme energy.
The agent loads the skill when they see a URL and a video request, and runs the full pipeline — capture, design, script, storyboard, voiceover, build, validate.
Agents also trigger this skill automatically when they see a URL and a video request.
3

Preview

npx hyperframes preview
Opens the video in your browser. Edits reload automatically.
4

Render to MP4

npx hyperframes render --output my-video.mp4
✓ Captured 750 frames in 12.4s
✓ Encoded to my-video.mp4 (25.0s, 1920×1080, 6.8MB)
You don’t need to run npx hyperframes capture manually — the skill instructs the agent to capture as the first step. The capture command is documented below for advanced use.

How the Pipeline Works

The skill runs 7 steps. Each produces an artifact that feeds the next:
StepOutputWhat happens
Capturecaptures/<name>/Extract screenshots, design tokens, fonts, assets, animations
DesignDESIGN.mdBrand reference — colors, typography, do’s and don’ts
ScriptSCRIPT.mdNarration text with hook, story, proof, CTA
StoryboardSTORYBOARD.mdPer-beat creative direction — mood, assets, animations, transitions
VO + Timingnarration.wav + transcript.jsonTTS audio with word-level timestamps
Buildcompositions/*.htmlAnimated HTML compositions, one per beat
ValidateSnapshot PNGsVisual verification before delivery

Video Types

The prompt determines the format. Include a duration and creative direction:
TypeDurationExample
Social ad10–15s”15-second Instagram reel. Energetic, fast cuts.”
Product launch20–30s”25-second product launch. Apple keynote energy.”
Product tour30–60s”45-second tour showing the top 3 features.”
Brand reel15–30s”20-second brand video. Celebrate the design.”
Feature announcement15–25s”Feature announcement highlighting the new AI agents.”
Teaser8–15s”10-second teaser. Super minimal. Just the hook.”
Creative direction matters more than format. “Playful, hand-crafted feel” or “dark, developer-focused, show code” shapes the storyboard and drives every visual decision the agent makes.

Enriching Captures with Gemini Vision

By default, captures describe assets using DOM context — alt text, nearby headings, CSS classes. Add a Gemini API key for richer AI-powered descriptions using vision. Create a .env file in your project root:
echo "GEMINI_API_KEY=your-key-here" > .env
- hero-bg.png — 582KB, section: "Hero", above fold
The agent knows the file exists and where it was on the page, but not what it looks like.
TierRate limitCost per image
Free5 RPMFree
Paid2,000 RPM~$0.001
A typical capture with 40 images costs about $0.04 on the paid tier.

Capture Command

The skill runs capture automatically, but you can run it directly for pre-caching, debugging, or using the data outside of video production.
npx hyperframes capture https://stripe.com
◇  Captured Stripe | Financial Infrastructure → captures/stripe-com

  Screenshots: 12
  Assets: 45
  Sections: 15
  Fonts: sohne-var
FlagDefaultDescription
-o, --outputcaptures/<hostname>Output directory
--timeout120000Page load timeout in ms
--skip-assetsfalseSkip downloading images and fonts
--max-screenshots24Maximum screenshot count
--jsonfalseOutput structured JSON for programmatic use

What Gets Captured

DataDescription
ScreenshotsViewport captures at every scroll depth — dynamic count based on page height
ColorsPixel-sampled dominant colors + computed styles, including oklch/lab conversion
FontsCSS font families + downloaded woff2 files
AssetsImages, SVGs with semantic names, Lottie animations, video previews
TextAll visible text in DOM order
AnimationsWeb Animations API, scroll-triggered animations, WebGL shaders
SectionsPage structure with headings, types, background colors
CTAsButtons and links detected by class names and text patterns

Snapshot Command

Capture key frames from a built video as PNGs — verify compositions without a full render:
npx hyperframes snapshot my-project --at 2.9,10.4,18.7
FlagDefaultDescription
--frames5Number of evenly-spaced frames
--atComma-separated timestamps in seconds
--timeout5000Ms to wait for runtime to initialize

Iterating

You don’t need to re-run the full pipeline to make changes:
  • Edit the storyboardSTORYBOARD.md is the creative north star. Change a beat’s mood or assets, then ask the agent to rebuild just that beat.
  • Edit a composition — open compositions/beat-3-proof.html directly and tweak animations, colors, or layout.
  • Rebuild one beat“Rebuild beat 2 with more energy. Use the product screenshot as full-bleed background.”

Troubleshooting

Increase the timeout for sites with Cloudflare or heavy client-side rendering:
npx hyperframes capture https://example.com --timeout 180000
Sites using frameworks like Framer lazy-load images via IntersectionObserver. The capture scrolls through the page to trigger loading, but very long pages may miss images near the bottom. Adding a Gemini key improves descriptions of captured assets, but doesn’t increase the count.
The capture uses pixel sampling combined with DOM computed styles. Dark sites should show dark colors in the palette. Check the scroll screenshots in captures/<name>/screenshots/ to see what the capture actually saw.
Verify skills are installed:
npx skills add heygen-com/hyperframes
Lead your prompt with “Use the /website-to-hyperframes skill” for the most reliable results. Agents also discover it automatically when they see a URL and a video request.

Next Steps

Quickstart

New to HyperFrames? Start here.

GSAP Animation

Animation patterns used in compositions.

Rendering

Render to MP4, MOV, or WebM.

CLI Reference

Full command reference.