---
id: "concept-beat-image-video"
type: "concept"
source_timestamps: ["00:03:42", "00:07:27"]
tags: ["video-editing", "b-roll", "scripting", "storyboarding"]
related: ["concept-higgsfield-mcp", "concept-claude-skills"]
definition: "A workflow that parses a script into sequential visual 'beats' and automatically generates corresponding static images or cinematic video clips."
---
# Beat Image / Beat Video Generation

## Definition

A workflow built as two distinct [[concept-claude-skills]] — **Beat Image Generator** and **Beat Video Generator** — that take a raw script, segment it into visual *beats*, and emit a sequential storyboard of media assets via the [[concept-higgsfield-mcp]].

## How beats are parsed

The Skill is instructed to insert a beat boundary every time:

- the topic shifts,
- a new metaphor or analogy is introduced, or
- the emotional register changes.

Each beat becomes a row in the output storyboard, paired with a generation prompt.

## Beat Image vs. Beat Video

| | **Beat Image** | **Beat Video** |
|---|---|---|
| Output | Static stills | Cinematic motion clips |
| Pace | Fast, flexible | Slow, hero-level |
| Use case | Cutaways, explainer visuals, carousels | Opening hooks, emotional payoffs |
| Volume | High | Low (1–3 per video) |

## Why this works

Visualizing a script is the biggest bottleneck in short-form video production. By embedding pacing rules and style guidelines inside the Skill (and combining with [[concept-claude-projects]] brand context), the output drops straight into an editing timeline with minimal cleanup.

## Caveat (from enrichment)

Auto-segmenting scripts into beats has commercial analogues (auto-B-roll features in tools like Pictory, Descript, etc.). The specific behavior of *this* Skill is creator-defined and not independently corroborated, so treat the implementation as a template rather than a benchmark.
