Curate 5–6 specific posts with high engagement (10K+ likes) and download/screenshot for slides
Higgsfield Cinema Studio — Quick Demo
Feature
What It Does
Workshop Relevance
50+ camera presets
Zoom, pan, dolly, drone, handheld — pick from menu
Eliminates "prompt engineering for camera"
Mood/lighting shift
Same scene → change to golden hour, neon night, moody morning
Show 3 versions of same product shot
Single-photo-to-video
Upload one product photo → animated video with camera move
Fastest path to social content
Viral presets library
Pre-built narrative structures based on high-performing content
"Don't guess what works — use what already works"
Free tier
Available for testing
Works for workshop demo
Competitor awareness (mention briefly):
Runway Gen-3/4 — Better for complex scenes, less camera control
Kling AI — Better motion physics, fewer presets
Pika 2.0 — Faster iteration, viral effects, less cinematic
The key insight: Higgsfield is purpose-built for commerce creators. Its presets encode "what went viral" into one-click templates. The other tools are general-purpose video generators that happen to be usable for commerce.
MODULE 1: AI Text Generation (45 min)
Pre-Made Case
"You're launching a smart water bottle that tracks hydration, glows to remind you to drink, and keeps water cold for 24 hours. Price: $49. Target: health-conscious millennials."
Why this matters: Text feeds everything else (image prompts, video scripts, VO)
—
Deliverables (each participant produces)
Format
Example
Purpose
Instagram caption
"I drank 8 glasses of water every day for a week. Here's what happened 👇"
Social engagement
Ad headline + body
"The Bottle That Won't Let You Forget." + 2-line description
Paid ads, product page
Video script (15 sec)
Shot-by-shot with voiceover script
Feeds into Module 3 + 5
Teaching Points
Hook-first: you have 1.5 seconds on social
AIDA: Attention → Interest → Desire → Action
PAS: Problem → Agitation → Solution (best for UGC)
The script you write here becomes the VO in Module 4 and the video in Module 3
MODULE 2: AI Image Creation (75 min)
Pre-Made Case
"Create a complete visual suite for an indie coffee brand called 'Mornings'. Cold brew, single-origin beans, minimalist packaging. Target: design-conscious coffee lovers, 25–40."
"Casual photo of person holding Mornings coffee bottle, sitting at kitchen table, morning sunlight through window, candid, unposed, phone camera quality, natural skin texture, genuine smile"
Storyboard frame
Cinematic product shot for video ad
"Cinematic wide shot, Mornings cold brew pouring into glass with ice, condensation on bottle, slow motion implied, golden hour backlight, shallow depth of field, 16:9, photorealistic, commercial quality"
COMPETITION: Reverse Prompt Engineering
Format:
You display a pre-generated reference image on screen (product photo, specific style)
Participants write a prompt they think will recreate it as closely as possible
All prompts submitted. You run them all through DALL·E.
Display results side-by-side with reference. Group votes.
Closest match wins.
Why this works:
No one "doesn't know what to make" — the target is right there
Forces precision: every word in the prompt matters
Teaches prompt anatomy better than any lecture
Competition = engagement = retention
Reference images to prepare (3 rounds, increasing difficulty):
Round
Image
Difficulty
Key Test
1
Simple product on white background
Easy
Basic subject + style
2
Product in lifestyle setting with specific lighting
Medium
Lighting + composition descriptors
3
UGC-style with specific "imperfections"
Hard
"Unpolished" aesthetic, camera artifacts
MODULE 3: AI Video Generation (75 min)
Pre-Made Case
"Create a 30-second product demo for 'PulsePods' — wireless earbuds with active noise cancellation. Key features: 36hr battery, water resistant, spatial audio. Price: $129."
Agenda
Time
Activity
Tool
0–10 min
Lecture: Three video generation approaches
Slides
10–18 min
Demo: Text-to-video live
Kling / Runway
18–25 min
Demo: Image-to-video + Start/End Frame technique
Kling / Runway / Higgsfield
25–30 min
Demo: Higgsfield camera presets (mood shift, angle change)
Higgsfield
30–60 min
Exercise: Generate 3 clips using 3 different methods
Kling / Runway / Higgsfield
60–70 min
Review: Best clips, what worked, what didn't
Group
70–75 min
Prep for next: Export clips for Module 5 assembly
—
Three Video Generation Methods
Method
How It Works
Best For
Tool
Text-to-Video
Prompt only. No image input. AI generates everything.
Abstract/product concepts, quick demos
Kling, Runway
Image-to-Video
Upload a keyframe image. AI animates it with motion.
Product shots, controlled starting point
Kling, Runway, Higgsfield
Start + End Frame
Upload image A (start) AND image B (end). AI fills the transition.
Before/after, transformation, reveal
Kling, Runway
Video Clip Assignments
Clip #
Method
Prompt / Input
1
Text-to-Video
"Cinematic close-up of wireless earbuds floating in space, pulsing sound waves visible, dark background with blue light trails. Smooth rotation. 5 seconds."
2
Image-to-Video
Upload: UGC photo from Module 2. Prompt: "Person puts earbuds in, expression shifts to delight as music starts, natural head nod, candid moment."
3
Start + End Frame
Frame A: Earbuds in closed case. Frame B: Earbuds in ears. Prompt: "Case opens, earbuds float out and into ears, smooth motion, tech aesthetic."
Higgsfield-Specific Demo
Load same product image
Apply 3 different camera presets (zoom, pan, drone)
Apply 3 different mood presets (golden hour, studio, night)
Show: "Same product. 9 different looks. Zero prompt engineering."
Teaching Points
Text-to-video = fastest, least control
Image-to-video = most control, needs good keyframe
Start/End frame = best for transformation/reveal narratives
Higgsfield = best for camera control without prompt expertise
Always generate video WITHOUT audio (add audio in Module 4)
MODULE 4: AI Audio Generation (45 min)
Pre-Made Case
"You need two completely different voiceovers for two different products. Product A: LuxeGlow serum (89,premiumskincare).ProductB:BeatBuds(49, budget tech). Voice, tone, pace, and music must match each product's audience."
Agenda
Time
Activity
Tool
0–8 min
Lecture: Voice as brand signal — premium vs casual, energy vs calm
Slides
8–15 min
Demo: Clone voice, generate VO for both products
ElevenLabs
15–25 min
Exercise: Generate Product A VO (luxury skincare)
ElevenLabs
25–35 min
Exercise: Generate Product B VO (budget tech)
ElevenLabs
35–42 min
Exercise: Generate matching music for both
Suno
42–45 min
Export: 2 VO tracks + 2 music tracks → ready for Module 5
—
The Two Voices Exercise
Attribute
Product A: LuxeGlow
Product B: BeatBuds
Voice style
Warm, smooth, aspirational
Energetic, punchy, relatable
Pace
Slow, deliberate
Fast, excited
Emotion tag
"[Warm, sophisticated, intimate]"
"[Energetic, excited, casual]"
Sample script
"Your skin deserves more than hope. It deserves science."
"36 hours of battery. Zero excuses. These just work."
Music prompt
"Luxury spa ambient, soft piano, gentle strings, 70bpm, C major"
"Upbeat tech pop, punchy drums, synth bass, 120bpm, E minor"
Teaching Points
Voice IS your brand. Premium skincare ≠ budget tech.
Emotion tags are not optional — they're the difference between "robot reading text" and "actor delivering a line"
Music tempo and key matter: fast + minor = urgency, slow + major = trust
Audio quality > video quality. Always export 48kHz WAV.
MODULE 5: Commercial & UGC Assembly (60 min)
Pre-Made Case
"You're an affiliate marketer for PulsePods. You need TWO content pieces: (A) A polished 15-second product commercial for TikTok/Reels. (B) A 30-second organic UGC review for the same product — 'I just tried these and honestly...'"
Agenda
Time
Activity
Tool
0–10 min
Lecture: Polished ad vs organic UGC — why you need BOTH
Slides
10–15 min
Demo: Assemble the polished ad on timeline
CapCut / Resolve
15–35 min
Exercise: Build your 15-sec polished ad
CapCut / Resolve
35–45 min
Exercise: Build your 30-sec organic UGC review
CapCut / Resolve
45–55 min
Add affiliate CTAs: where, what, when
—
55–60 min
Export both pieces
—
Two Content Types — Same Product, Different Vibes
Element
Polished Ad (15 sec)
Organic UGC (30 sec)
Opening
Product reveal with motion graphics
"Okay so I just got these..." (casual)
Visual style
Cinematic, color-graded, smooth transitions
Phone-selfie look, slight shake, natural light
Voiceover
Professional VO, scripted
Conversational, "thinking out loud"
Music
Brand-matched, polished
Trending audio or lo-fi
Text overlay
Key features, price
"Link in bio", "Honest review ↓"
CTA
"Available now. Link in bio."
"Not sponsored. I paid for these. Link below if you want to try."
Trust signal
Brand credibility
Personal testimony
Affiliate CTA Formats (teach these)
Platform
CTA Example
Placement
TikTok/Reels
"Link in bio 👆"
Last 2 seconds
Instagram Story
"Tap the link to shop"
Swipe-up sticker
YouTube Shorts
"Check the description"
End screen
All platforms
"Comment 'LINK' and I'll DM you"
Engagement bait
Assembly Checklist (per piece)
Video clips in sequence
VO aligned with visuals
Music layered, ducked during VO
Text overlays legible, timed correctly
CTA clear and placed at peak attention
Color grade: polished ad = consistent, UGC = "natural"
Export: 1080×1920 (9:16 vertical) for TikTok/Reels
MODULE 6: ComfyUI Workflow Builder (60 min)
Pre-Made Case
"You're a content creator who needs to produce 50 product images per week — different angles, different moods, different backgrounds. Build a workflow that does this in one click."
Agenda
Time
Activity
Tool
0–10 min
Lecture: What is node-based AI? Why automation matters for commerce
Slides
10–20 min
Demo: Build a product image workflow live
ComfyUI
20–40 min
Exercise: Load pre-built workflow → modify → run
ComfyUI
40–50 min
Show: Video workflow (Image → AnimateDiff → Save)
ComfyUI
50–60 min
Discussion: Where does this lead? (API automation, batch content, template systems)
—
Pre-Built Workflows (prepare these as .json files)
Workflow
What It Does
Nodes
product-image.json
Text prompt → product image on specified background
Same as above + batch processing with seed variation
product-video.json
Static product image → animated video with camera move
Load Image → AnimateDiff → VAE Decode → Save Video
Teaching Points
The manual work you did today (generate image → download → upload to video tool → download) can be ONE ComfyUI click
Node thinking = assembly line thinking. Each box does one job.
Start simple. Add complexity only when you need it.
The business case: 50 images/week manually = 5 hours. With ComfyUI = 10 minutes.
Figma Weave (cloud, polished) vs ComfyUI (local, unlimited) — same concept, different audience
BONUS: Lip Sync for Commerce (30 min, standalone)
When Commerce Content Needs Lip Sync
Talking-head UGC review → HeyGen or Sync.so
AI avatar product explainer → HeyGen
Character/spokesperson ad → Sync.so or Runway Act-One
Product demo with host → Sync.so
Text/image-only content → Skip (no lips visible)
Quick Demo
Take the UGC video from Module 3
Generate a VO in Module 4 voice
Upload both to Sync.so → download synced result
Compare: unsynced (voiceover) vs synced (lip-matched)
The Rule
Only sync when lips are VISIBLE and the shot is CLOSE.
For everything else: voiceover works.