Modern Creator Network
Tao Prompts · YouTube · 15:47

Create Seamless AI Films of ANY Length (GPT Image 2 + Seedance 2.0)

Tao Prompts' full pipeline for stitching long AI films from 15-second Seedance clips, using GPT Image 2 storyboards and last-frame chaining.

Posted
1 weeks ago
Duration
Format
Tutorial
educational
Channel
TP
Tao Prompts
§ 01 · The Hook

The bait, then the rug-pull.

Tao Prompts opens by playing the finished pieces — a 44-second fight scene and a 71-second short film — before he names a single tool. The pitch is the proof, not the promise. Then the two title cards drop: GPT Image 2 for the storyboard, Seedance 2.0 for the motion.

§ · Stated Promise

What the video promised.

stated at 00:05I'm showing you exactly how I generated this realistic forty four second long fight scene and how I made the seventy one second long short film, and I'll show you how to extend them even more to be as long as you want.delivered at 15:00
§ · Chapters

Where the time goes.

00:0000:19

01 · Cold open

Plays both finished AI films (44s fight scene, 71s short) before naming any tools. Proof-first hook.

00:1902:55

02 · Step 1: AI storyboard with GPT Image 2

Inside Higgsfield, uploads two reference photos (scientist in hazmat suit + robot companion) and prompts for a 12-panel storyboard at 16:9. Discusses fixing repetition by re-prompting individual panels.

02:5508:48

03 · Step 2: Animate rows with Seedance 2.0

Seedance maxes at 15s, so the storyboard is cropped into rows of 4 shots layered on a 16:9 canvas. Each prompt assigns a per-shot time range using the storyboard's own text descriptions. Adds 'no music, no subtitles' tail. Introduces the 4-column character reference sheet for cross-clip consistency.

08:4809:56

04 · First 45 seconds assembled

Three 15-second clips animated and combined into a continuous 45-second sequence covering the full 12-panel storyboard.

09:5610:44

05 · Step 3: Extend the storyboard

Upload original storyboard + character reference sheets back into GPT Image 2; prompt for the next 12 panels with a continuation hint. Animate the new page the same way for a 90s total (trimmed to 71s for repetition).

10:4415:00

06 · Step 4: Seamless transitions for action scenes

For motion that crosses clip boundaries (a chokehold mid-fight), use the Video Frame Extractor tool to save the last frame of clip N and upload it as the first-frame seed for clip N+1. Demonstrated on the fight-scene example.

15:0015:47

07 · Higgsfield eligibility note + outro CTA

Side note about Higgsfield rejecting some uploaded reference images for copyright reasons (retry tends to work). Plays the full fight-scene example. Ends pointing to his 10-practical-tips video.

§ · Storyboard

Visual structure at a glance.

talking head open
hooktalking head open00:00
71-second proof
hook71-second proof00:10
GPT Image 2 title
promiseGPT Image 2 title00:29
Higgsfield home
valueHiggsfield home00:40
Reference photos
valueReference photos01:04
GPT Image 2 prompt
valueGPT Image 2 prompt01:24
Storyboard prompt
valueStoryboard prompt02:04
12-panel storyboard
value12-panel storyboard02:54
Seedance 2.0 title
promiseSeedance 2.0 title02:55
15s constraint
value15s constraint03:36
Storyboard preview
valueStoryboard preview04:40
Video prompt structure
valueVideo prompt structure04:56
Drift problem
valueDrift problem05:32
Char ref sheet prompt
valueChar ref sheet prompt06:05
Upload references
valueUpload references07:12
Tagged prompts
valueTagged prompts07:44
First animated row
valueFirst animated row08:48
Storyboard page 2
valueStoryboard page 209:56
Page 2 panels
valuePage 2 panels10:44
Last-frame transition
valueLast-frame transition12:30
Higgsfield CTA
ctaHiggsfield CTA10:35
§ · Frameworks

Named ideas worth stealing.

02:55concept

Storyboard → Rows of 4 → Chain by Last Frame

The whole pipeline. Generate a 12-panel storyboard, split into rows of 4 shots, animate each row inside one 15-second Seedance clip with per-shot time ranges, chain clips together using last-frame seeds.

Steal forany 'use a constrained AI tool to make long-form output' tutorial — same shape works for AI music, AI animation, AI voice
03:50concept

Four Shots Per 15-Second Clip

Turn Seedance's 15-second cap into a structural feature by cropping the storyboard into 4-shot strips and prompting the model with explicit time ranges ([00:00-00:04], [00:04-00:07], ...) plus each panel's text description verbatim.

Steal forany constrained-window video tool — Pika, Runway, Kling
05:45concept

4-Column Character Reference Sheet

Use GPT Image 2 with a prompt that specifies four vertical columns showing front view, left profile, right profile, back view of the character. Plain barren background. Reusable as a character bible across image and video generations.

Steal forany AI workflow needing consistent characters across multiple generations
06:45concept

Tagged References in Prompts

Higgsfield lets you @-tag uploaded images inside the prompt (@image_1, @image_2). Lets you say 'the robot @char_ref_sheet walks along @storyboard_1' so the model knows which reference grounds which entity.

Steal forgeneral prompt-engineering pattern: named slots beat positional references
12:30concept

Last-Frame Seed for Seamless Transitions

When motion crosses a clip boundary (chokehold, fall, swing), extract the literal last frame of clip N with the Video Frame Extractor tool and upload it as the first-frame anchor for clip N+1. Eliminates the jump-cut tell.

Steal forany 'stitch short AI clips into a longer sequence' workflow — the #1 fix that separates 'AI demo reel' from 'short film'
04:57concept

Tail-Append 'No Music, No Subtitles'

Standard tail string added to every Seedance prompt so the model doesn't bake music or subtitle artifacts into the clip that you'd have to remove in post.

Steal forany video-gen prompt — clean output is easier to edit
§ · Quotables

Lines you could clip.

03:46
There just isn't enough time inside those fifteen seconds to animate this entire storyboard. So I'm gonna split the storyboard up.
names the central constraint and the central move in one breathTikTok hook for an AI-film short
05:32
We need an additional character reference sheet so that when we generate the long AI video sequence of them, they actually stay consistent throughout the entire scene.
pain-statement plus solutionIG reel cold open
04:42
Just a one sentence description is gonna be enough to create a full storyboard.
low-effort high-output promiseTikTok hook
12:42
Tell the AI to use the screenshot I just saved as the first frame and generate those next four shots starting from that initial screenshot.
names the entire transition-fix in one sentencenewsletter pull-quote
13:33
Using this method, you can generate endless continuous shots for your AI films.
outcome statementoutro card on a short
§ · Pacing

How they spent the runtime.

Hook length32s
Info densityhigh
Filler8%
§ · Resources Mentioned

Things they pointed at.

00:45toolGPT Image 2
02:58toolSeedance 2.0 (ByteDance)
12:30toolVideo Frame Extractor
15:21channel10 Practical Tips for Realistic AI Videos (his prior video)
§ · CTA Breakdown

How they asked for the click.

10:35link
I'm gonna put a link in the description for Higgsfield AI if you wanna go and generate your own long video sequences using GPT image two and Seedance two point o.

Soft affiliate drop right after the 71-second payoff lands — value-delivered first, ask second. End screen at 15:21 cross-promotes his '10 practical tips' video.

§ · The Script

Word for word.

HOOKopening / re-engagementCTAthe pitchmetaphorstory
00:00HOOKHere's the easy way to create AI videos. In this tutorial, I'm showing you exactly how I generated this realistic forty four second long fight scene and how I made the seventy one second long short film, and I'll show you how to extend them even more to be as long as you want. This is by far the simplest and easiest method to generate long AI videos. The first step is to use AI to generate a storyboard just like this one. It's gonna divide our long AI videos into these separate scenes. The tool we're gonna use for this is the newest GPT two image model.
00:32What this tool is amazing at is reasoning and generating text, and that's gonna come in really handy for us when we're trying to generate these lawn storyboards. To use GPT image two, I'm inside Higgs Field AI, and you can find it on the home page, or we can go to the list of image generators and find the new GPT image two model. For the first example I'm creating, I'm gonna generate an lawn AI video based around these two photos
00:59of the scientist in the hazmat suit and his robot companion who are exploring this toxic forest. So we have the GPT image two model chosen, and what I'm gonna do is upload those two reference images of my characters. Let's put those inside here. Next, I'm gonna ask the g p t image two model to create a full storyboard for us. The prompt can be super, super simple. Just a one sentence description is gonna be enough to create a full storyboard.
01:25Now you have some other options, uh, like choosing the quality, the resolution of the storyboard, and also the aspect ratio. I'd recommend setting this to 16 to nine. And then we just have to generate the AI storyboard. This is the full built out story. It's got 12 complete shots in it that tell a full story, and underneath each individual shot,
01:49there's a small text description. Those little text descriptions are gonna be super useful later on because we're gonna use those as part of the prompts for our AI video generator. I did notice sometimes that when you're generating a bunch of separate panels like this, there is some repetition. So if you look at this image in panel three, which is the robot gesturing upwards, it ends up being the same image as shown inside panel 11. So we need to make a minor edit to the storyboard so we don't have this repetition.
02:18So inside Hicksfield, what I'm gonna do is hit reference, which is gonna let me edit this storyboard. And so there's that storyboard added on into our image references. Let's delete all the other ones, um, and also change the prompt. What we wanna do is change
02:36HOOKpanel 11 so that it's not a repetition of panel three, and that's basically exactly what I'm gonna tell the AI to do. Just adjust shot 11 so it's not a repeat of shot number three. And looking at the result, it's swapped shot 11 to one of the scientists, uh, inspecting some kind of a toolkit. So now that we have the storyboard sequence, how do we leverage this and turn it into a lawn AI video? We're gonna use a AI video model called c dance two point o. So if we have a storyboard,
03:08HOOKand what cdance can do for us is to animate all of these in a single video generation at the same time. To use the cdance video model, I'm gonna look through the video models on Higgs field, and the top one is cdance two point o. So for this AI video model, the maximum duration of each video generation happens to be fifteen seconds. So then how do we create a super long AI video sequence
03:34of our entire storyboard? Well, if we try to fit this entire storyboard, all 12 shots into a single fifteen second video, there just isn't enough time inside those fifteen seconds to animate this entire storyboard. So I'm gonna split the storyboard up
03:52and crop out each individual role. So this is the first role I've cropped out where I'm just taking the first four shots of that storyboard. And what we're gonna do is animate these four shots inside the fifteen second video sequence, and it should be able to produce a pretty good result for us. A quick technical point, you will need to layer the cropped roll on top of a 16 by nine image. This is just to make sure that Higgs Field can actually use this image as a reference. For the prompt that I'm gonna use, what I'm gonna rely on is the pregenerated
04:27text descriptions that's already inside the storyboard. So what we're gonna do is tell it to generate a scene using the shots in the uploaded film storyboard. Then for each of the shots inside this fifteen second video generation, I'm gonna give it a time frame of when I want the shot to happen. For example, the first four seconds. And then for the description of what happens in the shot, I'm basically just gonna copy the text description that's already described
04:56in the storyboard, and I'm gonna fill out the rest of the prompt using the same method. You'll also wanna add in this line at the bottom that says no music and no subtitles. I found that that just makes it much easier later on when we're putting everything together. Before we go and use this prompt though, there's an really important addition that we need to add into this to preserve character consistency.
05:21So if we just used the prompt exactly written like this, we'll get results like this, which does basically animate the scene as expected except the robot doesn't look quite the same. For starters in this sequence, the robot's legs look super long for some reason. He should be much shorter than that, and there's a lot of different variation in the way that the character looks between different video generations. We need an additional character reference sheet so that when we generate the long AI video sequence of them, they actually stay consistent throughout the entire scene. To create a character reference sheet, I'm still gonna use the GPT image two model, and I'm gonna upload the original
06:01shot of my robot in the forest. And then in the prompt, I'm gonna tell it to create a character reference sheet. I'm gonna put the prompt in the description so you can go and copy it, and the prompt is gonna give us this really useful character reference sheet to help us maintain consistency inside the videos. Now that we have all these different assets, we can start putting together our long AI video sequence. So first off, I'm gonna upload into Higgs Field the first four shots of our storyboard
06:29along with a character reference sheet of our robot. Let's drag that inside there. And inside the prompt, we're gonna write it exactly like I previously described. So first telling it to generate a scene using the shots inside the uploaded film storyboard. And here, I'm actually gonna make a reference
06:48to the film storyboard, uh, uploaded image. So I'll type at, which lets me tag different references, and then I'm gonna write out each of the individual scenes. Just basically copy in the text descriptions already in the storyboard.
07:05And then inside this prompt, I'm also gonna need to reference our character sheet for the robot as well so that the AI video knows exactly what the robot should look like. So here, I'll add an additional tag next to the word robot
07:21for our character sheet. I'm actually also gonna do this in a few other places as well. And then in the settings, I'm just gonna make sure that I'm using the full fifteen second video duration. Now let's generate the video and see what it looks like.
07:53It actually looks like it animated five shots here instead of the four shots that are prompted for, but that's not a bad problem to have. It still followed all the shots inside the storyboard accurately and the robot looks great. And then using the same exact method, I went ahead and animated the rest of the storyboard. So here's the animation sequence for roll number two of the storyboard, which is shots five through eight.
08:24And here are the last four shots of the storyboard.
08:30And now we have three separate fifteen second video clips that are in sequence to each other, which means that we can combine them together into a full forty five second AI video sequence, but it doesn't stop here. We can actually extend our storyboard as many times as we want to. So looking at the sequence of shots inside the storyboard, what if we wanted to extend this so that they start exploring the forest even more? What we can do is actually use GPT image two to generate the next 12 panels of the storyboard as well. So here's what we're gonna do. Inside the g p t two image generator, we're gonna upload the original storyboard reference.
09:11So this is the first 12 shots that we just animated. And then I'm also gonna upload a character reference sheet for my scientist and also the character reference sheet that I generated for the robot. Then inside the prompt, I'm gonna ask it to simply generate the next 12 panels of the storyboard, and here's the prompt I'm gonna use. Generate the next page of the storyboard
09:32from image one, which continues the story with 12 panels using the uploaded images. The robot reveals hidden knowledge of the forest, guiding the scientists to a deeper, more dangerous core where the source of the toxic outbreak lies. And this is the extended storyboard that GPT
09:50image two has created for us. And then I can use the same exact technique of animating four separate shots at the same time to turn this storyboard into three fifteen second video generations as well.
10:08And if you do the math, using the two storyboard pages, we'll end up with six fifteen second video clips, which is a ninety second video sequence. I found that some of the sequences that I generated were a little bit repetitive, so I ended up trimming it down to a seventy one second
10:26CTAvideo instead. But you can use this method and extend your videos for as long as you want. I'm gonna put a link in the description for Higgs Field AI if you wanna go and generate your own long video sequences using GPT image two and c dense two point o. Now, one of the challenges we're gonna run into this especially when animating a more dynamic action packed scenes, for example, this fight scene that I created, is that because of how much action there is, I found that if we animate
10:58CTAeach of the roles inside the storyboard separately by themselves, it's much harder to combine them together seamlessly. So let's see how we can fix this problem.
11:09First off, for my fight scene, these are the characters and the environment that I want them to be in, and this is the storyboard and the prompt that I used inside GPT image two to create this. Now when I go and animate each row separately of the storyboard, each of the individual animations look really, really good. So this is the first role animated, and keep an eye on what happens in the last scene.
11:36The bounty hunter suddenly ambushes the female character and has her in this chokehold. Now when I go and animate the second row of the storyboard, it starts off with the first frame of them already engaged inside a fight scene. If we look at this individually, it's a really amazing looking AI video. However, if we try to combine
11:57those two separate fifteen second clips together, the transition is gonna look a little weird. So starting with the first video sequence and suddenly jumping into the second video sequence. That transition right there, it doesn't really make sense. She's in a choke hold and suddenly is free and engaged in a fight scene. So then how do we get the transitions to be seamless inside our lawn AI video? What we're gonna need to do is give the AI some extra information
12:27when generating each individual video clip. So first, I'm gonna use this tool called video frame extractor, and I'm gonna save that last image frame of our video sequence. Now when we go and generate the next four shots of our storyboard sequence, actually tell the AI to use the screenshot I just saved as the first frame and generate those next four shots starting from that initial screenshot.
12:54This is what it's gonna look like inside Higgs field first with all of our image references uploaded. And then inside the prompt, I'm gonna tell it to generate a scene using the uploaded film storyboard fight sequence starting with this image frame of the female character getting attacked. And then the rest of the prompt is written just like how it was done before,
13:17basically describing what happens inside each storyboard sequence. Now when we go and generate this scene, it should generate the next four shots starting with the initial image frames. And using this method, you can generate endless continuous shots for your AI films. Here's a quick note about using Higgs field. When you upload image references, what Higgs field is gonna do is check the eligibility
13:44of each image. This is to avoid any copyright issues. Now if you're using, like, a celebrity or a scene from a movie or something, it's gonna get denied.
13:57But sometimes when you upload images of your own characters as well, it can also get denied. So the first time that I tried to upload my reference sheet for my bounty hunter character, it actually got determined as not eligible. But then when I tried again, the next time it was determined
14:18CTAas eligible to be used. So if you try to do this and you upload an image of yourself or your characters and it gets denied, just try uploading that image reference a few different times and eventually it might work. Let's take a look at what the full animated fight scene looks like so you can get an idea of what this method is capable of.
15:21CTAIf you also want a complete breakdown of 10 practical tips to generate the most realistic possible AI videos, go watch his guide right here.
§ · For Joe

Steal the 'turn the limit into the structure' move.

How to teach a constrained AI tool

The 15-second cap was Seedance's biggest problem. Tao turns it into the unit of the entire workflow — and the whole tutorial has a shape because of it.

  • When demoing an AI tool with a hard ceiling (15s clips, 4 image refs, 4k token context), build the lesson around how that ceiling becomes the structural unit. 'Four shots per 15-second clip' is the entire video's spine.
  • Open with the receipts. Play the finished 44s + 71s clips before naming a single tool. Demo-first hook earns the 15 minutes that follow.
  • Name your workflows. 'Character reference sheet,' 'last-frame seed,' 'four-shots-per-clip' are all coinable phrases viewers can re-use. This is how a tutorial becomes a meme that other creators cite.
  • Bury the most valuable thing 70% in. The transition-fix at 10:44 is the part of this video most likely to go viral as a short — Tao left it where the algorithm has to reward the watch-time first. Mod-Boss / JoeFlow tutorials should do the same.
  • Tail-append your prompt boilerplate. 'No music, no subtitles' is a copy-paste rule that gets used 100% of the time. JoeFlow vocab + Mod-Boss session templates already do this; lean harder.
  • Use ratio-of-effort to amplify the pitch: 'one sentence prompt → 12-panel storyboard' is the math that hooks the audience. Always state the leverage explicitly.
§ · For You

If you want to make a 30-second-to-2-minute AI film yourself.

The actual recipe

Here's the recipe for making short AI films with consistent characters and no jump cuts.

  • Use GPT Image 2 (inside Higgsfield) with two reference photos and a one-sentence prompt to generate a 12-panel storyboard at 16:9.
  • Make a 4-column character reference sheet (front / left profile / right profile / back) for any character that needs to stay consistent across shots.
  • Crop the storyboard into rows of 4 shots. For each row, prompt Seedance 2.0 with explicit time ranges per shot ([00:00-00:04], [00:04-00:07], etc.) and the panel's own text description.
  • Always append 'No music, no subtitles' to the prompt.
  • If a shot's motion crosses into the next clip (someone falls, gets grabbed, swings), save the last frame of the previous clip with a frame-extractor tool and upload it as the first frame for the next generation. This is what kills the 'AI demo reel' look.
  • To go longer than ~45 seconds, upload your original storyboard back into GPT Image 2 with the character refs and ask for 'the next 12 panels' — then animate that page the same way.
  • Plan on Higgsfield occasionally rejecting your reference uploads for eligibility — just retry the upload, it usually works on the second try.
§ · Frame Gallery

Visual moments.