AI Film Intro

Self-initiated experiment testing whether an AI-first filmmaking workflow can achieve consistent character detail and cinematic continuity

Creative Direction AI Workflow Design Production

"AI filmmaking won't reward prompts. It will reward people who can turn creative direction into a system."

Character close-up: kanji tattoos, pink braids, headset
Full body character design: white streetwear, dynamic pose
Character in purple fog environment: full body pose
Low angle camera: dynamic movement in fog
Lateral close-up: tattoo detail and character consistency
Character in El Solitario MC brandwear
Character in Lacoste editorial
High-fashion editorial: black dress
Futuristic white outfit: kanji wall
Candid: character in supermarket aisle
Candid: bubblegum in supermarket
Fashion runway: character on catwalk
Gentle Monster sunglasses editorial
Character spray-painting JMD on concrete wall: front
Spray-painting JMD: tracking angle

"Consistency is workflow-dependent, not tool-dependent. The differentiator isn't access to the newest model. It's pipeline design, iteration discipline, and taste."

Mad Max world: character with hovering ship in desert
Creature design: front close-up
Creature design: full lateral view
POV from inside ship cockpit: desert chase
Tokyo street: noir envelope exchange
Walking through crowd of clones
Motocross bike in favela
Half-submerged meditation in ocean
Astronaut suit with chrome helmet
Top-down view: character looking up
Pushing wheelchair through supermarket
High-angle candid: character looking up
Retro pixel art game version of the character

A self-initiated experiment proving that AI filmmaking demands workflow design, not just tool access. By treating a deliberately difficult brief (character with fine-grain details) as a production system test, the project delivered coherent cinematic output and a documented pipeline model.

A self-initiated experiment to test whether an AI-first pipeline can produce a short film intro with real cinematic control (not just isolated good-looking frames). The goal was to stress-test character consistency, visual style continuity, environment coherence, and the practical limits of today's generative toolchain.

The result: a fully AI-produced intro film, a repeatable workflow model, and a clear answer. AI filmmaking won't reward prompts. It will reward people who can turn creative direction into a system.

Context & Challenge

AI video is advancing faster than anyone can track. A year and a half before this experiment, we were laughing at a deformed AI video of Will Smith eating spaghetti. Today, keeping up with model releases is nearly impossible. But speed of output isn't the same as quality of craft.

The real problem is continuity. Most AI-generated video still breaks where filmmaking actually matters. Characters drift between shots, micro-details mutate, visual language collapses, and there's no reliable way to maintain coherence across a sequence. Good-looking single frames are easy. A coherent narrative is not.

The test needed to be deliberately hard. I designed a character covered in kanji tattoos as a visual brief that punishes inconsistency. Small deviations are instantly visible. If the pipeline could hold fine-grain detail across multiple generations, it could hold anything.

Strategic Approach

Treat it like a production system test, not a creative experiment. The point wasn't to "make something cool with AI." It was to prove or disprove a pipeline. Can you build a repeatable workflow that gives a creative director actual control over AI-generated moving images?

Define consistency criteria before generating anything. Before a single frame was produced, I established what must stay fixed: character details, tattoo placement, colour palette, lighting logic. And what could vary: camera angle, environment, atmospheric mood. This constraint framework is what made coherence possible.

Separate the pipeline into controllable stages. Instead of throwing prompts at a single tool and hoping, I broke the process into distinct phases, each with its own tool, purpose, and quality gate: Ideation → Character design → Consistency lock → Motion generation → Sound → Edit. Each handoff was a potential failure point. Mapping them was half the work.

Direct for stability, not novelty. I used repeated visual cues and controlled variations to reduce the model's room to "freestyle." The direction was restrictive by design. In filmmaking, consistency is the craft.

Execution

Character Design & Art Direction

Designed the protagonist in Midjourney, establishing the tattoo system, body language, colour world, and visual tone. Every subsequent tool in the pipeline was measured against this reference.

Consistency Development

Used Nano Banana and Seedream 4 to develop and lock the character across poses and contexts, the critical step that most AI workflows skip entirely. This is where coherence is built or lost.

Multi-Model Video Generation

Generated motion across six different models: Veo 3.1, Kling 2.5, LTX, Sora, Runway, and SeedanceWaver (open source). I deliberately tested how each handled consistency, motion quality, and style fidelity. The goal wasn't to pick a winner. It was to understand which tool excels at what, and how to orchestrate them.

Sound Design & Voice

Music sourced from Artlist (Yarin Primak, "Back to Base"). Sound effects produced through Artlist and ElevenLabs. Audio was treated as a directorial layer, not an afterthought. Pacing and emphasis were shaped to match the visual rhythm of the edit.

Editorial Assembly

Edited in iMovie, intentionally lightweight to keep the focus on AI output quality and workflow design rather than post-production rescue. The edit had to prove the pipeline, not compensate for it.

Results & Impact

A fully AI-produced short film intro with consistent character detail, stable visual language, and coherent narrative flow. This proves the pipeline works when the workflow is designed with the same rigour as traditional production.

A practical, documented workflow model covering the full chain: character design, consistency development, multi-model generation strategy, audio production, and editorial logic. This includes a map of failure modes (what breaks first, where handoffs collapse, and why).

A core finding: consistency is workflow-dependent, not tool-dependent. The differentiator isn't access to the newest model; it's pipeline design, iteration discipline, and taste. The same tools that produce incoherent noise in one workflow can deliver controlled, directed results in another.

A strategic read on the moment. This feels like crypto in 2017: chaotic discovery, new tools daily, open-source versus centralisation tensions, and a rapid acceleration curve. The creative landscape is shifting not away from creativity, but toward creativity at scale. Ideas that lived on paper can now move into motion. The next great filmmakers may emerge from the most unexpected places.

Role
Concept · Creative Direction · AI Pipeline Design · Prompt Engineering · Art Direction · Video & Sound Fx Production
Year
2025

A self-initiated experiment to test whether an AI-first pipeline can produce a short film intro with real cinematic control (not just isolated good-looking frames). The goal was to stress-test character consistency, visual style continuity, environment coherence, and the practical limits of today's generative toolchain.

The result: a fully AI-produced intro film, a repeatable workflow model, and a clear answer. AI filmmaking won't reward prompts. It will reward people who can turn creative direction into a system.

Context & Challenge

AI video is advancing faster than anyone can track. A year and a half before this experiment, we were laughing at a deformed AI video of Will Smith eating spaghetti. Today, keeping up with model releases is nearly impossible. But speed of output isn't the same as quality of craft.

The real problem is continuity. Most AI-generated video still breaks where filmmaking actually matters. Characters drift between shots, micro-details mutate, visual language collapses, and there's no reliable way to maintain coherence across a sequence. Good-looking single frames are easy. A coherent narrative is not.

The test needed to be deliberately hard. I designed a character covered in kanji tattoos as a visual brief that punishes inconsistency. Small deviations are instantly visible. If the pipeline could hold fine-grain detail across multiple generations, it could hold anything.

Strategic Approach

Treat it like a production system test, not a creative experiment. The point wasn't to "make something cool with AI." It was to prove or disprove a pipeline. Can you build a repeatable workflow that gives a creative director actual control over AI-generated moving images?

Define consistency criteria before generating anything. Before a single frame was produced, I established what must stay fixed: character details, tattoo placement, colour palette, lighting logic. And what could vary: camera angle, environment, atmospheric mood. This constraint framework is what made coherence possible.

Separate the pipeline into controllable stages. Instead of throwing prompts at a single tool and hoping, I broke the process into distinct phases, each with its own tool, purpose, and quality gate: Ideation → Character design → Consistency lock → Motion generation → Sound → Edit. Each handoff was a potential failure point. Mapping them was half the work.

Direct for stability, not novelty. I used repeated visual cues and controlled variations to reduce the model's room to "freestyle." The direction was restrictive by design. In filmmaking, consistency is the craft.

Execution

Character Design & Art Direction

Designed the protagonist in Midjourney, establishing the tattoo system, body language, colour world, and visual tone. Every subsequent tool in the pipeline was measured against this reference.

Consistency Development

Used Nano Banana and Seedream 4 to develop and lock the character across poses and contexts, the critical step that most AI workflows skip entirely. This is where coherence is built or lost.

Multi-Model Video Generation

Generated motion across six different models: Veo 3.1, Kling 2.5, LTX, Sora, Runway, and SeedanceWaver (open source). I deliberately tested how each handled consistency, motion quality, and style fidelity. The goal wasn't to pick a winner. It was to understand which tool excels at what, and how to orchestrate them.

Sound Design & Voice

Music sourced from Artlist (Yarin Primak, "Back to Base"). Sound effects produced through Artlist and ElevenLabs. Audio was treated as a directorial layer, not an afterthought. Pacing and emphasis were shaped to match the visual rhythm of the edit.

Editorial Assembly

Edited in iMovie, intentionally lightweight to keep the focus on AI output quality and workflow design rather than post-production rescue. The edit had to prove the pipeline, not compensate for it.

Results & Impact

A fully AI-produced short film intro with consistent character detail, stable visual language, and coherent narrative flow. This proves the pipeline works when the workflow is designed with the same rigour as traditional production.

A practical, documented workflow model covering the full chain: character design, consistency development, multi-model generation strategy, audio production, and editorial logic. This includes a map of failure modes (what breaks first, where handoffs collapse, and why).

A core finding: consistency is workflow-dependent, not tool-dependent. The differentiator isn't access to the newest model; it's pipeline design, iteration discipline, and taste. The same tools that produce incoherent noise in one workflow can deliver controlled, directed results in another.

A strategic read on the moment. This feels like crypto in 2017: chaotic discovery, new tools daily, open-source versus centralisation tensions, and a rapid acceleration curve. The creative landscape is shifting not away from creativity, but toward creativity at scale. Ideas that lived on paper can now move into motion. The next great filmmakers may emerge from the most unexpected places.

Role
Concept · Creative Direction · AI Pipeline Design · Prompt Engineering · Art Direction · Video & Sound Fx Production
Year
2025
Next →