DreamDirector: AI-Driven Agentic Platform for Interactive Cinematic Storytelling

Course Instructor

Pramod Gupta

Lead Team Member Affiliation

Computer Science

Abstract

DreamDirector is an AI-driven cinematic storytelling platform designed to generate complete interactive narratives from a single user prompt. The goal of the project is to explore how modern generative AI models can be orchestrated together to create a unified, real-time storytelling experience that includes text, images, video, narration, and adaptive music. Instead of requiring users to have expertise in writing, illustration, audio design, or film production, DreamDirector acts as an end-to-end creative engine that automatically produces the components of a visual-novel style story while allowing the user to influence the narrative through meaningful choices.

At the core of the system is a multi-agent architecture built using LangGraph, where different AI agents collaborate to handle narrative writing, visual consistency, media generation, and mood-aware audio. The Story Director agent manages plot progression, branching choices, and the overall structure of the narrative. The Visual Consistency agent maintains coherent character appearances and scene settings across story segments by using reference-based prompting and image-to-image diffusion. The Media Orchestrator coordinates image generation with SDXL, short cinematic video creation using Google Veo, text-to-speech voice narration through ElevenLabs, and real-time adaptive soundscapes generated with Tone.js. Together, these agents enable DreamDirector to deliver a multimedia story experience that updates interactively as the user makes decisions.

The platform is implemented as a full-stack application using React for the front-end, FastAPI for the back-end, and SQLite for persistent story storage. The user interface features a cinematic presentation mode with transitions, subtitles, film grain effects, and a media gallery that organizes all generated assets. Additional tools such as the Casting Director, Location Scout, and Studio Selector extend the platform beyond narrative generation and into early film pre-production workflows by analyzing characters, suggesting real actors, identifying real-world environments, and aligning stories with professional studios.

This project demonstrates the growing potential of AI to support creative work by simulating a collaborative production pipeline across writing, art, audio, and planning. It also highlights practical engineering challenges involving latency, asynchronous media generation, visual consistency, and multimodal synchronization. DreamDirector illustrates how AI can expand access to storytelling and serve as an assistive companion for creators, students, and filmmakers seeking to rapidly prototype narrative ideas.

This document is currently not available here.

Share

COinS
 

DreamDirector: AI-Driven Agentic Platform for Interactive Cinematic Storytelling

DreamDirector is an AI-driven cinematic storytelling platform designed to generate complete interactive narratives from a single user prompt. The goal of the project is to explore how modern generative AI models can be orchestrated together to create a unified, real-time storytelling experience that includes text, images, video, narration, and adaptive music. Instead of requiring users to have expertise in writing, illustration, audio design, or film production, DreamDirector acts as an end-to-end creative engine that automatically produces the components of a visual-novel style story while allowing the user to influence the narrative through meaningful choices.

At the core of the system is a multi-agent architecture built using LangGraph, where different AI agents collaborate to handle narrative writing, visual consistency, media generation, and mood-aware audio. The Story Director agent manages plot progression, branching choices, and the overall structure of the narrative. The Visual Consistency agent maintains coherent character appearances and scene settings across story segments by using reference-based prompting and image-to-image diffusion. The Media Orchestrator coordinates image generation with SDXL, short cinematic video creation using Google Veo, text-to-speech voice narration through ElevenLabs, and real-time adaptive soundscapes generated with Tone.js. Together, these agents enable DreamDirector to deliver a multimedia story experience that updates interactively as the user makes decisions.

The platform is implemented as a full-stack application using React for the front-end, FastAPI for the back-end, and SQLite for persistent story storage. The user interface features a cinematic presentation mode with transitions, subtitles, film grain effects, and a media gallery that organizes all generated assets. Additional tools such as the Casting Director, Location Scout, and Studio Selector extend the platform beyond narrative generation and into early film pre-production workflows by analyzing characters, suggesting real actors, identifying real-world environments, and aligning stories with professional studios.

This project demonstrates the growing potential of AI to support creative work by simulating a collaborative production pipeline across writing, art, audio, and planning. It also highlights practical engineering challenges involving latency, asynchronous media generation, visual consistency, and multimodal synchronization. DreamDirector illustrates how AI can expand access to storytelling and serve as an assistive companion for creators, students, and filmmakers seeking to rapidly prototype narrative ideas.