From prompt to production
Each job lives in a Finder-visible .studio package with scripts, imported assets, generated outputs, captions, cache files, exports, and a local SQLite project store.
A native macOS creative production environment where AI-assisted image, video, film, and audio work stays inside the project — not scattered across chat threads.
Cast characters with voice identity. Build reusable settings and scene templates. Generate image, video, dialogue, and sound effects from a shared production bible. Every prompt, variant, failure, and decision lives in a portable .studio project you can open, review, and export.
Prompt improvements, variants, caption drafts, pacing proposals, and failed-job diagnoses stay attached to the project for review.
Each job lives in a Finder-visible .studio package with scripts, imported assets, generated outputs, captions, cache files, exports, and a local SQLite project store.
Define characters with visual identity, voice profiles, and speaking style. Build settings with reusable location descriptions and lighting language. Bind ElevenLabs voices to characters so dialogue stays consistent across every generation — not just another prompt, but a cast.
Pull character and setting context into generation specs without pasting raw JSON. Prompt iterations, model settings, reference images, dialogue turns, failures, retries, assistant proposals, and final selects stay recoverable across the life of the project.
Start with a creator-friendly timeline: video lanes, audio lanes, captions, thumbnails, waveforms, trimming, shot order, assistant pacing proposals, and MP4 export.
Studio keeps AI close to the work. Characters, settings, and voice profiles feed into generation. The interface prioritizes preview, timeline, asset state, and export readiness while exposing contextual assistant actions only where they help creators make a concrete next decision.
Define characters with visual identity, voice profiles, and speaking style. Build reusable settings, scene templates, props, and style guides. Pull entity context into generation specs without pasting raw JSON.
Bind ElevenLabs voice profiles to characters for consistent dialogue across shots. Generate sound effects, ambience, and two-speaker dialogue. OpenRouter speech fallback for quick TTS.
Image, video, and audio generation from OpenAI and OpenRouter. Typed creative controls — modes, durations, aspect ratios, reference frames, motion presets — mapped into provider-specific payloads.
Video lanes, audio lanes, captions, thumbnails, and waveforms. Trim, split, reorder, and arrange clips. Contextual assistant actions for pacing, captions, and shot order.
SwiftUI shell with AppKit-backed professional views for dense drag/drop, timeline precision, and keyboard-driven editing. AVFoundation-first playback and export.
SQLite/GRDB persists every edit decision, generation job, and provider payload. Finder-visible .studio packages stay portable and inspectable — your project, your files.
Early access is aimed at creators, marketers, filmmakers, and small studios exploring AI-assisted production with cast, voice, and scene continuity.