Search Phrase = clone
Automated Video Production Pipeline
This video guides you through setting up an automated video production pipeline, from selecting and testing brand voices using Eleven Labs to pairing them with digital avatars in HeyGen. By following the steps, you'll learn how to catalog and integrate voices, match them with visual characters, and generate preview videos for evaluation. Once you complete the video, you'll be able to efficiently create, test, and organize multiple spokesperson options for your brand's automated content generation. This process empowers you to streamline video production and build a scalable library of branded video assets.
Following are the key things you will be able to do after you watch this demo:
Identify suitable brand voices using generative AI tools.
Catalog and organize voice and avatar options for efficient selection.
Integrate third-party voices into video production platforms.
Pair voices with digital avatars to create compelling spokesperson combinations.
Generate and preview automated video content for evaluation.
Document and track production assets for streamlined workflow.
Select and finalize top spokesperson options for automated content generation.
Introduction to Automated Video Production Pipeline (00:00:00 – 00:00:59)
Josh kicks off the demo by outlining the goal: selecting brand-aligned voices and digital doubles (either your own clone or hired actors), organizing those assets, and laying out the end-to-end steps needed to spin up a fully automated video production pipeline.
Content Sequencing Concept and Cloning (00:00:59 – 00:02:20)
He explains the core idea of building a repeatable sequence of content—cloning a finished production over and over—so you can continually generate new videos by plugging different scripts into the same automated workflow.
Defining Digital Doubles and Voice Types (00:02:20 – 00:03:11)
Josh clarifies terminology (digital twin vs. digital double), walks through the two main “buckets” of voice assets (personality-based clones vs. spokesperson avatars), and discusses how to mix and match them depending on your brand needs.
Selecting Platforms for Generative AI and Deployment (00:03:11 – 00:04:00)
He emphasizes the importance of vetting your generative-AI tools—voice engines and video avatars—and making sure they’re compatible with your target platforms before committing to any given solution.
Brand-Focused Workflow and SRT Utilization (00:04:00 – 00:05:25)
Josh decides to focus on one streamlined method for this demo, using a single SRT transcript file as the “source of truth” for automation—underscoring that a clean, well-formatted SRT is absolute gold when you’re architecting an automated pipeline.
Importing SRT and Leveraging Automation (00:05:25 – 00:07:40)
He shows how to import the SRT into the voice-generation platform, highlighting how the time-coded script drives every subsequent step—from audio rendering to scene assembly.
Setting Up Voice Design in ElevenLabs (00:07:40 – 00:11:49)
A step-by-step walkthrough of testing voice presets, tweaking text lengths, integrating third-party voices, and crafting voice-design prompts to nail down the exact tone and style you want.
Managing Credits and Reviewing Generated Audio (00:11:49 – 00:15:46)
Josh demonstrates how to monitor and conserve your generation credits, preview the rendered audio, swap out placeholder text, and ensure you’re only spending resources on polished clips.
Applying Voiceover and Text Overlays to Video (00:15:46 – 00:19:08)
He attaches the finalized voice track to the video timeline, adds and styles text overlays (centering, contrast adjustments), and assembles the basic video composition ready for export.
Enhancing Prompts with AI Tools for Voice Design (00:19:08 – 00:22:04)
Introduces additional AI utilities for brainstorming and refining your voice-design prompts—showing how to iterate until you get a sample that truly matches your brand voice.
API Key Handling and Asset Export Configuration (00:22:04 – 00:27:28)
A practical guide on securely copying your ElevenLabs API key, configuring export settings (e.g., 4K output), and organizing all generated files into branded folders for easy access.
Frame Rate Considerations and Quality Checks (00:27:28 – 00:31:42)
Notes the default 25 fps setting, explains how frame rate impacts perceived motion, and walks through checking your export quality to avoid any unexpected artifacts.
Avatar Adjustments, Project Naming, and Fallbacks (00:31:42 – 01:05:16)
Josh covers fine-tuning avatar scale and positioning, updating project names for consistency, and setting up fallback workflows if you need to swap voices or visuals mid-pipeline.
Avatar Replacement and Cataloging (00:31:42 – 00:34:06)
Pair your chosen voice with visuals by replacing the default avatar, browsing through the 21 “looks” in each category, using the snipping tool to capture promising thumbnails, and logging each candidate’s name and category in your tracking spreadsheet.
Avatar Testing and Video Formatting (00:34:07 – 00:36:24)
Brainstorm voice–visual combinations (e.g. “August”), select a portrait-mode avatar, preview the static image, upload any custom avatars into the pipeline, drag your source video beneath the avatar layer, and confirm the composition and framing.
Voice-Avatar Sync and Quality Comparison (00:36:24 – 00:37:39)
Generate audio samples to compare HeyGen vs. ElevenLabs quality, force-refresh the clip to confirm it’s using the intended voice (e.g. Ryan Kirk), and watch for the spinning indicator to verify successful render.
Preview Generation and File Labeling (00:38:10 – 00:39:11)
Render a 4K preview of the voice-avatar pairing, then label the export asset with your convention (e.g. 001_RyanKirk_CharlieAvatar) so each test remains organized and easily identifiable.
Pipeline Duplication for Variant Testing (00:39:11 – 00:41:15)
Duplicate the entire sequence to create “Test 002,” swap in a new avatar (such as Colton), explore lifestyle/UGC categories, and note how background removal and frame size affect the final look.
Background Removal and Frame Adjustments (00:41:15 – 00:42:32)
Apply the background-remover tool to avatars with built-in backgrounds, observe any cut-offs (like arms being cropped), tweak the canvas framing, and decide between static vs. transparent backgrounds based on brand needs.
Third-Party Voice Integration Workflow (00:42:32 – 00:44:03)
In the “My Voices” tab, toggle on integrated voices (e.g. Charlie), heart your favorites so they surface first, preview each sample, and ensure the API integration is active before proceeding.
Voice Audition Labeling and Mood Board Documentation (00:44:03 – 00:47:09)
Name each audition (e.g. 002_CharlieAvatar), update your mood board with snipped thumbnails, record which browser tab or category each came from, and keep this documentation up to date for reproducibility.
Frame Rate and Credit Management (00:47:09 – 00:48:06)
Note the default 25 fps setting—mismatches can cause audio sync issues—toggle off “Avatar 4” if you’re on an unlimited plan, and monitor your generation credits to avoid unexpected limits.
Styling and Folder Organization (00:48:06 – 00:49:29)
Adjust text overlay colors to maintain contrast (match your brand palette), create new folders for each batch, and standardize your output directory structure so you know exactly where each rendered clip lives.
Option Preview and Cataloging Workflow (00:49:30 – 00:55:51)
Refresh thumbnails, scroll through voice-avatar combos, assign option numbers, screenshot grids of candidates, and log each pairing’s status (“Yes,” “Maybe,” “No”) in your spreadsheet.
Iteration Process and Consistency Notes (00:55:51 – 00:57:23)
Always regenerate every variation (never reuse stale renders), note any limitations (e.g. animated text can cover on-screen elements), and keep your naming and documentation consistent so the pipeline remains bullet-proof.
Ranking Options and Visual Separators (00:57:24 – 01:02:40)
Introduce visual separators in your catalog (e.g. blank rows), rank the top voice-avatar combos, screenshot your “definite yes” list, and preserve those as templates for future batches.
Additional Voice Integration: Amelia (01:02:40 – 01:04:33)
Search for “Amelia” in your voice library, verify whether it’s built-in or needs third-party integration, add it to favorites, preview the sample, and record its ID for consistent reuse.
Final Voice Candidate Integration (01:04:33 – 01:05:16)
Confirm Amelia’s render, then search for any last candidates (e.g. “Analore”), heart and test them, catalog the results, and ensure each new voice is fully integrated into the pipeline.
Final Pipeline Recap and Scale Duplication (01:07:40 – 01:08:34)
Recap how you’ve selected your final set of voices and avatars, finalize your naming conventions, and highlight that you can now duplicate this entire automated workflow to churn out an endless library of on-brand social-media videos.
Keywords: audio,recording,microphone,quality,live,studio,interface,phantom,power,sample,rate,uncompressed,format,pop,filter,level,balancing,Camtasia,Studio,file,organization,voice,clone,AI,avatar,sound,absorption
This video provides a comprehensive guide to professional audio recording for content creators, focusing on essential equipment and techniques for high-quality sound production. Viewers will learn how to select the right microphone, set up a proper recording environment, and use audio interfaces and editing tools to capture clean, professional-grade audio. By following Josh Lomelino's expert advice, participants will be able to create polished audio recordings suitable for workshops, demos, podcasts, and even AI-generated video content. The tutorial equips creators with practical skills to improve their audio recording process and produce more engaging, professional-sounding content.
This video provides a comprehensive guide to professional audio recording for content creators, focusing on essential equipment and techniques for high-quality sound production. Viewers will learn how to select the right microphone, set up a proper recording environment, and use audio interfaces and editing tools to capture clean, professional-grade audio. By following Josh Lomelino's expert advice, participants will be able to create polished audio recordings suitable for workshops, demos, podcasts, and even AI-generated video content. The tutorial equips creators with practical skills to improve their audio recording process and produce more engaging, professional-sounding content.
Here are the key things you will be able to do after you watch this demo:
Select an appropriate high-quality microphone for professional audio recording
Set up a clean, noise-free recording environment
Configure audio interfaces and software for optimal sound capture
Choose the correct sample rate and recording format
Use a pop filter and mic positioning techniques to improve audio quality
Perform audio test recordings and evaluate sound levels
Utilize audio editing tools for recording and post-production
Implement file organization strategies for audio projects
Export audio files in various formats for different content needs
Create consistent, professional-grade audio recordings for workshops, demos, and presentations
Prepare audio recordings for potential AI avatar or voice clone generation
Troubleshoot common audio recording and equipment setup challenges
Basic Method of Production 0:09
Josh Lomelino explains the simplicity and power of recording thoughts and ideas using just a microphone.
Live recordings during workshops or demos are more engaging but harder to edit if mistakes are made.
Studio recordings allow for pauses and polished takes but require maintaining a natural and conversational tone.
The importance of a high-quality microphone and a quiet, clean recording space is emphasized.
Microphone Setup and Recording Quality 1:31
Josh recommends the AKG condenser mic for its clean, detailed sound, which requires phantom power.
The Shure 57 microphone is mentioned as a versatile option for various recording situations.
The Zoom H6 USB audio interface is preferred for its compatibility with various software like Camtasia.
Recording at 48,000 Hz instead of the default 44.1 Hz is suggested to preserve audio detail.
Audio Recording Practices 3:18
Josh advises recording in an uncompressed format like WAV until the final export to avoid audio degradation.
Ensuring the computer and audio interface are set to the same sample rate prevents speed mismatches.
The use of a pop filter and an adjustable mic arm helps maintain consistent audio quality.
Test recordings and listening on different devices help ensure balanced sound levels.
Audio Editing and Tools 4:53
Josh mentions various audio editing tools like Audacity, Adobe Audition, Pro Tools, and FL Studio.
Camtasia Studio is recommended for its convenience in recording and managing audio projects.
The Auto Normalize feature in Camtasia helps maintain consistent volume throughout recordings.
Exporting recordings as MP3s allows for generating on-camera videos using AI avatars.
File Organization and Studio Setup 5:55
A consistent naming system for recordings and exports is crucial for easy retrieval and updates.
Avoiding rooms with echo and using soft materials to absorb sound helps improve recording quality.
A good studio setup, including soundproofing and proper equipment, is essential for high-quality recordings.
Josh hints at a future demo on creating a voice clone, which requires clean and consistent audio recordings.
In this video, Josh Lomelino demonstrates how to create an AI-powered digital voice replica using 11 Labs, enabling content creators to rapidly generate high-quality audio and video content at scale. By training the system with a consistent audio sample, users can produce automated voice performances that sound like their own, allowing them to create lectures, demos, and other content quickly and efficiently. The method involves uploading 1-3 hours of controlled audio recordings, fine-tuning voice settings, and integrating with platforms like HeyGen to automate video production. After watching this tutorial, viewers will be able to develop their own AI voice clone, streamline content creation, and overcome time constraints by generating multiple scripts and videos with minimal manual effort.
In this video, Josh Lomelino demonstrates how to create an AI-powered digital voice replica using 11 Labs, enabling content creators to rapidly generate high-quality audio and video content at scale. By training the system with a consistent audio sample, users can produce automated voice performances that sound like their own, allowing them to create lectures, demos, and other content quickly and efficiently. The method involves uploading 1-3 hours of controlled audio recordings, fine-tuning voice settings, and integrating with platforms like HeyGen to automate video production. After watching this tutorial, viewers will be able to develop their own AI voice clone, streamline content creation, and overcome time constraints by generating multiple scripts and videos with minimal manual effort.
Here are the key things you will be able to do after you watch this demo:
Train an AI voice synthesis system using personal audio recordings
Generate consistent voice replicas with controlled audio samples
Optimize AI-generated voice settings for natural-sounding output
Integrate voice cloning technology with video production platforms
Create automated content at scale using text-to-speech technologies
Manage AI voice generation credits efficiently
Export and store audio files in multiple formats for different applications
Prototype and refine scripts using AI voice technology
Develop a workflow for rapid content creation across lectures, demos, and presentations
Leverage AI tools to overcome time constraints in content production
Creating a Voice Replica Using AI 0:09
Josh Lomelino discusses the use of AI-powered voice synthesis to create a voice replica, emphasizing the challenge of matching human recordings.
He highlights the effectiveness of using text prompts to quickly prototype, test, and revise scripts or generate finished audio files.
Josh mentions his preference for the 11 labs tool, which offers a studio mode for producing longer form audio tracks.
He shares his initial struggles with the tool and how contacting their support provided helpful suggestions.
Training the System for Consistent Output 1:24
Josh explains the importance of training the system with a consistent audio sample to avoid unnatural variations in volume and tone.
He describes his initial mistake of using diverse recordings from different sessions, which led to inconsistent results.
Josh emphasizes the need for a controlled environment with a single, consistent audio sample for better results.
He plans to demonstrate the settings that produce the best results for replicating his voice in the user interface.
Optimizing Generated Audio Files 2:56
Josh advises generating audio sparingly to avoid exhausting monthly credits and recommends starting with smaller sections of text.
He explains the process of refining the output and generating both wave and mp3 audio files for different applications.
Josh mentions the importance of storing both wave and mp3 files for secure storage and project organization.
He notes that it may take several attempts to develop a method that works well for the user.
Exporting and Integrating Audio Files 4:19
Josh describes two methods for uploading audio files to virtual avatars: exporting both wave and mp3 versions or integrating the 11 labs API directly with Hey Gen.
He prefers using the wave audio file for higher quality and to avoid double compression but acknowledges the need to export the mp3 format for larger tracks.
Josh explains the integration of the 11 labs API with Hey Gen, which allows for rapid development of prototypes and large volumes of content.
He mentions the need to break up scripts into manageable sections for efficient processing by the software.
Automating Video Production with AI 6:02
Josh discusses the ability to produce videos at scale by automating both audio and video avatars from text.
He highlights the productivity gains from using AI to generate video scripts and produce audio and video automatically.
Josh notes the cost of AI-generated voice and the strategy of using high-quality audio only when necessary.
He explains the use of draft versions of scripts with Hey Gen's voice replica to refine the script without incurring additional costs.
Finalizing and Exporting Scripts 8:04
Josh describes the process of finalizing scripts and either reading and recording them manually or using the 11 labs integration within Hey Gen.
He mentions the use of a side-by-side display setup with a Google document and video avatar performance for quick edits.
Josh emphasizes the usefulness of this method for high-end projects that require detailed polishing and iteration.
He concludes the demo by encouraging the use of digital voice replicas to scale beyond time constraints and improve productivity.
Keywords: automation,content,creation,production,studio,digital,doubles,video,avatar,text,script,cloud-based,tools,slide,decks,PowerPoint,Canva,training,programs,staff,development,retention,coding,Academy,method,four
Method Four of the Ultimate Content Creation Workflow enables creators to automate their entire video production process by leveraging cloud-based tools and digital technology. By mastering this method, content creators can clone their voice, generate video avatars, and produce high-quality training videos and presentations with minimal time and effort. The workflow allows you to transform a simple text script into a fully automated video production, complete with synchronized audio, visuals, and slide decks. Ultimately, this approach empowers busy professionals to scale their content creation without being constrained by traditional time-consuming production methods.
Method Four of the Ultimate Content Creation Workflow enables creators to automate their entire video production process by leveraging cloud-based tools and digital technology. By mastering this method, content creators can clone their voice, generate video avatars, and produce high-quality training videos and presentations with minimal time and effort. The workflow allows you to transform a simple text script into a fully automated video production, complete with synchronized audio, visuals, and slide decks. Ultimately, this approach empowers busy professionals to scale their content creation without being constrained by traditional time-consuming production methods.
Here are the key things you will be able to do after you watch this demo:
clone your voice for digital content creation
Generate automated video avatars
Transform text scripts into complete video presentations
Automate slide deck production in PowerPoint and Canva
Scale content creation with minimal time investment
Develop training materials efficiently
Leverage cloud-based production tools
Create digital doubles of yourself
Streamline video production workflows
Produce high-quality educational content without extensive technical skills
Ultimate Content Creation Workflow Overview 0:08
Josh Lomelino introduces method four, which automates the entire content creation process.
This method combines the first three methods but focuses on automation, making it more efficient.
Josh emphasizes the importance of mastering the first three methods before attempting method four.
The method allows for the creation of high-quality content with minimal time, effort, and budget.
Method Four's Impact on Production 1:09
Josh describes the transformative power of method four, which revolutionized his production process.
A potential customer expressed interest in using the method for staff development and retention.
Josh explains how he creates digital doubles of himself to automate the production process.
The method enables large-scale production without the time constraints typically associated with video creation.
Addressing Time Constraints in Content Creation 1:49
Josh shares experiences of customers who face time constraints in creating training programs and classes.
He highlights the challenges of maintaining a busy schedule while keeping up with production demands.
Method four allows for the cloning of voices and creation of audio tracks to generate video avatars.
The method significantly reduces the time required to produce multiple videos.
Automation Capabilities of Method Four 2:29
Josh explains that everything in the final video is fully automated, starting from a text script.
The process involves copying and pasting the script into cloud-based production tools.
High-end computers are not necessary as most of the heavy lifting is done in the cloud.
The method also automates the creation of slide decks in tools like PowerPoint or Canva.
Step-by-Step Process Walkthrough 2:48
Josh mentions that he will walk through each part of the process in the following sections.
The detailed steps will provide a comprehensive understanding of method four.
The process aims to make content creation more efficient and less time-consuming.
Josh emphasizes the importance of understanding each step to effectively implement the method.
Keywords: Automated, performance, text, video, Otter, AI, voice, clone, Eleven Labs, HeyGen, audio, multilingual
In this video, Josh demonstrates how to create fully automated video performances directly from text using tools like Otter AI, 11 Labs, and HeyGen. Viewers will learn how to generate high-quality voice clones, prototype video scripts, and produce professional-looking content with minimal effort by leveraging AI-powered voice and video generation technologies. The workflow allows content creators to transform written or spoken text into polished video presentations quickly and efficiently. By following Josh's method, users can generate multiple video iterations, edit audio precisely, and create digital avatars that replicate their voice and performance with remarkable accuracy.
In this video, Josh demonstrates how to create fully automated video performances directly from text using tools like Otter AI, 11 Labs, and HeyGen. Viewers will learn how to generate high-quality voice clones, prototype video scripts, and produce professional-looking content with minimal effort by leveraging AI-powered voice and video generation technologies. The workflow allows content creators to transform written or spoken text into polished video presentations quickly and efficiently. By following Josh's method, users can generate multiple video iterations, edit audio precisely, and create digital avatars that replicate their voice and performance with remarkable accuracy.
Following are the key things you will be able to do after you watch this demo:
Generate video scripts from transcribed audio using AI tools
Create high-quality voice clones with consistent audio recordings
Prototype video content using free and paid AI platforms
Optimize voice training for digital avatars
Manage content production across multiple AI environments
Edit audio tracks with minimal credit consumption
Develop a systematic workflow for automated video creation
Replicate personal performance using digital voice technology
Transform text-based content into professional video presentations
Implement cost-effective strategies for video and audio generation
Creating a Fully Automated Performance from Text 0:08
Josh Lomelino explains the process of creating a fully automated performance directly from text, including generating audio prompts using Otter AI.
He describes how he brainstorms ideas while walking and exports the subtitle transcript file, SRT, to process it with AI tools like Claude or ChatGPT.
Josh mentions breaking up long scripts into manageable blocks of 1800 characters and generating a year's worth of content for various platforms.
He emphasizes the use of text, whether written manually or spoken and transcribed, to craft a video script using two primary methods.
Generating High-Quality Voice clones 1:51
Josh discusses creating a high-quality voice clone using 11 Labs, initially finding the results artificial but later perfecting the settings.
He highlights the importance of using a consistent audio clip for training the voice digital double, ideally around three hours of spoken audio.
Josh explains the challenges of recording consistently for three hours and how he stitches together previous demo recordings to create a large audio clip.
He stresses the need for meticulous tracking of audio settings to ensure uniformity and avoid sudden changes in volume or tonal quality.
Optimizing Audio Recording for Consistency 3:36
Josh shares his experience of recording multiple live sessions with an audience, which infused the audio with personality and energy.
He explains the importance of having consistently dialed-in audio for generating a high-quality performance, as the AI listens to everything in the audio track.
Josh mentions the time and cost involved in using 11 Labs, which can take up to six to eight hours to analyze a voice and build a model.
He advises against using cheaper models, such as the multilingual version one model or turbo 2.5, and recommends upgrading to the multilingual version two model for better results.
Using Hey Gen for Cost-Effective Prototyping 5:35
Josh introduces Hey Gen as an alternative for creating generative content when 11 Labs burns through credits too quickly.
He explains how he trains Hey Gen on his voice by uploading a 10 to 15-minute audio clip and generates unlimited videos for free, depending on the subscription plan.
Josh describes the process of creating prototypes, making real-time adjustments to the script, and rendering multiple takes.
He mentions using his phone in split screen mode while walking to make adjustments on the fly and then copying and pasting the revised script into Hey Gen.
Switching Between Hey Gen and 11 Labs 7:44
Josh explains how he can switch the voice in Hey Gen to the high-quality production voice in 11 Labs with a click of a button.
He highlights the downside of using Hey Gen, which is the risk of losing all credits if there are issues with the audio track in the final video.
Josh prefers using the Studio tool in 11 Labs for targeted editing, which allows regenerating just portions of the audio without redoing the entire clip.
He mentions the benefit of being able to download the WAV file and MP3 file from the Studio tool in 11 Labs as a fail-safe.
Organizing Video Production Phases 9:21
Josh describes his workflow of treating production as two phases: the cheap, free voice phase and the final phase.
He explains the process of pasting the text directly into the Hey Gen editor, listening to the prototype, and resolving issues before creating a new file in Hey Gen.
Josh organizes his videos into two folders: a prototype folder and a final folder, for easy organization of his methods.
He mentions using the multilingual version two model for cost-effective throwaway tests and training his voice with Hey Gen for free prototyping.
Leveraging Digital Doubles for High-Quality Videos 10:34
Josh shares how he uses his digital doubles to replicate a performance of his voice and generate a corresponding video composite.
He explains how he creates a script using Otter AI during a walk, copies and pastes it into his automated workflow, and produces a high-end video with minimal effort.
Josh highlights the benefits of this workflow, which allows him to deliver excellence without skipping a beat, even when small inconsistencies would have derailed the process before.
He concludes by mentioning the next steps in the following videos, which will cover adding automated visual elements on screen behind the virtual avatar.
Keywords: AI, Claude, Chat GPT, brainstorming, video, script, otter, SRT, transcription, generative audio, bulk export, workflow
Generate Ideas with Otter and Claude
Josh demonstrates how to use AI tools like Otter AI, ChatGPT, and Hey Gen to quickly transform brainstorming transcripts into polished video scripts. By leveraging AI's capabilities, creators can capture their ideas, generate scripts, and create content with minimal manual editing. The workflow allows users to convert spoken thoughts into text, refine the script through AI assistance, and produce a final video with a digital avatar or voice clone. Viewers will learn a streamlined process for content creation that dramatically reduces production time and enables rapid, creative video generation.
Following are the key things you will be able to do after you watch this demo:
Capture brainstorming ideas using Otter AI transcription
Export SRT files from recorded thoughts
Convert raw transcripts into structured video scripts
Leverage AI tools to refine and edit content automatically
Break down long scripts into manageable character blocks
Identify and correct potential AI pronunciation challenges
Generate video scripts with minimal manual editing
Prepare scripts for digital avatar or voice clone production
Batch process multiple transcripts simultaneously
Create content at scale using AI-assisted workflows
Using AI Tools for Content Creation 0:09
Josh Lomelino explains how AI tools help him capture ideas and generate content directly from brainstorming sessions.
He uses Otter AI to record his thoughts verbatim, which he then exports as an SRT file for transcription.
The SRT file contains every word spoken along with time codes, making it easy to generate a full video script.
Josh leverages AI tools like 11 Labs and Hey Gen to produce audio and video content from the transcribed text.
Generating Video Scripts from Transcripts 2:00
Josh describes the process of generating a video script from the transcribed text using AI tools.
He explains the difference between having a clear plan and a vague notion for the script.
The AI can capture random ideas and generate multiple scripts within the Otter AI application.
Josh then uses tools like Claude AI or ChatGPT to expand and refine the generated scripts.
Collaborative Writing with AI 2:35
Josh aims to create a video script that his digital double can read aloud, reducing the need for extensive editing.
He explains the collaborative writing process between himself and AI tools to generate drafts and revisions.
The ultimate goal is to use AI to create a polished video script without spending hours on manual editing.
Josh emphasizes the importance of spending time to perfect the AI prompting process.
Workflow for Converting SRT Files 3:51
Josh demonstrates the workflow for converting an SRT file into a video script using Otter AI and Notepad.
He highlights the importance of checking the prompts document for time-saving methods.
Josh explains two methods for creating video scripts: word-for-word transcription and general direction.
He provides detailed prompts for ChatGPT to convert SRT files into 1800-character blocks.
Handling Rough Brainstorming Transcripts 7:40
Josh discusses handling rough brainstorming transcripts that require more assistance from AI tools.
He explains the need to be mindful of checking each word when using AI to generalize the transcript.
Josh provides a prompt for ChatGPT to convert the SRT file into a video script and fix grammatical issues.
He emphasizes the importance of ensuring the script is readable by the AI digital double.
Challenges with AI-Generated Scripts 10:06
Josh mentions potential challenges with AI-generated scripts, such as mispronunciation by the digital double.
He explains the time-consuming process of manually correcting AI-generated scripts.
Josh introduces a prompt for a cleanup pass to automatically correct readability issues.
He advises copying and pasting the corrected script into the video script document for backup.
Finalizing the Video Script 12:23
Josh explains the final steps of rendering the script as a prototype using a free voice clone.
He advises listening to the playback and adjusting the script for pronunciation issues.
Once satisfied with the prototype, the final audio can be generated using tools like 11 Labs.
The final audio clip can then be uploaded to a virtual avatar software for the final on-screen performance.
Batch Processing Multiple SRT Files 13:21
Josh highlights the option to bulk export multiple SRT files from the Otter AI app for time savings.
He explains how this process can be applied to a whole folder of SRT files.
This method allows for the creation of massive amounts of content quickly and easily.
Josh concludes the demo by encouraging viewers to try the process for themselves.
Discover how to unlock your product’s potential with this hands-on demo! Learn to identify your audience’s biggest challenges, craft compelling scripts using leading marketing frameworks, and leverage AI-powered tools to create engaging vision videos. Walk away ready to prototype voiceovers, iterate on creative ideas, and connect with your audience through actionable storytelling that drives real results.
Discover how to unlock your product’s potential with this hands-on demo! Learn to identify your audience’s biggest challenges, craft compelling scripts using leading marketing frameworks, and leverage AI-powered tools to create engaging vision videos. Walk away ready to prototype voiceovers, iterate on creative ideas, and connect with your audience through actionable storytelling that drives real results.
This video guides viewers through recognizing and addressing key challenges like lack of clarity, inconsistency, and information overload. By following the step-by-step vision presented, viewers will learn how the app helps them transform these obstacles into opportunities for personal growth and productivity. After watching, audiences will be equipped to download the app, leverage its key features to build better habits, and take actionable steps toward positive change. The video empowers viewers to begin their own transformation journey right away.
Following are the key things you will be able to do after you watch this demo:
Creating a Vision Video Using Marketing Frameworks 0:10
Josh Lomelino explains the initial steps for creating a vision video, emphasizing the importance of the Ray Edwards framework.
The process involves identifying and amplifying pain points, telling a story, and transforming the narrative to lead to a call to action.
Josh introduces the Jeff Walker framework, which follows a similar pain-agitate-solve structure.
He discusses the use of ChatGPT to unearth pain points and personas, integrating this information into the script writing process.
Script Writing and User Problems 5:13
Josh details the process of writing a script using the Ray Edwards framework, focusing on the top three common problems.
He lists the top three problems: lack of clarity, inconsistency, and lack of accountability.
The script aims to show a transformation from pain to breakthrough, with a vision video lasting two to three minutes.
Josh emphasizes the importance of defining marketing before finishing the product to connect with the audience effectively.
Iterating the Script and Using Generative AI 10:44
Josh explains the process of creating multiple versions of the script, using ChatGPT and Claude AI for brainstorming and refining.
He highlights the importance of providing detailed instructions to the AI tools to ensure they stay within the desired framework.
Josh discusses the use of teleprompter scripts to ensure the spoken words are accurate and readable.
He mentions the use of 11 Labs for generating voiceovers, which helps in prototyping and refining the script.
Finalizing the Script and Preparing for Video Production 27:00
Josh talks about the importance of testing different versions of the script with focus groups to get valuable market feedback.
He explains the process of creating a Google Doc to keep track of different versions of the script and related content.
Josh introduces the Jeff Walker framework, which is used for product launches, and compares it with the Ray Edwards framework.
He discusses the final steps of creating the vision video, including generating animatics, storyboards, and visual content.
Generating Audio and Selecting Voices 36:23
Josh demonstrates the use of 11 Labs to generate audio performances from the script, using his own voice as a clone.
He explains the process of selecting and applying different voices from the 11 Labs library to experiment with different tones and styles.
Josh highlights the importance of exporting the audio in WAV format for higher quality and flexibility in editing.
He discusses the potential use of multiple voices to create a cast of characters in the vision video.
Editing and Refining the Vision Video 58:53
Josh outlines the next steps for editing the audio and video content, including creating animatics and storyboards.
He emphasizes the importance of aligning the visuals with the audio track to ensure the narrative flows smoothly.
Josh discusses the use of AI-generated video content for B-roll footage to show the app in use.
He concludes by summarizing the overall process of creating a vision video, from script writing to final production, and the role of various tools and frameworks in achieving this.
Unlock the power of AI to supercharge your product design process! This demo guides you through capturing raw ideas via voice recordings, organizing them into agile user stories with Otter and ChatGPT, and rapidly turning those insights into working app prototypes using Figma Make. You’ll learn to mine your own thoughts for powerful features and pain points, map these to real user needs, and supercharge your workflow with cutting-edge tools. By the end, you’ll be ready to turn any burst of inspiration into design-ready prototypes and actionable development steps.
Unlock the power of AI to supercharge your product design process! This demo guides you through capturing raw ideas via voice recordings, organizing them into agile user stories with Otter and ChatGPT, and rapidly turning those insights into working app prototypes using Figma Make. You’ll learn to mine your own thoughts for powerful features and pain points, map these to real user needs, and supercharge your workflow with cutting-edge tools. By the end, you’ll be ready to turn any burst of inspiration into design-ready prototypes and actionable development steps.
In this video, you'll learn how to transform your brainstorming sessions and unstructured ideas into actionable agile user stories using AI tools and Otter transcription. By following the process demonstrated, you'll discover how to mine your thoughts for key features and pain points, then organize them into structured requirements. Viewers will see how to use these user stories to generate rapid app prototypes with tools like Figma Make and refine them for a real-world project. By the end, you'll have the methods and confidence to turn your random ideas into clear, design-ready prototypes and workflows.
Following are the key things you will be able to do after you watch this demo:
Here is the template you can clone to define your app.
Click here to get the ultimate prompt cheat sheet of every prompt used end to end.
Click here to get the 10 step workflow summary guide and supplemental resources.
AI-Driven Prototype Development Process 0:09
Josh Lomelino explains the process of creating AI-driven prototypes using tools like Figma, Proto.io, and others.
The goal is to create a template that can be integrated into manual prototypes, eventually leading to a full app experience using tools like Lovable or Bubble.
Emphasis on the importance of a clear product definition and agile user stories for successful AI development.
Josh demonstrates how to train a chat on app features and user stories, using his app "Reclaim You" as an example.
Training ChatGPT for User Stories 4:30
Josh shows how to train ChatGPT on audio brainstorming sessions using Otter for transcription.
He explains the process of exporting SRT files from Otter and using them as inputs for ChatGPT.
The goal is to capture random thoughts and ideas, which AI can then organize into structured user stories.
Josh demonstrates how to ask ChatGPT to learn from the audio files and generate actionable insights for app features and user stories.
Data Mining and Feature Identification 10:13
Josh discusses the importance of data mining and research to identify core pain points and features for the app.
He shows how to ask ChatGPT to create lists of pain points, issues, and challenges from the data set.
The process involves categorizing pain points into broad buckets like health and wellness, planning and process, motivation and mindset, teaching and engagement.
Josh emphasizes the need for a clear understanding of pain points to develop effective product solutions.
Generating Agile User Stories 17:52
Josh explains how to use ChatGPT to create detailed agile user stories based on the identified pain points.
He demonstrates the process of training ChatGPT on the framework of pain to solution for creating user stories.
The goal is to generate a comprehensive list of user stories that can be used to guide the development of the app.
Josh shows how to create personas for different user groups and generate user stories for each persona.
Prototype Generation with Figma Make 25:43
Josh introduces Figma Make as a tool to generate prototype screens based on the agile user stories.
He explains the process of describing the app in Figma Make, including the app store description and features.
The tool generates HTML code for the prototype screens, which can then be manually refined.
Josh emphasizes the importance of using multiple tools and integrating their outputs to create a comprehensive prototype.
UI Framework and Stencils 35:30
Josh discusses the importance of selecting a UI framework for the final app experience.
He demonstrates how to use UI kits like Bootstrap UI and Material UI to create a consistent UI workflow.
The goal is to ensure that the prototype screens match the final app experience as closely as possible.
Josh shows how to use stencils to quickly create UI elements and save time in the development process.
Reviewing and Refining the Prototype 45:41
Josh explains the importance of reviewing and refining the prototype to ensure it meets the project requirements.
He demonstrates how to identify and fix broken links and other issues in the prototype.
The process involves iterating on the prototype, incorporating feedback, and refining the UI elements.
Josh emphasizes the need for a clear and accurate input to get the best output from AI tools.
Final Steps and Best Practices 46:18
Josh outlines the final steps in the AI-driven prototype development process.
He emphasizes the importance of saving chat history and project documentation for future reference.
The goal is to create a comprehensive and accurate prototype that can be used as a starting point for the final app development.
Josh encourages the use of multiple tools and integrating their outputs to create a robust and functional prototype.
This demo walks you through a dynamic, actionable framework for transforming user pain points into a Minimum Viable Product (MVP) using Agile user stories. You'll learn how to brainstorm and document personas, reframe core problems as needs, and translate those into actionable features with step-by-step guidance. The session demonstrates both hands-on manual methods and the use of powerful AI tools—like ChatGPT—to rapidly generate and refine user stories, making the product development process more efficient. By the end, you'll be equipped with practical strategies to build, organize, and launch your own MVP, leveraging proven workflows and modern AI support.
This demo walks you through a dynamic, actionable framework for transforming user pain points into a Minimum Viable Product (MVP) using Agile user stories. You'll learn how to brainstorm and document personas, reframe core problems as needs, and translate those into actionable features with step-by-step guidance. The session demonstrates both hands-on manual methods and the use of powerful AI tools—like ChatGPT—to rapidly generate and refine user stories, making the product development process more efficient. By the end, you'll be equipped with practical strategies to build, organize, and launch your own MVP, leveraging proven workflows and modern AI support.
In this video, viewers will learn a practical framework for developing a Minimum Viable Product by systematically transforming user pains into actionable features through needs analysis and Agile user stories. By following along, participants will discover how to brainstorm and document user personas, define core problems, and leverage AI tools to accelerate feature development. Completing the video equips viewers with the skills to map out, organize, and build a comprehensive feature backlog tailored to their users' needs. Ultimately, you’ll be able to apply this structured approach to launch your own MVP efficiently and effectively.
Here is the template you can clone to define your app.
Click here to get the 10 step workflow summary guide and supplemental resources.
Click here to get the ultimate prompt cheat sheet of every prompt used end to end. Below is the Prompt Atlas showing the 8 categories of prompts in the prompt cheat sheet guide.

Following are the key things you will be able to do after you watch this demo:
Introduction to MVP and Ecosystem Mapping (0:00:09)
Josh introduces the concept of the Minimum Viable Product (MVP), emphasizing the importance of defining pain points, needs, and features for different stakeholders and mapping the overall project ecosystem
Three-Part Agile User Story Process (0:00:09)
Josh outlines a sequential framework: converting pains into needs and then into features, which are written as Agile user stories, highlighting the value of progressing step-by-step rather than jumping directly to user stories.
Business Vision Planning and Personas (0:08:36)
The importance of developing user personas is discussed, showing how various archetypes (e.g., medical doctors, programmers, teachers) help clarify specific needs by associating real-life pains and developing empathy-driven solutions.
Pain, Need, and Feature Transformation (0:18:15)
Demonstrates, with the example of a medical doctor, how individual pains are rewritten into user needs and further translated into actionable features, using empathetic design as a core principle.
Manual and AI-Assisted User Story Generation (0:27:50)
Josh explains how manually defining the framework prepares for leveraging AI tools, which accelerate the process of drafting needs and Agile user stories for each persona.
Leveraging AI for Brainstorming and Data Mining (0:37:51)
Describes incorporating AI chatbots (e.g., ChatGPT, Gemini) to quickly generate, format, and expand user stories and pain points, as well as methods for data mining common user problems using AI.
Persona and Backlog Development Workflow (0:47:00)
Shares how organizing and summarizing ideas with AI and mind mapping tools streamlines persona development and backlog creation, allowing for comprehensive project planning and smoother team collaboration.
Iterative Process and Documentation Completion (0:55:56)
Details the ongoing, iterative approach of developing user personas and scenarios, curating and refining user stories, and documenting all features in structured formats for future product development steps.
There are no Main Site search results.