Learn how to scale interactive, personalized video with HeyGen enterprise tools, API workflows, and live avatars to boost training, sales, and support.
This session from HeyGen Bootcamp pulled together three big pieces of the “next-gen” stack:
- Personalization and interactivity in AI Studio (Karina Jhangiani, Head of Enterprise Product)
- Personalized video at scale via API and integrations (Kevin Raheja, Head of Partnerships))
- LiveAvatar: real-time, conversational AI humans (Sunny Liao, Head of Business for LiveAvatar)
Here’s a walk-through of what they shared and how you can actually use it.
1. Personalize your content without being a video pro
Karina kicked things off with a simple premise: HeyGen doesn’t expect you to be a video creator. The product is designed so personalization is basically “click a button, type a few words, done.”
Start with Video Agent for one-click drafts
Video Agent is HeyGen’s autopilot for video creation. You:
- Give it a prompt (and optionally slides/assets)
- It writes a script
- Picks images and B-roll
- Chooses an avatar and voice
- Adds transitions and motion
You get a fully-formed video you can personalize and refine without touching a timeline.
Perfect for:
- Social videos
- Announcements and updates
- “First draft” marketing and training content
Fine-tune everything in AI Studio
If you want more control, you move into AI Studio, where editing a video feels more like editing a document:
- Script panel: Change wording, tone, pacing
- Avatar and voice controls: Adjust style, emotion, language
- Gestures and delivery: Use tools like voice director and pausing to shape how lines land
Because it’s all script-driven, anyone on your team (not just editors) can personalize videos without professional production skills.
Screen capture + avatars for internal training
For internal enablement and technical content, you can layer:
- Screen recordings
- AI narration
- Your avatar or a brand avatar
HeyGen then helps you:
- Auto-generate narration
- Add captions and chapters
- Turn raw product walkthroughs into polished training modules
That means your training stays accurate as products change. You update the script or recording, not reshoot a whole talking-head video.
Localize once, reach everywhere
Personalization isn’t just “Hi {{first_name}}.” It’s also language.
HeyGen’s video translation stack lets you:
- Translate videos into 175+ languages and dialects
- Preserve the original speaker’s voice and style
- Maintain lip sync with realistic VT (video translation)
Viewers can even toggle languages in a multilingual player, choosing how they want to watch. No duplicate uploads, no extra editing for you.
2. Make videos interactive: Quizzes, branches, and CTAs
Personalization isn’t only what the video says; it’s how viewers engage with it.
Karina also showed how HeyGen now bakes interactivity directly into AI Studio.
Quizzes: Turn viewers into participants
You can drop quizzes right into your video:
- Reinforce key concepts
- Check comprehension
- Keep people engaged instead of passively watching
This is especially powerful for:
- L&D and training programs
- Customer education
- Internal onboarding and compliance
Branching: Build “choose your own path” experiences
Brand-new and very fun: branching logic.
On the timeline you can:
- Add a question or choice
- Define where each branch goes (Scene A vs Scene B, etc.)
- Let viewers “choose their path” inside the same video
Use cases:
- Scenario-based training (“What would you do next?”)
- Roleplay for sales and support
- Guided product tours where viewers pick their use case
Instead of a flat linear video, you get an adaptive experience that responds to the viewer’s decisions.
Embedded links: Drive action without losing attention
You can also add links and CTAs directly in the player:
- Send learners to documents or LMS modules
- Send prospects to product pages or trial signup
- Link internal viewers to SOPs, Notion docs, or knowledge base articles
All without making them leave the video tab or hunt around for “that link from the email.”
Purpose-built L&D solution
All of this comes together in HeyGen’s L&D solution, which:
- Verticalizes the platform for training teams
- Adds LMS-friendly outputs:
- SCORM export
- LMS-ready HTML embed codes
- Makes it easier for L&D leaders to create, localize, manage, and measure training at scale
If you’re running education programs, this is where personalization + interactivity really stack up.
3. Personalized videos at scale via API and integrations
Next up, Kevin walked through how teams are automating personalized video using HeyGen’s API.
The big use cases that are exploding
He highlighted four categories where personalization is taking off:
- Sales
- Cold outreach
- Warm follow-ups
- Marketing
- Lifecycle campaigns
- Event confirmations and thank-yous
- Customer success
- Onboarding walkthroughs
- Feature education
- Internal comms
- Employee recognition
- New hire intros
Anywhere you’d send a personalized email, you can now send a personalized video.
How template-based personalization works
The workflow is straightforward:
- Create a template in HeyGen
- Pick an avatar (public or your digital twin)
- Write a base script
- Add variables
- e.g. {{first_name}}, {{company}}, {{plan_type}}
- These tokens get swapped out per contact
- Connect your data
- CSV upload
- Google Sheets
- CRM (via API or tools like Zapier, Make, n8n, etc.)
- Generate at scale
- HeyGen autogenerates a unique video + share URL for every row in your data
Result: “Hi Kevin…” “Hi Priya…” “Hi Luis…”—hundreds or thousands of truly personalized videos created from one template.
Connect your stack: CRMs, design tools, automations
HeyGen plugs into a wide set of tools so you’re not rebuilding your workflows from scratch:
- Automations
- Zapier, Make, n8n, Pabbly, viaSocket, and more
- Design and editing
- Canva
- Adobe Express
- Hosting and sharing
- Vimeo sync
- Logic and content
- A ChatGPT integration for script generation and refinement
You can think of HeyGen as the video generation engine behind your existing stack: CRM, ESP, CDP, automations, and design tools.
4. LiveAvatar: Real-time, conversational AI humans
Finally, Sunny showed where things are headed when you take hyper-real avatars and hook them up to real-time conversations.
What is LiveAvatar?
LiveAvatar (the evolution of “Interactive Avatar”) lets you:
- Spin up an avatar that can talk live with users
- Handle back-and-forth conversations in real time
- Plug into your own knowledge base or systems via API
It runs at low latency and supports enterprise-level SLAs so teams can safely build it into customer-facing or mission-critical workflows.
You can use:
- Public avatars from the HeyGen library
- Or your own custom avatar, trained from ~2 minutes of footage following HeyGen’s guidelines
Where people are already using it
Sunny shared four categories where customers are seeing real traction:
1. E-learning and assessment
Use LiveAvatar as an:
- AI tutor
- AI coach
- Practice partner
Available 24/7, speaking multiple languages, testing understanding, and simulating real conversations.
2. AI recruitment
Here, the avatar becomes the first-round interviewer:
- Runs structured interview flows
- Asks follow-up questions
- Captures responses consistently
Interestingly, many candidates actually feel more comfortable with an AI interviewer. Less judged, less biased, and easier for non-native speakers.
3. Sales training and roleplay
LiveAvatar can be the:
- Tough prospect
- Objection-heavy buyer
- Scenario partner
Reps can practice:
- Negotiations
- Objection handling
- Product storytelling
Anytime, without needing another human schedule to line up.
4. L&D (internal)
For fast-moving products or industries, LiveAvatar can:
- Sit on top of an always-updated knowledge base
- Answer employee questions in real time
- Simulate customer scenarios or internal workflows
Instead of static training that ages quickly, you get a living, conversational trainer.
Getting started with LiveAvatar
To try it:
- Go to liveavatar.com (separate site from HeyGen)
- Log in (you can reuse your HeyGen account)
- Talk to a public avatar to feel the experience
- When you’re ready:
- Record ~2 minutes of footage for your own custom avatar
- Let HeyGen train it (typically under 24 hours)
- Integrate via the API in full mode (HeyGen manages ASR/TTS/LLM) or custom mode (you bring your own stack, HeyGen handles the avatar layer)
Current limitation: LiveAvatar is cloud-only (no on-prem yet), but you can embed it into kiosks, websites, apps, and more as long as they can call the cloud APIs.
Where to go from here
If you’re already using HeyGen for “classic” avatar videos, the next steps look something like:
- Personalize more deeply
- Try Video Agent → refine in AI Studio
- Translate an existing video into a new language
- Add a quiz or a simple branching choice
- Connect your stack
- Build one personalized video template
- Connect it to a CSV or CRM list
- Trigger a small batch of personalized videos as a pilot
- Experiment with live experiences
- Talk to a public LiveAvatar on liveavatar.com
- Map one use case (e.g., interview simulation, product Q&A, internal training)
- Prototype with API docs and a small internal audience
The pattern across all three layers (Studio, API, Live) is the same:
- Start small
- Treat it like an experiment
- Measure time saved and impact
- Scale what works
Video production doesn’t have to be a bottleneck anymore. With AI Studio, personalized video via API, and LiveAvatar, it becomes a system you can actually manage and scale.








