Tess AI's Avatar tool lets you create videos with a digital presenter that “speaks” a script you provide, with lip sync. This makes it possible to create professional-looking content without a camera, studio, in-person recording, or traditional editing — ideal for scale and standardization.
In this process, the AI animates the avatar, generates or uses the audio, syncs the lip movements with the speech, and renders the final video for download.
Models available in Tess
To activate it, just find the Avatar option in the tools button; there you’ll find models like:
HeyGen
Omni Human
Wan
Each option tends to have different settings and performance (avatar style, realism, expressiveness, lip sync quality, language/voice options, etc.).
When to use it (ideal cases)
onboarding modules
product and process trainings
internal policies and standardized announcements
announcement videos (short)
feature presentations
welcome messages and “product tour” with consistent identity
team/project updates
leadership announcements (with standardization and speed)
short educational videos (Reels/TikTok)
weekly series with the same visual identity
Important tip: if you want, you can combine Avatar + Narration (Speech) for full control, especially if you want maximum voice consistency (tone, rhythm, timbre).
How to write scripts that sound natural in avatar
Write “to be spoken”, not like an article text
Use short and direct sentences
Avoid long paragraphs
Add natural pauses using punctuation
For acronyms, preferably write them out in full the first time (e.g.: “Customer Success” before “CS”)
If there are technical terms, include a context sentence to reduce “robotic reading”
Credit usage and generation time
Avatar videos tend to use more credits than simple text and narration, because they involve rendering. They can also take a bit longer to be ready, especially for long videos or higher quality settings.
For anything you need, you can reach out to our support team at: support@tessai.io.