Instant Dialogues — Overview
Instant Dialogues gives the fastest possible way to generate AAA quality dialogues in Unreal Engine, simply from text. It uses ElevenLabs API for text2speech transformation, and MetaHuman Performance for speech 2 lipsync. All this happens under the hood.
A game-changing moment was when ElevenLabs launched model "Eleven V3".
It creates a theatrical quality voice, which is then very nicely translated to lip-sync with MetaHuman Performance.
The design concept is to provice the simplest and fastest possible workflow, but to allow customization and fine tuning when and where needed.
This includes:
- Voice can be generated from elevenlabs by pressing one button, but you can upload your own voice.
- Facial Animation is generated with one click, but you can upload your own custom lipsync animation.
- You can specify body animation montage which will be played along with the voice.
- Default camera preset can be changed for specific dialogue part.
- Dialogue controller points the face towards the following actor by default, but this can be overriden; for example in situation when A is talking to B, but suddenly person C cuts in. In such graph, the default face of A would be directed towards C (because his line is next). We can override this by specifying B as listener.
- Player branching - choice options by default contain the full text of the following line, but if you specify choice text, it will be used instead. For example, if player is going to say "I can help you, but you should tell me exactly what happened", you can specify choice text "Ok, I will help you.
