Real-Time AI Conversations & Facial Animation for MetaHumans in Unreal Engine Using Convai

By
Convai Team
February 25, 2026

Imagine walking up to an NPC in your game, or a virtual brand avatar in your training simulation and having a natural, unscripted conversation about anything, stargazing, your company’s information and more, complete with perfectly synchronized lip movements and nuanced facial expressions.

With Convai’s updated Unreal Engine plugin, creating highly expressive, low-latency Conversational AI characters is easier than ever.

Watch the full tutorial below:

Why Should You Use Convai for Real-Time AI Conversations and Facial Animation?

Traditional dialogue tree

Interactive storytelling and virtual simulations have been bottlenecked by static dialogue trees and pre-baked animations. If a player or trainee asked a question outside the script, the immersion broke.

Today, AI-powered characters are changing the landscape of game development, virtual training, and enterprise simulations. By integrating Large Language Models (LLMs) with high-fidelity 3D avatars, developers can offer users unprecedented agency. Players can speak into their microphones naturally, and the AI agent will understand context, recall backstory, and reply dynamically. This doesn't just save hundreds of hours in manual animation and voice acting—it creates a deeply personalized, immersive experience that was previously impossible.

How Does Convai Enhance Real-Time Lipsync in Unreal Engine?

Convai's plugin banner

Convai’s latest Unreal Engine integration is a massive leap forward for Embodied AI, focusing heavily on speed, realism, and ease of use. Here is what the latest upgrade brings to your Unreal Engine 5 projects:

  • NeuroSync Animation System: Convai’s in-house neural AI model processes live audio streams instantly, eliminating the need for offline processing or pre-baked animations.
  • Realtime Lipsync & Facial Animation: The system natively drives MetaHuman’s 250+ facial blend shapes. The result? Nuanced, dynamic expressions that adapt to the emotional tone of every single word spoken.
  • Ultra-Low Latency: Optimized for real-time interactions, ensuring the pause between a user speaking and the AI Avatar responding feels as natural as a human conversation.
  • Handsfree Voice AI: New blueprint components allow for seamless push-to-talk or completely hands-free continuous listening.

The Tech Behind the Realism: NeuroSync and MetaHumans

Convai's lipsync banner

When your character responds to a user's voice, NeuroSync goes to work. Instead of relying on rigid, pre-programmed facial movements, NeuroSync generates perfectly synchronized lip movements on the fly based on the generated text-to-speech audio.

Whether you are building interactive NPCs, virtual instructors for AI in XR, or enterprise sales roleplay simulations, this pipeline guarantees a deeply immersive user experience without bogging down your development timeline.

Read Also: Give Your AI Characters Eyes & Ears in Unreal Engine: Streaming Vision + Hands-Free AI Voice Interaction

What Are the Steps to Animate AI MetaHumans in Unreal Engine Using Convai?

  1. Enable the Convai Plugin: Install from the Fab Store, enable it via Edit > Plugins, and authenticate your account.
  2. Import MetaHuman via Quixel Bridge: Export the character directly into your project.
  3. Add the Chatbot Component: Connect your MetaHuman to Convai’s LLM by adding the BP_ConvaiChatbot component and inputting your Character ID.
  4. Add the FaceSync Component: Enable lipsync and assign Convai MetaHuman animation classes to the face and body.
  5. Set Up Player Speech Input: Add BP_ConvaiPlayer to your player blueprint and configure push-to-talk or handsfree voice AI.

Step-by-Step Guide: Animating AI MetaHumans in Unreal Engine

Demo snippet

Let's get into the editor. For this tutorial, we are starting with a newly created Unreal Engine project using the First Person template.

Step 1: Enable the Convai Plugin

Enabling Convai Plugin in Unreal Engine

If you haven't installed the Convai plugin yet, grab it from the Fab Store and check out our documentation for Unreal Engine.

  1. Go to Edit > Plugins.
  2. Search for Convai and check the box to enable it.
  3. Restart your project.
  4. Upon reopening, sign in to your Convai account to authenticate. You can verify your connection by clicking the Convai icon in your toolbar.

Step 2: Import Your MetaHuman via Quixel Bridge

Metahuman selection in Quixel Bridge
  1. Open Window > Quixel Bridge and navigate to the MetaHumans section.
  2. Select your desired MetaHuman and click Export to send it directly to your project.
  3. In your Content Browser, locate the MetaHumans folder and double-click the Blueprint (BP) file for your character.
  4. If prompted to enable missing plugins, click Enable and restart the engine.

Step 3: Add the "Brain" (Chatbot Component)

Convai Character Screen

Now, we need to connect our MetaHuman to Convai's LLM so it can think and speak.

  1. Open your MetaHuman Blueprint.
  2. Click Add Component and search for the BP_ConvaiChatbot component. This acts as the brain of your character.
  3. In the Details panel of the Chatbot Component, find the field for the Character ID.
  4. Head over to the Convai Dashboard, create a new character, customize their backstory and voice, and copy the Character ID.
  5. Paste this ID back into the Unreal Engine Blueprint.

Step 4: Add the "Animation" (FaceSync Component)

Convai Character Blueprint Navigation

To ensure the AI's voice perfectly matches the facial movements, we need to interpret the audio data into animation.

  1. Add the BP_ConvaiFaceSync component to your MetaHuman Blueprint.
  2. In the Details panel, ensure Lip Sync is enabled.
  3. Set the Lip-Sync Mode to MetaHuman Blend Shapes.
  4. Select the Body component, go to Animation Class, and select Convai MetaHuman Body Anim.
  5. Select the Face component, go to Animation Class, and choose Convai MetaHuman Face Anim.

Step 5: Set Up Player Speech Input

Enabling Push-To-Talk

To enable voice interactions:

  1. Press Shift + F1 to regain mouse control and click Detach in the toolbar to navigate the editor while in Play mode.
  2. Select your Player Character in the scene and click Edit Blueprint.
  3. Add the BP_ConvaiPlayer component to your player blueprint.
  4. In the details, you can choose to enable Push-to-Talk or disable it to allow for completely Handsfree Voice AI interactions.

Hit Play! Walk up to your MetaHuman and start talking.

Example Scenarios to Build Today

Sample Character from Convai

Not sure what to build first? Here are two powerful, ready-to-use, Convai characters to integrate with your MetaHumans in your next project:

1. The Medical Sales Trainer

  • Character Name: Dr. Clara Reynolds
  • Description: You are Dr. Clara Reynolds, a professional pharmaceutical sales trainer at MediTech Solutions, a company specializing in cardiovascular medications… (Read the full description here
  • Ready-to-use Character ID: 7a933556-c8c5-11ef-ad5e-42010a7be016
  • The Setup: Build a clinic-themed room in Unreal Engine
  • The Interaction: Trainees can use hands-free voice AI to interact with Dr. Clara, asking critical questions about the craft of pharmaceutical sales. She will respond in real-time, offering verbal corrections with realistic, empathetic facial expressions.

2. The Enterprise Mock Interviewer

  • Character Name: Alex
  • Description: A highly experienced and friendly Software Development Engineer (SDE) interviewer conducting realistic mock interviews… (Read the full description here)  
  • Ready-to-use Character ID: 45998160-c8d0-11ef-842c-42010a7be016
  • The Setup: Dress a MetaHuman in professional attire in an office type environment
  • The Interaction: Customers visiting your 3D office setup can ask Alex to start the interview with something like "Hello Sir, I am ready for the interview. Can we begin?” Alex will process the intent and ask questions, and answer empathetically with flawless Realtime Lipsync.

Also Watch: Reallusion Avatars with Conversational AI, Real-time Lipsync & Facial Animation | Convai UE Tutorial 

How Can Convai Improve Real-Time Voice AI NPCs in Unity and Other Engines?

While this tutorial focuses on Unreal Engine, Convai’s NPC AI Engine supports real-time voice AI and dynamic dialogue systems in multiple platforms including Unity. Developers can use Convai’s SDK to integrate NPC speech that reacts to game states, enabling seamless player interactions without scripted dialogue limitations.

Frequently Asked Questions (FAQs)

Q: Do I need to create my own facial animations for the AI to speak? 

A: No! Convai's NeuroSync system processes the generated audio stream in real-time and automatically drives over 250 MetaHuman facial blend shapes. You do not need to manually animate the lip-sync or facial expressions.

Q: Can I use this setup for VR and XR applications? 

A: Yes. The Convai Unreal Engine plugin is highly optimized for AI in XR. Because it processes audio and animation streams with very low latency, it is perfect for maintaining immersion in Virtual Reality.

Q: Does Convai work with avatars other than MetaHumans? 

A: Absolutely. While this tutorial focuses specifically on MetaHumans, Convai is avatar agnostic and supports various avatar systems, including Reallusion, and custom 3D characters.

Q: Do I need heavy coding experience to integrate this? 

A: No. As shown in the tutorial, Convai utilizes Unreal Engine Blueprints. By simply adding the Chatbot, FaceSync, and Player components to your existing Blueprints, you can achieve conversational AI without writing C++ code.

Q: Is the voice interaction actually real-time? 

A: Yes. By combining fast Speech-to-Text, an optimized LLM pipeline, and rapid Text-to-Speech generation, the latency is kept incredibly low to simulate a natural, flowing conversation.

Q: What is Convai’s NeuroSync technology? A: NeuroSync is Convai’s neural AI model that processes live audio streams instantly to drive real-time lipsync and facial animation on MetaHumans. It eliminates the need for pre-baked animations, enabling dynamic and expressive AI characters.

Q: Can Convai’s system be used in VR and XR environments? A: Yes. Convai’s Unreal Engine plugin is optimized for low latency and high fidelity, making it ideal for immersive VR and XR applications that require natural voice AI and facial animation synchronization.

Q: Do I need to write code to integrate Convai with Unreal Engine? A: No. Convai leverages Unreal Engine Blueprints, allowing developers to add Chatbot, FaceSync, and Player components without heavy coding. This makes conversational AI accessible even to those with minimal programming experience.

Q: Is Convai compatible with avatars other than MetaHumans? A: Absolutely. While this tutorial focuses on MetaHumans, Convai supports multiple avatar systems, including Reallusion and custom 3D characters, offering flexibility across different projects.

Q: How real-time is the voice interaction using Convai? A: Convai combines fast Speech-to-Text, optimized large language models, and rapid Text-to-Speech generation to ensure ultra-low latency, enabling natural, flowing conversations with AI-driven characters.

Join the Convai Community

Convai Forum

Ready to start building your own intelligent, fully interactive AI agents?

  • Try it out for free: Sign up at Convai.com and explore our extensive documentation.
  • Need technical support? Visit the Convai Developer Forum to connect with our community, report bugs, and share your incredible AI creations.

Don't forget to subscribe to our YouTube channel for more deep dives into Convai UE Integrations!