Our product enables to animate audio on the fly (1) in a browser (2) with your personal 3D model (3) called flexatar.
This means
- Animating in real-time enables you to implant flexatars to any service, from WebRTC online meetings to chatbots with AI
- The capability to run on the device makes scaling trivial and eliminates the need for GPU-powered backend
- Your customers can easily make 3D models by their own, and it takes about 15 seconds
Last but not least, we are provideing effects like mixing different models and introducing emotions
Please take a look at our demo to get an idea how it looks. It makes sense to start with Web Radio App.
Obtain FLEXATAR_API_SECRET from Flexatar Web Page.
export FLEXATAR_API_SECRET=your_api_key
cd src/server
npm install
npm run build
npm run start
Visit the demo server in your browser at: localhost:8081/main
To make integration seamless, we provide JavaScript libraries, each utilizing an underlying i-frame but offering additional functionality and UI for different use cases.
The Virtual Camera Library v-cam takes an audio source as input and outputs a synchronized (audio+video) stream. It includes an iframe-based UI for controlling the video output, allowing users to create Flexatars, switch between different Flexatars and adjust emotions. This solution can be used in various scenarios, such as:
-
WebRTC Virtual Camera – Acts as a virtual webcam, providing a new capture device alongside physical cameras.
-
AI Virtual Assistant – Can be used to create interactive AI-driven avatars that respond naturally. In this case, the UI for creating and deleting Flexatars can be hidden for a more streamlined experience.
The Video Generator Library v-gen provides a UI and functionality for generating videos with talking heads synchronized to audio. It allows users to create and customize video content easily within your webpage.
Our solutions are designed for easy embedding and customization, enabling you to integrate talking avatars into your applications effortlessly.
All out I-frames from above are interworing with our backend. Our customers are getting Secrets from us, that enables them to request Tokens to serve their end-users.
We are also providing test access to help our customers tune interfacing with both our backend and end-users.
The node.js example of customer-side backend can be found here
The terms of use are presented at our website
Flexatar technology comprises the following:
- Our unique proprietary algorithm to create truely 3D flexatars from photo(s) or videos (for enchanced version). Computing is performed on our backend, and it takes about 15 seconds.
- In-browser real-time animation of 3D flexatars with effects. The tiny delay is introduced to compute phoneme animation, that is combined with animtion patterns making flexatars look alive. Flexatar applied for animating can be either original one or a morph of flexatars with emotions and effects which are controlled by a user. The animaton delay is a compromise between an adequate voice-driven animation and the processing duration. To compensate such a delay and ensure lipsync, the delayed audio track is provided alongside with video.
Flexatars are created and stored in our AWS-hosted backend, being supplied to a browser directly. Our customer is expected to request Tokens per user session from backend by Secret. These Tokens are used to account end-users activities.