Adding Mouth Movement to your Pika Labs AI Videos

December 13, 2023

Pika Labs is one the most popular tools next to RunwayML for generating AI videos from text. Today, we'll show you how you can add mouth movement (a.k.a. lip sync) your Pika Labs videos to make your content even more dynamic.

 

Step 1. Generating a Video from Pika Labs

Create an account on Pika Labs and generate a video by typing /create or you can also animate an existing image via the /animate command. For our example, we use /create with an image parameter to generate the video. Below is a sample prompt made by another user.

screen

The video should looks something like you see below.

 

Step 2. Upload the Video on Polymorf

Login to https://polymorf.me. You should see a library of your current videos, or if none, you should see a "Create Video" button. Click that, and it'll take you to the editor

Library

Once you're on the editor, click on the videos tab, and click on the "Click to Upload" button on videos library on the left drawer to upload the video from pika labs.

Editor

 

Step 3. Enter a script/sentence

You can enter a script to lipsync the video. Most pika labs videos are only 3 seconds in duration currently, so make sure the duration is not too long. We can verify the video duration by clicking the "Render" button and checking the duration field. It shows that the text "Is a genie going to come out of here?" as having 2 seconds duration.

Script

 

Step 4. Render the video

Once that's done, you can click "Render", and with enough credits, generate a video with mouth movement. See the results below.

 

Conclusion

As you can see, the process is pretty simple and straightforward, and you can do this either on a website or mobile device. If you want to add mouth movement and make your pika labs videos talk, try using Polymorf AI for free, and you can check it out at https://polymorf.me/lipsync-video .