-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
My tests show the character has a mask on his face #16
Comments
This is probably caused from the color distribution in your video. The face mask is defined correctly but the merged lip-sync output is simply showing no details for the lips since the model wasn't trained on similar skin colors, like in this video. Would be better if you try changing the skin color in this video. |
Is there any other good driving method |
if you wanted to animate a single image as it's shown above then D-ID would be your best bet, you can get free credits on every email you sign up with, pretty sure they allow facial hair and skin discoloration, other than that there is wav2lip + upscaling but results would be poor, the rest of the tools out there are trained on real people without facial hair and normal skin so you would get similar results as the above if an error didn't appear, possible sadtalker would work too & that's opensource, and if you needed to animate it you could look into putting the results onto the original video with masking in a video editor although pretty sure sadtalker can copy the head movements |
I also downloaded that repo it's shame they deleted their code and didn't release their training scripts no point looking more into it if they are not continuing (i'm assuming they plan to keep it closed now) and yeah d-id is based off of sadtalker they just did some magic to it |
The other magic is https://github.com/thygate/stable-diffusion-webui-depthmap-script You can use depthmaps to create videos from images. (Or the new stable video diffusion SVD) Depthmaps + sadtalker and you get an awesome talking faces with animated videos. Actually you can do everything with automatic1111 and centralized in just one app. You are welcome. David Martin Rius |
By the way, I think the code was never released. There was just a readme, but finally deleted the repo |
yeah code was released (not training code) I downloaded the pretrained model + code then they disappeared 2 days later. unfortunately the model is person specific/ not generalized. |
Thank you very much for your reply, I found this project, the author said the effect is too good, did not publish the inference model And the latest project, GAIA: Zero-shot Talking Avatar Generation Unfortunately, the project homepage is also down |
|
Hi @Inferencer you said that you downloaded the pretrained model and code of stylelipsync.. can you share it privately? we can figure out how we can train new models etc.. I run a service related to lipsync and we have a team that will help us on this. |
https://drive.google.com/drive/folders/1W9RAyqu2hwrieaWtGG19GmSjkkhreyIA?usp=sharing |
I recently found the new homepage for gaia https://gaiavatar.github.io/gaia/ |
Did you guys manage to reproduce the StyleLipSync training algorithm? |
Nope, but this is coming next month, could drive it with a 3dmm or something so it is ctrl with audio rather than driving vid |
Do you have any recommendations for open-source modelos for lip sync that can be used commercially? All the ones that I'm finding have (1) no code or (2) non-commercial license |
result:
src:
What causes this and how can you get rid of it
The text was updated successfully, but these errors were encountered: