Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Where do the Sync-C and Sync-D benchmarks come from? #26

Open
mulanlew opened this issue Dec 4, 2024 · 3 comments
Open

Where do the Sync-C and Sync-D benchmarks come from? #26

mulanlew opened this issue Dec 4, 2024 · 3 comments

Comments

@mulanlew
Copy link

mulanlew commented Dec 4, 2024

Hi I would like to know where you got the Sync-C and Sync-D benchmarks from, thanks!

@xuyangcao
Copy link
Collaborator

Hi I would like to know where you got the Sync-C and Sync-D benchmarks from, thanks!

We follow pipelines in the Wav2lip: https://github.com/Rudrabha/Wav2Lip

@mulanlew
Copy link
Author

mulanlew commented Dec 5, 2024

Thanks!

@sampagon
Copy link

sampagon commented Jan 9, 2025

We follow pipelines in the Wav2lip: https://github.com/Rudrabha/Wav2Lip

I went ahead and setup the test bench from Wav2Lip but started with cherry-picking videos just to see what the sync-net scores the confidence and distance. It seems to be almost arbitrary... Using videos from your project page, Echomimics, HDTF, etc... all yield results with little to no pattern. For example, I tried 4-5 ground truth clips from HDTF that all yield a 0.x confidence scores which doesn't make sense considering they should have higher confidence than anything that is generated by an audio driven solution. What was your experience with this benchmark? @xuyangcao

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants