-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Implement more synching options #31
Comments
Hi there, I strongly encourage you to check Mario Guggenberger's AudioAlign (which relays on Aurio fingerprint library), Ben Miller's Audalign and audio-offset-finder. I also suggest you try to avoid the "reinventing the wheel" issue by focusing your efforts on aspects (e.g. phase/polarity adjustments after synching) that other projects have not yet considered/implemented. In other hand, for visual synchronization approach, @cgtuebingen's Learning Robust Video Synchronization without Annotations seems the most advanced/promising one instead. Hope that inspires. |
Hi @MarcoRavich, Thanks for the thoughtful comment, and including us in your resource.
As far as a phase polarity post-processing step goes, it's not top on our list, as we're focused on synchronizing videos for use in motion capture, where downstream audio isn't very important. I would gladly review a contribution of it if it's something you're motivated to implement though! Also, for the Learning Robust Video Synchronization without Annotations, it seems to be a different kind of video synchronization than we're interested in. They perform spatial alignment (align images in space), while we're interested in temporal alignment (align frames by when they were shot) |
The current synching methods (audio cross correlation and brightness contrast thresholding) both work in the majority of cases. But there are no options for a different method if a user's videos do not synchronize properly (although the brightness contrast method can be run with different thresholds to change the results). Adding methods will help ensure a method works for each set of videos, and different synching methods will be robust to different types of noise. I would like to have at least 2 audio and two brightness methods.
As described in this academic poster by Mark Koh, you can use the Python package Dejavu to match audio files and use the output to synchronize videos.
Since we already take the brightness over time of videos in the brightness contrast thresholding method, we can simply use the cross correlation methods from the audio cross correlation method and apply it to the brightness time series.
The text was updated successfully, but these errors were encountered: