-
Notifications
You must be signed in to change notification settings - Fork 11
Description
Dear Author,
According to the UMMAFormer paper, extracting features for this dataset requires multiple complex preprocessing steps, including:
Frame extraction and optical flow computation using MMAction (v0.24.1)
Audio separation and feature extraction using BYOL-A
Integration of pre-trained TSN models for RGB and flow features
Given the complexity of this pipeline and the substantial computational resources required, I was wondering if you would be willing to share the pre-extracted features (both visual and audio) that you used in your experiments with the AV-Deepfake1M dataset.
This would greatly facilitate our reproduction efforts and help ensure the accuracy of our implementation. We would be happy to:
Sign any necessary data usage agreements
Provide proper attribution in our work
Share our reproduction results with you for verification
If the features cannot be shared publicly, we would be grateful for any guidance on the specific parameters and configurations you used in your feature extraction pipeline.
Thank you for considering this request. I greatly admire your work and look forward to the possibility of learning from your implementation.