Skip to content

Request for Pre-extracted Features of AV-Deepfake1M for UMMAFormer Reproduction #25

@123308

Description

@123308

Dear Author,
According to the UMMAFormer paper, extracting features for this dataset requires multiple complex preprocessing steps, including:

Frame extraction and optical flow computation using MMAction (v0.24.1)

Audio separation and feature extraction using BYOL-A

Integration of pre-trained TSN models for RGB and flow features

Given the complexity of this pipeline and the substantial computational resources required, I was wondering if you would be willing to share the pre-extracted features (both visual and audio) that you used in your experiments with the AV-Deepfake1M dataset.

This would greatly facilitate our reproduction efforts and help ensure the accuracy of our implementation. We would be happy to:

Sign any necessary data usage agreements

Provide proper attribution in our work

Share our reproduction results with you for verification

If the features cannot be shared publicly, we would be grateful for any guidance on the specific parameters and configurations you used in your feature extraction pipeline.

Thank you for considering this request. I greatly admire your work and look forward to the possibility of learning from your implementation.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions