We using the pre-extracted features coming from this awesome paper R$^2$-Tuning, can be downloaded from HuggingFace Hub directly. And We express our sincere gratitude for their contribution to the community.
Please follow our baseline to prepare the dataset and place the corresponding files in the correct directory. And change the config file to the correct path.
Here are the origin video datasets download links:
# Single GPU
python tools/launch.py <path-to-config>
# Multiple GPUs on a single node (elastic)
torchrun --nproc_per_node=<num-gpus> tools/launch.py <path-to-config>
Arguments of tools/launch.py
config
The config file to use--checkpoint
The checkpoint file to load from--resume
The checkpoint file to resume from--work_dir
Working directory--eval
Evaluation only--dump
Dump inference outputs--seed
The random seed to use--amp
Whether to use automatic mixed precision training--debug
Debug mode (detectnan
during training)--launcher
The job launcher to use
python tools/launch.py <path-to-config> --checkpoint <path-to-checkpoint> --eval
If problems occur when reproducing the results, please feel free to contact us at github or email.
Maybe you need to change the config
file to the correct path.
Some issues may be fixed by these issues in Baseline Repository
We would like to express our sincere gratitude to the following authors for their contributions to the community: