Skip to content

Latest commit

 

History

History
85 lines (59 loc) · 3.26 KB

README.md

File metadata and controls

85 lines (59 loc) · 3.26 KB

Pretraining of GeoMIM

The pretraining is implemented with the bevfusion framework. We use the LiDAR-only model in bevfusion (69.28 NDS) for pretraining.

Pretrained Models

We provide pretrained models that have been trained on the NuScenes dataset. These models can be used as a starting point for your own tasks or fine-tuning.

Config Epoch Download
Swin-Base 50 Model
Swin-Large 50 Model

Usage

Installation

The code is built with following libraries:

After installing these dependencies, please run this command to install the codebase:

python setup.py develop

Data Preparation

nuScenes

Please follow the instructions from here to download and preprocess the nuScenes dataset. After data preparation, you will be able to see the following directory structure (as is indicated in mmdetection3d):

mmdetection3d
├── mmdet3d
├── tools
├── configs
├── data
│   ├── nuscenes
│   │   ├── maps
│   │   ├── samples
│   │   ├── sweeps
│   │   ├── v1.0-test
|   |   ├── v1.0-trainval
│   │   ├── nuscenes_database
│   │   ├── nuscenes_infos_train.pkl
│   │   ├── nuscenes_infos_val.pkl
│   │   ├── nuscenes_infos_test.pkl
│   │   ├── nuscenes_dbinfos_train.pkl

LiDAR and MixMAE models setups

To set up the LiDAR model for pretraining, follow these steps:

Download the LiDAR model weights from bevfusion.

Download the Swin-Base/Swin-Large weights from MixMAE.

Pre-Training

We provide instructions to reproduce our results on nuScenes. You can use pytorch or slurm for distributed training.

For example, the Swin-Base model can be pretrained with:

sh run_pretrain.sh partition 8 config/pretrain_base_50ep.yaml runs/pretrain/pretrain_base_50ep

The large model can be pretrained as the same.

Acknowledgements

The pretraining code is based on bevfusion.