TL;DR: Introducing Loci-Looped, an extension to Loci, with a parameterized percept gate that learns to slot-wise fuse imaginations with sensations. As a result, Loci learns to track objects through occlusions and to imagine their trajectory. See the project page: Page. See our paper for more details: arxiv
LociLooped_adept_control.mp4
A suitable conda environment named loci-l
can be created
and activated with:
conda env create -f environment.yaml
conda activate loci-l
Preprocessed datasets together with pretrained models and results can be found here
Download the pretrained folder and place it in the /out folder. Download the data folder and place it in the /data folder.
The model's performances on the testsets are stored as csv files in the respective results folders. To re-generate these results run:
sh run_evaluation.py
To test a single Loci model on the testset run:
python -m scripts.exec.eval -cfg [path-to-config-file] -load [path-to-model-file] -n replica
To analyse the generated results, run the evaluation notebooks in the /evaluation folder. You may need to modify the root_path variable in the notebooks pointing to the respective results folder.
Loci-Looped can be trained using with this command:
python -m scripts.exec.train -cfg [path-to-config-file] -n [run_description]
The original config files used for training can be found in the cfg folder. To continue training from an existing model use this command:
python -m scripts.exec.train -cfg [path-to-config-file] -load [path-to-model-file] -n [run_description]
MOT metrics are computed with py-motmetrics. The visual evaluation metrics script is borrowed from the Slotformer repository.