-
Notifications
You must be signed in to change notification settings - Fork 783
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Low mIoU on Cityscapes #35
Comments
What is the backbone you used in your model? |
The backbone is Resnet, as I have seen in another issue that Xception may have some problems. Am I right? |
Yes, previous code miss bn after depth wise cnn. I have fixed it. But I do not have pretrained aligned xception model right now. So I dont recommend you use this backbone. |
Hi, it is an error in |
Are you sure about that? In the cityscapes dataset here you are referring to the method I linked before, with 0 filling, so you are actually telling your network that all the padding you have belongs to a class. This is on |
Hi, I think you are right. |
I will test the model trained with the updated code and report the performances to you. Right now, I am training with really resized images (512x256), batch size 16 and lr=0.007. |
Update on the results: very low mIoU (around 0.6). I think there is some additional problem with the code, because I trained also on another dataset (bdd100k) and the results are equally bad. Or maybe my training parameters are wrong, but they are similar to yours... |
Same here... low mIoU (around 0.6), base_size = 1024, crop=600 |
I tried to train ResNet based model on cityscapes dataset. My training setting is batch_size 16 (8 TITAN XP GPUs), base_size and crop_size 796, lr 0.01, epoch 200. Currently on epochs 50, and best mIoU is 0.65. So I think we need to tune training parameters in order to reach higher mIoU. |
Maybe you could explore pytorch-encoding or drn repos to try their training setting and data augmentation methods on Cityscapes dataset. I believe their code could help you. |
Same Low IOU on cityscapes, 200 epoch, 8 GPUs, BS = 16, lr 0.01, base_size 1024, crop_size 769. The Final result is 67.98% |
But the use of whether general or weighted mIou should not cause such a large performance gap. |
@jfzhang95 @fabvio How to do a single inference?, I have train in cityscapes I got only 62% mIoU. After I am doing testing I got this result. Why the result seems very bad? I have doing the normalization first before feeding to the network like substracted mean and unit variance. |
Hi, thanks for the great repo. I try to train the cityscapes dataset with the suggested parameters. However, the performance of network is weird. The epoch loss decrease all the way. However, it is weird to note that the miou first rise then descend and the highest miou is 57.5%. Could you please try to give a suggestion on this? Thanks a lot. |
This repo use nn.Dropout instead of nn.Dropout2d. Maybe this is one reason. |
Hello ,have u solve this problem yet? can you share your results? I also only got 68% mIOU,and dont know how to improve . Thank you very much! @GeoffreyChen777 |
I found the problem for my case probably caused by the function of SynchronizedBN. When I turn of the SynBN and use the nn.BatchNorm2d, I have the highest performance 72% in val dataset without coco pretrained. However, it is weird to note that the miou descends in the middle of training. By the way, my setting is bs=4, base_size=796, crop_size=796, backbone=resnet lr=0.01 1GPU(v100). The current miou is still far away from the state-of-the-art performance, maybe there are other problems. |
Trained the model on Cityscapes for 100 epochs, evaluated using official Cityscapes script. Obtained mIoU of 0.694. Evaluation using the repo I used the whole image i.e, 1024*2048 during evaluation. The mIoU improved slightly to 0.6943 Evaluation using official Cityscapes scripts |
the metrics Acc np.diag(self.confusion_matrix) / self.confusion_matrix.sum(axis=1) may be wrong,please check it. |
Hi, thanks for sharing your code. I trained the model on the cityscapes dataset (btw, you missed a
self
here) without code edits and I can get only 68% mIoU. Do you have any pretrained models, or can you describe your training strategy? I trained on a single Tesla v100 GPU, with lr = 0.007 and batch size 8.Thanks in advance
The text was updated successfully, but these errors were encountered: