Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Cross Validation #19

Open
osivaz61 opened this issue May 11, 2022 · 2 comments
Open

Cross Validation #19

osivaz61 opened this issue May 11, 2022 · 2 comments

Comments

@osivaz61
Copy link

osivaz61 commented May 11, 2022

Hi,

I have two questions:

First : Why didn't you use K Fold Cross Validation?
Second : What is the reason use different learning rate for classifier? Is it for faster convergence?

I am trying to adapt CSRA to EfficientNetB3 on my multi-label dataset. Although I try various head and lambda numbers, I am getting worse results according to baseline model. What is your opinion? Is there also something different to try?

Also there is class imbalance in my dataset. Is there need to make data augmentation to prevent class imbalance? Is CSRA a method affected by data augmentation?

Thanks

@Kevinz-code
Copy link
Owner

Hi,
Thanks for your question.

For your first question, can refer to ASL, Q2L for a comprehensive pipeline of multi-label classification code (which contains standard validation code similar to ours).

The ten times larger learning rate of the classifier is used for faster convergency.

As far as we're concerned, there is a dropout layer between the avgpool the the final FC in EfficientnetB3, which might cause different struture settings when CSRA is applied. Try using resnet for a stable improvement.

For imblance, refer to recent multi-label long tail classification paper, where specific loss has been added to cure performance degradation.

Best,

@osivaz61
Copy link
Author

osivaz61 commented Jun 4, 2022

Hi,

I have an idea about CSRA performance. I want to know your idea also.

In voco2007 dataset there are 5011 images and 7306 objects. The multi-label objects ratio is (7306-5011)/5011 = % 45.79
In coco dataset there are 82081 images and 241035 objects. The multi-label objects ratio is (241035-82081)/82081 = % 193.64

The best performance gain was obtained for the coco dataset (for VIT-L16 (80.4 -> 86.5) %6.1).

For my dataset multi-label ratio is %9. I think the reason is why I can not get better results is lower multi-label ratio.
I am right?

Thanks
Orhan SIVAZ

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants