You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It appears that, in all CNN models, the last dropout, i.e., embedding = F.dropout(x, p=0.5, training=self.training), is actually disconnected from the output linear layer, i.e., self.fc_audioset(x).
Indeed, the forward method of these models reads:
By reading the arXiv paper, it seems that the last dropout should have instead connected the 2048-embedding layer to the 527-output layer. Indeed, the paper reads:
"Dropout [38] is applied after each downsampling operation and fully connected layers to prevent systems from overfitting."
It appears that, in all CNN models, the last dropout, i.e.,
embedding = F.dropout(x, p=0.5, training=self.training)
, is actually disconnected from the output linear layer, i.e.,self.fc_audioset(x)
.Indeed, the
forward
method of these models reads:By reading the arXiv paper, it seems that the last dropout should have instead connected the 2048-embedding layer to the 527-output layer. Indeed, the paper reads:
"Dropout [38] is applied after each downsampling operation and fully connected layers to prevent systems from overfitting."
Therefore, I expected to see the following:
Am I missing something?
Thank you,
Alessandro
The text was updated successfully, but these errors were encountered: