You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In the paper, when stride=2, the left branch (Figure 5 (c)) has two depth-wise layers. Their output channels are the same, which is expand_ratio * C. But in the codes, due to the chain rule, the second layer's output has expand_ratio * expand_ratio * C channels.
Also, the expansion is with respect to the input dimension, instead of the output dimension.
Thank you for reporting the issue! I have to admit that I'm quite out of the loop regarding ML, as this was part of my studies. I have created a pull request that should fix the double expension, feel free to have a look.
Also, the expansion is with respect to the input dimension, instead of the output dimension.
I'm not entirely sure what you mean by that, and how I'd go about fixing that. I'd be happy to accept a pull request :)
In the paper, when
stride=2
, the left branch (Figure 5 (c)) has two depth-wise layers. Their output channels are the same, which isexpand_ratio * C
. But in the codes, due to the chain rule, the second layer's output hasexpand_ratio * expand_ratio * C
channels.Also, the expansion is with respect to the input dimension, instead of the output dimension.
bisenetv2-tf2/model.py
Lines 14 to 37 in e018b3b
The text was updated successfully, but these errors were encountered: