|
@ -259,7 +259,7 @@ class CustomAttentionFFN(nn.Module): |
|
|
nn.Linear(dim, dim * 4), |
|
|
nn.Linear(dim, dim * 4), |
|
|
nn.GELU(), |
|
|
nn.GELU(), |
|
|
nn.Linear(dim * 4, dim), |
|
|
nn.Linear(dim * 4, dim), |
|
|
nn.DropPath(proj_drop) |
|
|
nn.Dropout(proj_drop) |
|
|
) |
|
|
) |
|
|
self.norm1 = nn.LayerNorm(dim) |
|
|
self.norm1 = nn.LayerNorm(dim) |
|
|
self.norm2 = nn.LayerNorm(dim) |
|
|
self.norm2 = nn.LayerNorm(dim) |
|
@ -342,7 +342,7 @@ class ViTSam(BaseBackbone): |
|
|
nn.Linear(embed_dim, embed_dim * 4), |
|
|
nn.Linear(embed_dim, embed_dim * 4), |
|
|
nn.GELU(), |
|
|
nn.GELU(), |
|
|
nn.Linear(embed_dim * 4, embed_dim), |
|
|
nn.Linear(embed_dim * 4, embed_dim), |
|
|
nn.DropPath(drop_rate) |
|
|
nn.Dropout(drop_rate) |
|
|
) |
|
|
) |
|
|
self.sam_norm = norm_layer(embed_dim) |
|
|
self.sam_norm = norm_layer(embed_dim) |
|
|
|
|
|
|
|
|