30
30
31
31
class LayerNorm2d (nn .LayerNorm ):
32
32
def forward (self , x : Tensor ) -> Tensor :
33
- # TODO: Benchmark this against the approach described at https://github.com/pytorch/vision/pull/5197#discussion_r786251298
34
33
x = x .permute (0 , 2 , 3 , 1 )
35
34
x = F .layer_norm (x , self .normalized_shape , self .weight , self .bias , self .eps )
36
35
x = x .permute (0 , 3 , 1 , 2 )
@@ -219,7 +218,7 @@ def _convnext(
219
218
220
219
class ConvNeXt_Tiny_Weights (WeightsEnum ):
221
220
IMAGENET1K_V1 = Weights (
222
- url = "https://download.pytorch.org/models/convnext_tiny-47b116bd .pth" , # TODO: repackage
221
+ url = "https://download.pytorch.org/models/convnext_tiny-983f1562 .pth" ,
223
222
transforms = partial (ImageNetEval , crop_size = 224 , resize_size = 236 ),
224
223
meta = {
225
224
** _COMMON_META ,
@@ -233,7 +232,7 @@ class ConvNeXt_Tiny_Weights(WeightsEnum):
233
232
234
233
class ConvNeXt_Small_Weights (WeightsEnum ):
235
234
IMAGENET1K_V1 = Weights (
236
- url = "https://download.pytorch.org/models/convnext_small-9aa23d28 .pth" , # TODO: repackage
235
+ url = "https://download.pytorch.org/models/convnext_small-0c510722 .pth" ,
237
236
transforms = partial (ImageNetEval , crop_size = 224 , resize_size = 230 ),
238
237
meta = {
239
238
** _COMMON_META ,
@@ -247,7 +246,7 @@ class ConvNeXt_Small_Weights(WeightsEnum):
247
246
248
247
class ConvNeXt_Base_Weights (WeightsEnum ):
249
248
IMAGENET1K_V1 = Weights (
250
- url = "https://download.pytorch.org/models/convnext_base-3b9f985d .pth" , # TODO: repackage
249
+ url = "https://download.pytorch.org/models/convnext_base-6075fbad .pth" ,
251
250
transforms = partial (ImageNetEval , crop_size = 224 , resize_size = 232 ),
252
251
meta = {
253
252
** _COMMON_META ,
@@ -261,7 +260,7 @@ class ConvNeXt_Base_Weights(WeightsEnum):
261
260
262
261
class ConvNeXt_Large_Weights (WeightsEnum ):
263
262
IMAGENET1K_V1 = Weights (
264
- url = "https://download.pytorch.org/models/convnext_large-d73f62ac .pth" , # TODO: repackage
263
+ url = "https://download.pytorch.org/models/convnext_large-ea097f82 .pth" ,
265
264
transforms = partial (ImageNetEval , crop_size = 224 , resize_size = 232 ),
266
265
meta = {
267
266
** _COMMON_META ,
0 commit comments