rwightman HF staff commited on
Commit
ecfbd22
·
1 Parent(s): 387c31c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -1
README.md CHANGED
@@ -34,7 +34,6 @@ The models are trained at 256x256 (working on 384 variants) image resolution.
34
 
35
  At 256x256, the ConvNext-Large-D used roughly 1/2 the training FLOPs to achieve accuracy greater than previous L/14 model trained on LAION-2B. L/14 model is ~1.65x more GMAC, 1.45x more activations, and 1.22x more parameters. The ConvNeXt was trained with 26B samples-seen and L/14 with 34B.
36
 
37
- All models in this series were trained for 13B samples and have ImageNet Zero-Shot top-1 of >= 70.8%. Comparing to ViT-B/16 at 34B SS with zero-shot of 70.2% (68.1% for 13B SS) this suggests the ConvNeXt architecture may be more sample efficient in this range of model scale. More experiments needed to confirm.
38
 
39
  | Model | Dataset | Resolution | AugReg | Top-1 ImageNet Zero-Shot (%) |
40
  | ----- | ------- | ---------- | ------------ | --------- |
 
34
 
35
  At 256x256, the ConvNext-Large-D used roughly 1/2 the training FLOPs to achieve accuracy greater than previous L/14 model trained on LAION-2B. L/14 model is ~1.65x more GMAC, 1.45x more activations, and 1.22x more parameters. The ConvNeXt was trained with 26B samples-seen and L/14 with 34B.
36
 
 
37
 
38
  | Model | Dataset | Resolution | AugReg | Top-1 ImageNet Zero-Shot (%) |
39
  | ----- | ------- | ---------- | ------------ | --------- |