vision
jun-untitled's picture
Update README.md
c9fe1ed
---
#license: not yet
tags:
- vision
datasets:
- kakaobrain/coyo-labeled-300m
annotations_creators:
- machine-generated
license:
- apache-2.0
pretty_name: COYO-Labeled-300M
task_categories:
- image-classification
task_ids:
- multi-label-image-classification
- multi-class-image-classification
inference: false
---
# Vision Transformer (large-sized model)
Vision Transformer (ViT) model pre-trained on [COYO-Labeled-300M](https://github.com/kakaobrain/coyo-dataset/tree/main/subset/COYO-Labeled-300M) (300 million images, 21,841 classes) at resolution 224x224. It was introduced in the paper [An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale](https://arxiv.org/abs/2010.11929) by Dosovitskiy et al. However, since the JFT-300M is a private dataset, we tried to reproduce it using the publicly available [COYO-Labeled-300M](https://github.com/kakaobrain/coyo-dataset/tree/main/subset/COYO-Labeled-300M) dataset.
Thanks to Hugging Face team for converting weights of ViT trained in Tensorflow to be used on Pytorch, JAX/Flax and Tensorflow in Hugging Face.
## Model description
The Vision Transformer (ViT) is a transformer model pretrained on a large collection of images in a supervised fashion, namely [COYO-Labeled-300M](https://github.com/kakaobrain/coyo-dataset/tree/main/subset/COYO-Labeled-300M), at a resolution of 224x224 pixels.
Images are presented to the model as a sequence of fixed-size patches (resolution 16x16), which are linearly embedded. One also adds a [CLS] token to the beginning of a sequence to use it for classification tasks. One also adds absolute position embeddings before feeding the sequence to the layers of the Transformer.
This ViT model is finetuned with Imagenet-1k at resolution 384x384. Please see details [here](https://github.com/kakaobrain/coyo-vit)
## Intended uses & limitations
You can use weights from ViT models for image classification, downstream. Codes for reproduction are also provided. Please see this [github repository](https://github.com/kakaobrain/coyo-vit) for pretraining and finetuning code.
### How to use
Here is how to use this model in PyTorch:
```python
WIP
```
Here is how to use this model in JAX/Flax:
```python
WIP
```
Here is how to use this model in Tensorflow:
```python
WIP
```
## Training data
The ViT model was pretrained on [COYO-Labeled-300M](https://github.com/kakaobrain/coyo-dataset/tree/main/subset/COYO-Labeled-300M), a dataset consisting of 300 million images and 21k classes.
## Training procedure
### Preprocessing
The exact details of preprocessing of images during training/validation can be found [here](https://github.com/kakaobrain/coyo-vit).
Images are inception-cropped to the same resolution (224x224) and normalized across the RGB channels with mean (0.5, 0.5, 0.5) and standard deviation (0.5, 0.5, 0.5).
### Pretraining
The model was trained on TPUv3 hardware. All model variants are trained with a batch size of 4096 and learning rate warmup of 10k steps. Pre-training resolution is 224.
More detail, please see [here](https://github.com/kakaobrain/coyo-vit)
### Evaluation results
| Model | Upstream Dataset | Resolution | ImageNet (downstream) | ImageNet-ReaL (dwonstream) | Public |
|---------- |------------------- |------------ |----------------------- |---------------------------- |-------- |
| ViT-L/16 | JFT-300M | 512 | 87.76 | 90.54 | X |
| ViT-L/16 | COYO-Labeled-300M | 512 | 87.24 (-0.52) | 90.03 (-0.51) | O |
| ViT-L/16 | JFT-300M | 384 | 87.12 | 89.99 | X |
| ViT-L/16 | COYO-Labeled-300M | 384 | 86.72 (-0.4) | 89.84 (-0.15) | O |
## Citation
```bibtex
@misc{kakaobrain2022coyo-vit,
title = {COYO-ViT},
author = {Lee, Sungjun and Park, Beomhee},
year = {2022},
howpublished = {\url{https://github.com/kakaobrain/coyo-vit}},
}
```
```bibtex
@misc{kakaobrain2022coyo-700m,
title = {COYO-700M: Image-Text Pair Dataset},
author = {Byeon, Minwoo and Park, Beomhee and Kim, Haecheon and Lee, Sungjun and Baek, Woonhyuk and Kim, Saehoon},
year = {2022},
howpublished = {\url{https://github.com/kakaobrain/coyo-dataset}},
}
```
```bibtex
@misc{dosovitskiy2020image,
title = {An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale},
author = {Alexey Dosovitskiy and Lucas Beyer and Alexander Kolesnikov and Dirk Weissenborn and Xiaohua Zhai and Thomas Unterthiner and Mostafa Dehghani and Matthias Minderer and Georg Heigold and Sylvain Gelly and Jakob Uszkoreit and Neil Houlsby},
year = {2020},
eprint = {2010.11929},
archivePrefix = {arXiv},
primaryClass = {cs.CV}
}
```
## License
The source codes are licensed under Apache 2.0 License.