Skip to content

Commit b4d93f2

Browse files
[docs] Add UNITER model to website
Add citation and training instructions under projects/uniter. ghstack-source-id: 96dd13e Pull Request resolved: #1144
1 parent 69d0a2a commit b4d93f2

File tree

3 files changed

+40
-1
lines changed

3 files changed

+40
-1
lines changed

website/docs/notes/model_zoo.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -21,6 +21,7 @@ Here is the list of models currently implemented in MMF:
2121
| Movie MCAN | movie_mcan | vqa2 | [paper](https://arxiv.org/abs/2004.11883) |
2222
| Pythia | pythia | textvqa, vizwiz, vqa2, visual_genome | [paper](https://arxiv.org/abs/1904.08920) |
2323
| Unimodal | unimodal | hateful_memes | |
24+
| UNITER | uniter | vqa2, masked_coco | [paper](https://arxiv.org/abs/1909.11740) |
2425
| VilBERT | vilbert | hateful_memes, coco, conceptual_captions, vqa2, mmimdb, nlvr2, visual_entailment, vizwiz, vqa2 |[paper](https://arxiv.org/abs/1908.02265)|
2526
| Visual BERT | visual_bert | gqa, hateful_memes, localized_narratives, coco, conceptual_captions, sbu, vqa2, mmimdb, nlvr2, visual_entailment, vizwiz|[paper](https://arxiv.org/abs/1908.03557)|
2627

website/docs/projects/uniter.md

Lines changed: 38 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,38 @@
1+
---
2+
id: uniter
3+
sidebar_label: UNITER
4+
title: "UNITER: UNiversal Image-TExt Representation Learning"
5+
---
6+
7+
This repository contains the code for pytorch implementation of UNITER model, released originally under this ([repo](https://github.com/ChenRocks/UNITER)). Please cite the following papers if you are using UNITER model from mmf:
8+
9+
* Chen, Y.-C., Li, L., Yu, L., Kholy, A. E., Ahmed, F., Gan,
10+
Z., Cheng, Y., and jing Liu, J. *Uniter: Universal imagetext representation learning.* In European Conference on
11+
Computer Vision, 2020b. ([arXiV](https://arxiv.org/pdf/1909.11740))
12+
```
13+
@inproceedings{chen2020uniter,
14+
title={Uniter: Universal image-text representation learning},
15+
author={Chen, Yen-Chun and Li, Linjie and Yu, Licheng and Kholy, Ahmed El and Ahmed, Faisal and Gan, Zhe and Cheng, Yu and Liu, Jingjing},
16+
booktitle={ECCV},
17+
year={2020}
18+
}
19+
```
20+
21+
## Installation
22+
23+
Follow installation instructions in the [documentation](https://mmf.readthedocs.io/en/latest/notes/installation.html).
24+
25+
## Training
26+
27+
To train a fresh UNITER model on the VQA2.0 dataset, run the following command
28+
```
29+
mmf_run config=projects/uniter/configs/vqa2/defaults.yaml run_type=train_val dataset=vqa2 model=uniter
30+
```
31+
32+
To pretrain UNITER on the masked COCO dataset, run the following command
33+
```
34+
mmf_run config=projects/uniter/configs/masked_coco/defaults.yaml run_type=train_val dataset=masked_coco model=uniter
35+
```
36+
37+
38+
Based on the config used and `do_pretraining` defined in the config, the model can use the pretraining recipe described in the UNITER paper, or be finetuned on downstream tasks.

website/sidebars.js

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -45,6 +45,6 @@ module.exports = {
4545
'challenges/textvqa_challenge',
4646
'challenges/vqa_challenge',
4747
],
48-
Projects: ['projects/butd', 'projects/m4c', 'projects/m4c_captioner', 'projects/movie_mcan', 'projects/unit'],
48+
Projects: ['projects/butd', 'projects/m4c', 'projects/m4c_captioner', 'projects/movie_mcan', 'projects/unit', 'projects/uniter'],
4949
},
5050
};

0 commit comments

Comments
 (0)