Skip to content

Commit ff7fe24

Browse files
committed
Rename TransformerTextualHead and USE_LOOKAHEAD config.
1 parent 11c5793 commit ff7fe24

22 files changed

+72
-71
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ model = torch.hub.load("kdexd/virtex", "resnet50", pretrained=True)
3636

3737
### Note (For returning users before January 2021):
3838

39-
The pretrained models in our model zoo have changed in [`v1.0`](https://github.com/kdexd/virtex/releases/tag/v1.0).
39+
The pretrained models in our model zoo have changed from [`v1.0`](https://github.com/kdexd/virtex/releases/tag/v1.0) onwards.
4040
They are slightly better tuned than older models, and reproduce the results in our
4141
CVPR 2021 accepted paper ([arXiv v2](https://arxiv.org/abs/2006.06666v2)).
4242
Some training and evaluation hyperparams are changed since [`v0.9`](https://github.com/kdexd/virtex/releases/tag/v0.9).

configs/_base_bicaptioning_R_50_L1_H1024.yaml

Lines changed: 15 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -3,9 +3,9 @@
33
# ResNet-50 + (L = 1, H = 1024) transformer trained for 500K iterations.
44
# -----------------------------------------------------------------------------
55
RANDOM_SEED: 0
6-
AMP: True
7-
CUDNN_BENCHMARK: True
8-
CUDNN_DETERMINISTIC: False
6+
AMP: true
7+
CUDNN_BENCHMARK: true
8+
CUDNN_DETERMINISTIC: false
99

1010
DATA:
1111
ROOT: "datasets/coco"
@@ -31,33 +31,36 @@ DATA:
3131
- "normalize"
3232

3333
USE_PERCENTAGE: 100.0
34-
USE_SINGLE_CAPTION: False
34+
USE_SINGLE_CAPTION: false
3535

3636
MODEL:
37-
NAME: "bicaptioning"
37+
NAME: "virtex"
3838
VISUAL:
3939
NAME: "torchvision::resnet50"
4040
PRETRAINED: false
4141
FROZEN: false
4242
TEXTUAL:
43-
NAME: "transformer_postnorm::L1_H1024_A16_F4096"
43+
NAME: "transdec_postnorm::L1_H1024_A16_F4096"
4444
DROPOUT: 0.1
4545

4646
OPTIM:
4747
OPTIMIZER_NAME: "sgd"
4848
SGD_MOMENTUM: 0.9
4949
WEIGHT_DECAY: 0.0001
50-
NO_DECAY: ".*textual.(embedding|transformer).*(norm.*|bias)"
51-
CLIP_GRAD_NORM: 10
5250

53-
USE_LOOKAHEAD: True
54-
LOOKAHEAD_ALPHA: 0.5
55-
LOOKAHEAD_STEPS: 5
51+
LOOKAHEAD:
52+
USE: true
53+
ALPHA: 0.5
54+
STEPS: 5
5655

5756
BATCH_SIZE: 256
5857
CNN_LR: 0.2
5958
LR: 0.001
6059
NUM_ITERATIONS: 500000
6160

6261
WARMUP_STEPS: 10000
63-
LR_DECAY_NAME: cosine
62+
LR_DECAY_NAME: "cosine"
63+
64+
NO_DECAY: ".*textual.(embedding|transformer).*(norm.*|bias)"
65+
CLIP_GRAD_NORM: 10.0
66+

configs/depth_ablations/bicaptioning_R_50_L2_H1024.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,4 +2,4 @@ _BASE_: "../_base_bicaptioning_R_50_L1_H1024.yaml"
22

33
MODEL:
44
TEXTUAL:
5-
NAME: "transformer_postnorm::L2_H1024_A16_F4096"
5+
NAME: "transdec_postnorm::L2_H1024_A16_F4096"

configs/depth_ablations/bicaptioning_R_50_L3_H1024.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,4 +2,4 @@ _BASE_: "../_base_bicaptioning_R_50_L1_H1024.yaml"
22

33
MODEL:
44
TEXTUAL:
5-
NAME: "transformer_postnorm::L3_H1024_A16_F4096"
5+
NAME: "transdec_postnorm::L3_H1024_A16_F4096"

configs/depth_ablations/bicaptioning_R_50_L4_H1024.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,4 +2,4 @@ _BASE_: "../_base_bicaptioning_R_50_L1_H1024.yaml"
22

33
MODEL:
44
TEXTUAL:
5-
NAME: "transformer_postnorm::L4_H1024_A16_F4096"
5+
NAME: "transdec_postnorm::L4_H1024_A16_F4096"

configs/downstream/imagenet_clf.yaml

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,8 @@
11
RANDOM_SEED: 0
22
# Don't need AMP to train a tiny linear layer.
3-
AMP: False
4-
CUDNN_BENCHMARK: True
5-
CUDNN_DETERMINISTIC: False
3+
AMP: false
4+
CUDNN_BENCHMARK: true
5+
CUDNN_DETERMINISTIC: false
66

77
DATA:
88
ROOT: "datasets/imagenet"
@@ -24,7 +24,8 @@ OPTIM:
2424
SGD_MOMENTUM: 0.9
2525
WEIGHT_DECAY: 0.0
2626
NO_DECAY: "none"
27-
USE_LOOKAHEAD: False
27+
LOOKAHEAD:
28+
USE: false
2829

2930
LR: 0.3
3031
WARMUP_STEPS: 0

configs/downstream/inaturalist_clf.yaml

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
RANDOM_SEED: 0
2-
AMP: True
3-
CUDNN_BENCHMARK: True
4-
CUDNN_DETERMINISTIC: False
2+
AMP: true
3+
CUDNN_BENCHMARK: true
4+
CUDNN_DETERMINISTIC: false
55

66
DATA:
77
ROOT: "datasets/inaturalist"
@@ -23,7 +23,8 @@ OPTIM:
2323
SGD_MOMENTUM: 0.9
2424
WEIGHT_DECAY: 0.0001
2525
NO_DECAY: "none"
26-
USE_LOOKAHEAD: False
26+
LOOKAHEAD:
27+
USE: false
2728

2829
LR: 0.025
2930
WARMUP_STEPS: 0

configs/task_ablations/bicaptioning_R_50_L1_H2048.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,4 +2,4 @@ _BASE_: "../_base_bicaptioning_R_50_L1_H1024.yaml"
22

33
MODEL:
44
TEXTUAL:
5-
NAME: "transformer_postnorm::L1_H2048_A32_F8192"
5+
NAME: "transdec_postnorm::L1_H2048_A32_F8192"

configs/task_ablations/captioning_R_50_L1_H2048.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,4 +3,4 @@ _BASE_: "../_base_bicaptioning_R_50_L1_H1024.yaml"
33
MODEL:
44
NAME: "captioning"
55
TEXTUAL:
6-
NAME: "transformer_postnorm::L1_H2048_A32_F8192"
6+
NAME: "transdec_postnorm::L1_H2048_A32_F8192"

configs/task_ablations/masked_lm_R_50_L1_H2048.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,4 +3,4 @@ _BASE_: "../_base_bicaptioning_R_50_L1_H1024.yaml"
33
MODEL:
44
NAME: "masked_lm"
55
TEXTUAL:
6-
NAME: "transformer_postnorm::L1_H2048_A32_F8192"
6+
NAME: "transdec_postnorm::L1_H2048_A32_F8192"

0 commit comments

Comments
 (0)