| { |
| |
| "cache_dir":"/tmp", |
| "downstream_fusion":false, |
| "downstream_fusion_layers":1, |
| "downstream_fusion_method":"elmo", |
| "drop_rate":0.1, |
| "freeze_RoBERTa":false, |
| "freeze_ViT":false, |
| "freeze_layer_count_roberta":false, |
| "freeze_layer_count_vit":false, |
| "head_hidden_scale":2, |
| "hidden_size":768, |
| "image_size":288, |
| "input_text_embed_size":768, |
| "link_tower_shared":false, |
| "link_tower_type":"add", |
| "log_dir":"log_dir", |
| "loss_names":{"contras": 0, |
| "irtr": 0, |
| "itm": 0, |
| "mlm": 0, |
| "mpp": 0, |
| "nlvr2": 0, |
| "snli": 0, |
| "vcr": 0, |
| "vcr_qar": 0, |
| "vqa": 1}, |
| "max_text_len":50, |
| "mlp_ratio":4, |
| "model_type":"bridgetower", |
| "num_heads":12, |
| "num_layers":6, |
| "num_nodes":1, |
| "only_load_cross_modal_from_meter":false, |
| "patch_size":16, |
| "resolution_before":224, |
| "stop_gradient":false, |
| "task_head_layers":2, |
| "test_only":false, |
| "tokenizer":"roberta-base", |
| "unfreeze_RoBERTa_attention":false, |
| "unfreeze_RoBERTa_embeddings":false, |
| "unfreeze_RoBERTa_encoder":false, |
| "unfreeze_RoBERTa_layernorm":false, |
| "unfreeze_ViT_attention":false, |
| "unfreeze_ViT_layernorm":false, |
| "vit":"ViT-B/16", |
| "vit_layernorm_init_from_vit":false, |
| "vit_layernorm_shared":true, |
| "vit_remove_last":false, |
| "vocab_size":50265 |
| } |
|
|