| from transformers import PretrainedConfig |
|
|
|
|
| class MELPEncoderConfig(PretrainedConfig): |
| model_type = "melp" |
|
|
| def __init__( |
| self, |
| model_size: str = "small", |
| shared_emb_dim: int = 256, |
| embed_dim_caption: int = 768, |
| use_attentional_pool_contrast: bool = True, |
| use_attentional_pool_caption: bool = True, |
| n_queries_contrast: int = 14, |
| n_queries_caption: int = 128, |
| attn_pooler_heads: int = 8, |
| proj: str = "linear", |
| drop: float = 0., |
| proj_bias: bool = False, |
| num_leads: int = 12, |
| **kwargs |
| ): |
| self.model_size = model_size |
| self.shared_emb_dim = shared_emb_dim |
| self.embed_dim_caption = embed_dim_caption |
| self.use_attentional_pool_contrast = use_attentional_pool_contrast |
| self.use_attentional_pool_caption = use_attentional_pool_caption |
| self.n_queries_contrast = n_queries_contrast |
| self.n_queries_caption = n_queries_caption |
| self.attn_pooler_heads = attn_pooler_heads |
| self.proj = proj |
| self.drop = drop |
| self.proj_bias = proj_bias |
| self.num_leads = num_leads |
| super().__init__(**kwargs) |
|
|