| { |
| "architectures": [ |
| "ModernBertForSequenceClassification" |
| ], |
| "attention_bias": false, |
| "attention_dropout": 0.0, |
| "bos_token_id": 50281, |
| "classifier_activation": "gelu", |
| "classifier_bias": false, |
| "classifier_dropout": 0.0, |
| "classifier_pooling": "mean", |
| "cls_token_id": 50281, |
| "decoder_bias": true, |
| "deterministic_flash_attn": false, |
| "dtype": "float32", |
| "embedding_dropout": 0.0, |
| "eos_token_id": 50282, |
| "global_attn_every_n_layers": 3, |
| "global_rope_theta": 160000.0, |
| "gradient_checkpointing": false, |
| "hidden_activation": "gelu", |
| "hidden_size": 768, |
| "id2label": { |
| "0": 7, |
| "1": 17, |
| "10": 58, |
| "100": 231, |
| "101": 232, |
| "102": 233, |
| "103": 234, |
| "104": 235, |
| "105": 236, |
| "106": 237, |
| "107": 238, |
| "108": 239, |
| "109": 240, |
| "11": 64, |
| "110": 241, |
| "111": 242, |
| "112": 243, |
| "113": 244, |
| "114": 245, |
| "115": 246, |
| "116": 247, |
| "117": 248, |
| "118": 249, |
| "119": 250, |
| "12": 67, |
| "120": 251, |
| "121": 252, |
| "122": 253, |
| "123": 254, |
| "124": 255, |
| "125": 256, |
| "126": 257, |
| "127": 258, |
| "128": 259, |
| "129": 260, |
| "13": 83, |
| "130": 261, |
| "131": 262, |
| "132": 263, |
| "133": 264, |
| "134": 265, |
| "135": 266, |
| "136": 267, |
| "137": 268, |
| "138": 269, |
| "139": 270, |
| "14": 91, |
| "140": 271, |
| "141": 272, |
| "142": 273, |
| "143": 274, |
| "144": 275, |
| "145": 276, |
| "146": 277, |
| "147": 278, |
| "148": 279, |
| "149": 280, |
| "15": 100, |
| "150": 281, |
| "151": 282, |
| "152": 283, |
| "153": 284, |
| "154": 154, |
| "155": 155, |
| "156": 156, |
| "157": 157, |
| "158": 158, |
| "159": 159, |
| "16": 102, |
| "160": 160, |
| "161": 161, |
| "162": 162, |
| "163": 163, |
| "164": 164, |
| "165": 165, |
| "166": 166, |
| "167": 167, |
| "168": 168, |
| "169": 169, |
| "17": 126, |
| "170": 170, |
| "171": 171, |
| "172": 172, |
| "173": 173, |
| "174": 174, |
| "175": 175, |
| "176": 176, |
| "177": 177, |
| "178": 178, |
| "179": 179, |
| "18": 127, |
| "180": 180, |
| "181": 181, |
| "182": 182, |
| "183": 183, |
| "184": 184, |
| "185": 185, |
| "186": 186, |
| "187": 187, |
| "188": 188, |
| "189": 189, |
| "19": 128, |
| "190": 190, |
| "191": 191, |
| "192": 192, |
| "193": 193, |
| "194": 194, |
| "195": 195, |
| "196": 196, |
| "197": 197, |
| "198": 198, |
| "199": 199, |
| "2": 31, |
| "20": 131, |
| "200": 200, |
| "201": 201, |
| "202": 202, |
| "203": 203, |
| "204": 204, |
| "205": 205, |
| "206": 206, |
| "207": 207, |
| "208": 208, |
| "209": 209, |
| "21": 135, |
| "210": 210, |
| "211": 211, |
| "212": 212, |
| "213": 213, |
| "214": 214, |
| "215": 215, |
| "216": 216, |
| "217": 217, |
| "218": 218, |
| "219": 219, |
| "22": 138, |
| "220": 220, |
| "221": 221, |
| "222": 222, |
| "223": 223, |
| "224": 224, |
| "225": 225, |
| "226": 226, |
| "227": 227, |
| "228": 228, |
| "229": 229, |
| "23": 153, |
| "230": 230, |
| "231": 231, |
| "232": 232, |
| "233": 233, |
| "234": 234, |
| "235": 235, |
| "236": 236, |
| "237": 237, |
| "238": 238, |
| "239": 239, |
| "24": 154, |
| "240": 240, |
| "241": 241, |
| "242": 242, |
| "243": 243, |
| "244": 244, |
| "245": 245, |
| "246": 246, |
| "247": 247, |
| "248": 248, |
| "249": 249, |
| "25": 155, |
| "250": 250, |
| "251": 251, |
| "252": 252, |
| "253": 253, |
| "254": 254, |
| "255": 255, |
| "256": 256, |
| "257": 257, |
| "258": 258, |
| "259": 259, |
| "26": 156, |
| "260": 260, |
| "261": 261, |
| "262": 262, |
| "263": 263, |
| "264": 264, |
| "265": 265, |
| "266": 266, |
| "267": 267, |
| "268": 268, |
| "269": 269, |
| "27": 157, |
| "270": 270, |
| "271": 271, |
| "272": 272, |
| "273": 273, |
| "274": 274, |
| "275": 275, |
| "276": 276, |
| "277": 277, |
| "278": 278, |
| "279": 279, |
| "28": 158, |
| "280": 280, |
| "281": 281, |
| "282": 282, |
| "283": 283, |
| "284": 284, |
| "29": 159, |
| "3": 38, |
| "30": 160, |
| "31": 161, |
| "32": 162, |
| "33": 163, |
| "34": 164, |
| "35": 165, |
| "36": 166, |
| "37": 167, |
| "38": 168, |
| "39": 169, |
| "4": 39, |
| "40": 171, |
| "41": 172, |
| "42": 173, |
| "43": 174, |
| "44": 175, |
| "45": 176, |
| "46": 177, |
| "47": 178, |
| "48": 179, |
| "49": 180, |
| "5": 40, |
| "50": 181, |
| "51": 182, |
| "52": 183, |
| "53": 184, |
| "54": 185, |
| "55": 186, |
| "56": 187, |
| "57": 188, |
| "58": 189, |
| "59": 190, |
| "6": 43, |
| "60": 191, |
| "61": 192, |
| "62": 193, |
| "63": 194, |
| "64": 195, |
| "65": 196, |
| "66": 197, |
| "67": 198, |
| "68": 199, |
| "69": 200, |
| "7": 48, |
| "70": 201, |
| "71": 202, |
| "72": 203, |
| "73": 204, |
| "74": 205, |
| "75": 206, |
| "76": 207, |
| "77": 208, |
| "78": 209, |
| "79": 210, |
| "8": 53, |
| "80": 211, |
| "81": 212, |
| "82": 213, |
| "83": 214, |
| "84": 215, |
| "85": 216, |
| "86": 217, |
| "87": 218, |
| "88": 219, |
| "89": 220, |
| "9": 56, |
| "90": 221, |
| "91": 222, |
| "92": 223, |
| "93": 224, |
| "94": 225, |
| "95": 226, |
| "96": 227, |
| "97": 228, |
| "98": 229, |
| "99": 230 |
| }, |
| "initializer_cutoff_factor": 2.0, |
| "initializer_range": 0.02, |
| "intermediate_size": 1152, |
| "label2id": { |
| "7": "0", |
| "17": "1", |
| "31": "2", |
| "38": "3", |
| "39": "4", |
| "40": "5", |
| "43": "6", |
| "48": "7", |
| "53": "8", |
| "56": "9", |
| "58": "10", |
| "64": "11", |
| "67": "12", |
| "83": "13", |
| "91": "14", |
| "100": "15", |
| "102": "16", |
| "126": "17", |
| "127": "18", |
| "128": "19", |
| "131": "20", |
| "135": "21", |
| "138": "22", |
| "153": "23", |
| "154": "154", |
| "155": "155", |
| "156": "156", |
| "157": "157", |
| "158": "158", |
| "159": "159", |
| "160": "160", |
| "161": "161", |
| "162": "162", |
| "163": "163", |
| "164": "164", |
| "165": "165", |
| "166": "166", |
| "167": "167", |
| "168": "168", |
| "169": "169", |
| "170": "170", |
| "171": "171", |
| "172": "172", |
| "173": "173", |
| "174": "174", |
| "175": "175", |
| "176": "176", |
| "177": "177", |
| "178": "178", |
| "179": "179", |
| "180": "180", |
| "181": "181", |
| "182": "182", |
| "183": "183", |
| "184": "184", |
| "185": "185", |
| "186": "186", |
| "187": "187", |
| "188": "188", |
| "189": "189", |
| "190": "190", |
| "191": "191", |
| "192": "192", |
| "193": "193", |
| "194": "194", |
| "195": "195", |
| "196": "196", |
| "197": "197", |
| "198": "198", |
| "199": "199", |
| "200": "200", |
| "201": "201", |
| "202": "202", |
| "203": "203", |
| "204": "204", |
| "205": "205", |
| "206": "206", |
| "207": "207", |
| "208": "208", |
| "209": "209", |
| "210": "210", |
| "211": "211", |
| "212": "212", |
| "213": "213", |
| "214": "214", |
| "215": "215", |
| "216": "216", |
| "217": "217", |
| "218": "218", |
| "219": "219", |
| "220": "220", |
| "221": "221", |
| "222": "222", |
| "223": "223", |
| "224": "224", |
| "225": "225", |
| "226": "226", |
| "227": "227", |
| "228": "228", |
| "229": "229", |
| "230": "230", |
| "231": "231", |
| "232": "232", |
| "233": "233", |
| "234": "234", |
| "235": "235", |
| "236": "236", |
| "237": "237", |
| "238": "238", |
| "239": "239", |
| "240": "240", |
| "241": "241", |
| "242": "242", |
| "243": "243", |
| "244": "244", |
| "245": "245", |
| "246": "246", |
| "247": "247", |
| "248": "248", |
| "249": "249", |
| "250": "250", |
| "251": "251", |
| "252": "252", |
| "253": "253", |
| "254": "254", |
| "255": "255", |
| "256": "256", |
| "257": "257", |
| "258": "258", |
| "259": "259", |
| "260": "260", |
| "261": "261", |
| "262": "262", |
| "263": "263", |
| "264": "264", |
| "265": "265", |
| "266": "266", |
| "267": "267", |
| "268": "268", |
| "269": "269", |
| "270": "270", |
| "271": "271", |
| "272": "272", |
| "273": "273", |
| "274": "274", |
| "275": "275", |
| "276": "276", |
| "277": "277", |
| "278": "278", |
| "279": "279", |
| "280": "280", |
| "281": "281", |
| "282": "282", |
| "283": "283", |
| "284": "284" |
| }, |
| "layer_norm_eps": 1e-05, |
| "local_attention": 128, |
| "local_rope_theta": 10000.0, |
| "max_position_embeddings": 8192, |
| "mlp_bias": false, |
| "mlp_dropout": 0.0, |
| "model_type": "modernbert", |
| "norm_bias": false, |
| "norm_eps": 1e-05, |
| "num_attention_heads": 12, |
| "num_hidden_layers": 22, |
| "pad_token_id": 50283, |
| "position_embedding_type": "absolute", |
| "problem_type": "single_label_classification", |
| "repad_logits_with_grad": false, |
| "sep_token_id": 50282, |
| "sparse_pred_ignore_index": -100, |
| "sparse_prediction": false, |
| "transformers_version": "4.57.3", |
| "vocab_size": 50368 |
| } |
|
|