haoyang-amd commited on
Commit
b9241dc
·
verified ·
1 Parent(s): 82daf63
config.json CHANGED
@@ -26,7 +26,7 @@
26
  "num_key_value_heads": 4,
27
  "output_router_logits": false,
28
  "quantization_config": {
29
- "algo_config": null,
30
  "exclude": [
31
  "model.layers.0.mlp.gate",
32
  "model.layers.1.mlp.gate",
@@ -126,7 +126,7 @@
126
  "quant_method": "quark",
127
  "quant_mode": "eager_mode",
128
  "softmax_quant_spec": null,
129
- "version": "0.10"
130
  },
131
  "rms_norm_eps": 1e-06,
132
  "rope_scaling": null,
@@ -134,7 +134,7 @@
134
  "router_aux_loss_coef": 0.001,
135
  "sliding_window": null,
136
  "tie_word_embeddings": false,
137
- "transformers_version": "4.56.2",
138
  "use_cache": true,
139
  "use_sliding_window": false,
140
  "vocab_size": 151936
 
26
  "num_key_value_heads": 4,
27
  "output_router_logits": false,
28
  "quantization_config": {
29
+ "algo_config": [],
30
  "exclude": [
31
  "model.layers.0.mlp.gate",
32
  "model.layers.1.mlp.gate",
 
126
  "quant_method": "quark",
127
  "quant_mode": "eager_mode",
128
  "softmax_quant_spec": null,
129
+ "version": "0.10+b8ad5c1d29"
130
  },
131
  "rms_norm_eps": 1e-06,
132
  "rope_scaling": null,
 
134
  "router_aux_loss_coef": 0.001,
135
  "sliding_window": null,
136
  "tie_word_embeddings": false,
137
+ "transformers_version": "4.57.1",
138
  "use_cache": true,
139
  "use_sliding_window": false,
140
  "vocab_size": 151936
generation_config.json CHANGED
@@ -9,5 +9,5 @@
9
  "temperature": 0.6,
10
  "top_k": 20,
11
  "top_p": 0.95,
12
- "transformers_version": "4.56.2"
13
  }
 
9
  "temperature": 0.6,
10
  "top_k": 20,
11
  "top_p": 0.95,
12
+ "transformers_version": "4.57.1"
13
  }
special_tokens_map.json CHANGED
@@ -14,18 +14,6 @@
14
  "<|image_pad|>",
15
  "<|video_pad|>"
16
  ],
17
- "eos_token": {
18
- "content": "<|im_end|>",
19
- "lstrip": false,
20
- "normalized": false,
21
- "rstrip": false,
22
- "single_word": false
23
- },
24
- "pad_token": {
25
- "content": "<|endoftext|>",
26
- "lstrip": false,
27
- "normalized": false,
28
- "rstrip": false,
29
- "single_word": false
30
- }
31
  }
 
14
  "<|image_pad|>",
15
  "<|video_pad|>"
16
  ],
17
+ "eos_token": "<|endoftext|>",
18
+ "pad_token": "<|endoftext|>"
 
 
 
 
 
 
 
 
 
 
 
 
19
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
- size 11422654
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9417dfa2470f086897a0fa5acf4c11e1b05646717bdd7f9d4dc119332c65d421
3
+ size 11422919
tokenizer_config.json CHANGED
@@ -228,11 +228,12 @@
228
  ],
229
  "bos_token": null,
230
  "clean_up_tokenization_spaces": false,
231
- "eos_token": "<|im_end|>",
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
  "model_max_length": 1010000,
235
  "pad_token": "<|endoftext|>",
 
236
  "split_special_tokens": false,
237
  "tokenizer_class": "Qwen2Tokenizer",
238
  "unk_token": null
 
228
  ],
229
  "bos_token": null,
230
  "clean_up_tokenization_spaces": false,
231
+ "eos_token": "<|endoftext|>",
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
  "model_max_length": 1010000,
235
  "pad_token": "<|endoftext|>",
236
+ "padding_side": "left",
237
  "split_special_tokens": false,
238
  "tokenizer_class": "Qwen2Tokenizer",
239
  "unk_token": null