keko24 commited on
Commit
e0a33f7
·
verified ·
1 Parent(s): a0c1986

Upload Qwen3ForCausalLM

Browse files
Files changed (2) hide show
  1. config.json +5 -4
  2. model.safetensors +2 -2
config.json CHANGED
@@ -19,7 +19,7 @@
19
  "num_key_value_heads": 8,
20
  "quantization_config": {
21
  "config_groups": {
22
- "group_1": {
23
  "input_activations": {
24
  "actorder": null,
25
  "block_structure": null,
@@ -50,14 +50,15 @@
50
  }
51
  }
52
  },
53
- "format": "pack-quantized",
54
  "global_compression_ratio": null,
55
  "ignore": [
56
  "lm_head"
57
  ],
58
  "kv_cache_scheme": null,
59
  "quant_method": "compressed-tensors",
60
- "quantization_status": "compressed"
 
61
  },
62
  "rms_norm_eps": 1e-06,
63
  "rope_scaling": null,
@@ -69,4 +70,4 @@
69
  "use_cache": true,
70
  "use_sliding_window": false,
71
  "vocab_size": 151936
72
- }
 
19
  "num_key_value_heads": 8,
20
  "quantization_config": {
21
  "config_groups": {
22
+ "group_0": {
23
  "input_activations": {
24
  "actorder": null,
25
  "block_structure": null,
 
50
  }
51
  }
52
  },
53
+ "format": "int-quantized",
54
  "global_compression_ratio": null,
55
  "ignore": [
56
  "lm_head"
57
  ],
58
  "kv_cache_scheme": null,
59
  "quant_method": "compressed-tensors",
60
+ "quantization_status": "compressed",
61
+ "sparsity_config": {}
62
  },
63
  "rms_norm_eps": 1e-06,
64
  "rope_scaling": null,
 
70
  "use_cache": true,
71
  "use_sliding_window": false,
72
  "vocab_size": 151936
73
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9dbfdc6bcd07183093c48cf1b47eca099b99e534d0e147e0ee656278993d95e7
3
- size 762101160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df89ca56d8fa8a03a708fc1e9ebab6d4061561e5e00b21e42a3f323e7c209d9c
3
+ size 758636928