Upload optimized ONNX model

#1
by Xenova HF Staff - opened
.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ onnx/model.onnx_data filter=lfs diff=lfs merge=lfs -text
37
+ onnx/model_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
38
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "EuroBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token": "<|begin_of_text|>",
8
+ "bos_token_id": 128000,
9
+ "classifier_pooling": "late",
10
+ "dtype": "float32",
11
+ "eos_token": "<|end_of_text|>",
12
+ "eos_token_id": 128001,
13
+ "head_dim": 64,
14
+ "hidden_act": "silu",
15
+ "hidden_dropout": 0.0,
16
+ "hidden_size": 32,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 64,
19
+ "is_causal": false,
20
+ "is_decoder": false,
21
+ "mask_token": "<|mask|>",
22
+ "mask_token_id": 128002,
23
+ "max_position_embeddings": 8192,
24
+ "mlp_bias": false,
25
+ "model_type": "eurobert",
26
+ "num_attention_heads": 2,
27
+ "num_hidden_layers": 4,
28
+ "num_key_value_heads": 2,
29
+ "pad_token": "<|end_of_text|>",
30
+ "pad_token_id": 128001,
31
+ "pretraining_tp": 1,
32
+ "rms_norm_eps": 1e-05,
33
+ "rope_parameters": {
34
+ "rope_theta": 250000,
35
+ "rope_type": "default"
36
+ },
37
+ "tie_word_embeddings": false,
38
+ "transformers_version": "5.3.0.dev0",
39
+ "use_cache": false,
40
+ "vocab_size": 128256,
41
+ "transformers.js_config": {
42
+ "use_external_data_format": {
43
+ "model.onnx": 1,
44
+ "model_fp16.onnx": 1
45
+ }
46
+ }
47
+ }
onnx/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af5f6f633eb6816892c78074279c1b70d049f88335ffaed44389033179299383
3
+ size 33962
onnx/model.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16b9b860541ce3b4f6ae30b6ea9eebbd3e49ec42e27f2c866622576eeb36c0fb
3
+ size 18874368
onnx/model_fp16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2345d45ba058a3f9c8110ec578abf419894e05222c777404333f3f5144760148
3
+ size 33632
onnx/model_fp16.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3cf5fb0e3ba472897a40f5dbbca8c5f99174168d65249068650ddcb205d9bd8
3
+ size 9437184
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1df814fd3a872f1189d5347e5eb190d570397f4b798f258fa1d41493d77b271a
3
+ size 11574811
tokenizer_config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "bos_token": "<|begin_of_text|>",
4
+ "clean_up_tokenization_spaces": true,
5
+ "eos_token": "<|end_of_text|>",
6
+ "is_local": false,
7
+ "mask_token": "<|mask|>",
8
+ "max_length": null,
9
+ "model_input_names": [
10
+ "input_ids",
11
+ "attention_mask"
12
+ ],
13
+ "model_max_length": 1000000000000000019884624838656,
14
+ "pad_to_multiple_of": null,
15
+ "pad_token": "<|pad|>",
16
+ "pad_token_type_id": 0,
17
+ "padding_side": "right",
18
+ "tokenizer_class": "TokenizersBackend"
19
+ }