SreyanG-NVIDIA commited on
Commit
155a1a1
·
verified ·
1 Parent(s): 8e288bc

Upload MusicFlamingoForConditionalGeneration

Browse files
Files changed (3) hide show
  1. config.json +83 -0
  2. generation_config.json +7 -0
  3. model.safetensors +3 -0
config.json ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MusicFlamingoForConditionalGeneration"
4
+ ],
5
+ "audio_config": {
6
+ "activation_dropout": 0.0,
7
+ "activation_function": "gelu",
8
+ "attention_dropout": 0.0,
9
+ "dropout": 0.0,
10
+ "hidden_size": 1280,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 5120,
13
+ "layerdrop": 0.0,
14
+ "max_source_positions": 1500,
15
+ "model_type": "musicflamingo_encoder",
16
+ "num_attention_heads": 20,
17
+ "num_hidden_layers": 32,
18
+ "num_mel_bins": 128,
19
+ "scale_embedding": false
20
+ },
21
+ "audio_token_id": 151667,
22
+ "dtype": "bfloat16",
23
+ "model_type": "musicflamingo",
24
+ "projector_bias": true,
25
+ "projector_hidden_act": "gelu",
26
+ "text_config": {
27
+ "attention_dropout": 0.0,
28
+ "bos_token_id": 151668,
29
+ "eos_token_id": 151645,
30
+ "hidden_act": "silu",
31
+ "hidden_size": 3584,
32
+ "initializer_range": 0.02,
33
+ "intermediate_size": 18944,
34
+ "layer_types": [
35
+ "full_attention",
36
+ "full_attention",
37
+ "full_attention",
38
+ "full_attention",
39
+ "full_attention",
40
+ "full_attention",
41
+ "full_attention",
42
+ "full_attention",
43
+ "full_attention",
44
+ "full_attention",
45
+ "full_attention",
46
+ "full_attention",
47
+ "full_attention",
48
+ "full_attention",
49
+ "full_attention",
50
+ "full_attention",
51
+ "full_attention",
52
+ "full_attention",
53
+ "full_attention",
54
+ "full_attention",
55
+ "full_attention",
56
+ "full_attention",
57
+ "full_attention",
58
+ "full_attention",
59
+ "full_attention",
60
+ "full_attention",
61
+ "full_attention",
62
+ "full_attention"
63
+ ],
64
+ "max_position_embeddings": 32768,
65
+ "max_window_layers": 28,
66
+ "model_max_length": 8192,
67
+ "model_type": "qwen2",
68
+ "num_attention_heads": 28,
69
+ "num_hidden_layers": 28,
70
+ "num_key_value_heads": 4,
71
+ "pad_token_id": 151669,
72
+ "rms_norm_eps": 1e-06,
73
+ "rope_parameters": {
74
+ "rope_theta": 1000000.0,
75
+ "rope_type": "default"
76
+ },
77
+ "sliding_window": null,
78
+ "use_cache": false,
79
+ "use_sliding_window": false,
80
+ "vocab_size": 151672
81
+ },
82
+ "transformers_version": "5.0.0.dev0"
83
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151668,
3
+ "eos_token_id": 151645,
4
+ "max_new_tokens": 2048,
5
+ "pad_token_id": 151669,
6
+ "transformers_version": "5.0.0.dev0"
7
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33ed1c5b06a8789dec6e2301f65f356e14995d02fdd3bf61856422e44fe205db
3
+ size 16534531864