jeffzyu commited on
Commit
296b82c
·
verified ·
1 Parent(s): edcc0fc

Add files using upload-large-folder tool

Browse files
Files changed (47) hide show
  1. ._chat_template.jinja +0 -0
  2. ._config.json +0 -0
  3. ._generation_config.json +0 -0
  4. ._model-00035-of-00050.safetensors +3 -0
  5. ._model-00036-of-00050.safetensors +3 -0
  6. ._model-00037-of-00050.safetensors +3 -0
  7. ._model-00038-of-00050.safetensors +3 -0
  8. ._model-00039-of-00050.safetensors +3 -0
  9. ._model-00040-of-00050.safetensors +3 -0
  10. ._model-00041-of-00050.safetensors +3 -0
  11. ._model-00042-of-00050.safetensors +3 -0
  12. ._model-00043-of-00050.safetensors +3 -0
  13. ._model-00044-of-00050.safetensors +3 -0
  14. ._model-00045-of-00050.safetensors +3 -0
  15. ._model-00046-of-00050.safetensors +3 -0
  16. ._model-00047-of-00050.safetensors +3 -0
  17. ._model-00048-of-00050.safetensors +3 -0
  18. ._model-00049-of-00050.safetensors +3 -0
  19. ._model-00050-of-00050.safetensors +3 -0
  20. ._model.safetensors.index.json +0 -0
  21. ._preprocessor_config.json +0 -0
  22. ._processor_config.json +0 -0
  23. ._special_tokens_map.json +0 -0
  24. ._tokenizer.json +0 -0
  25. ._tokenizer_config.json +0 -0
  26. .cache/._huggingface +0 -0
  27. config.json +227 -205
  28. generation_config.json +13 -0
  29. model-00001-of-00050.safetensors +3 -0
  30. model-00002-of-00050.safetensors +3 -0
  31. model-00006-of-00050.safetensors +3 -0
  32. model-00039-of-00050.safetensors +3 -0
  33. model-00040-of-00050.safetensors +3 -0
  34. model-00041-of-00050.safetensors +3 -0
  35. model-00042-of-00050.safetensors +3 -0
  36. model-00043-of-00050.safetensors +3 -0
  37. model-00044-of-00050.safetensors +3 -0
  38. model-00045-of-00050.safetensors +3 -0
  39. model-00046-of-00050.safetensors +3 -0
  40. model-00047-of-00050.safetensors +3 -0
  41. model-00048-of-00050.safetensors +3 -0
  42. model-00049-of-00050.safetensors +3 -0
  43. model-00050-of-00050.safetensors +3 -0
  44. model.safetensors.index.json +0 -0
  45. preprocessor_config.json +36 -0
  46. processor_config.json +6 -0
  47. tokenizer_config.json +0 -1
._chat_template.jinja ADDED
Binary file (4.1 kB). View file
 
._config.json ADDED
Binary file (4.1 kB). View file
 
._generation_config.json ADDED
Binary file (4.1 kB). View file
 
._model-00035-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00036-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00037-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00038-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00039-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00040-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00041-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00042-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00043-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00044-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00045-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00046-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00047-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00048-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00049-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model-00050-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0e8836dc636a4bef230020c6ffe89f1e343d1c94b8e53001e72b302747c603
3
+ size 4096
._model.safetensors.index.json ADDED
Binary file (4.1 kB). View file
 
._preprocessor_config.json ADDED
Binary file (4.1 kB). View file
 
._processor_config.json ADDED
Binary file (4.1 kB). View file
 
._special_tokens_map.json ADDED
Binary file (4.1 kB). View file
 
._tokenizer.json ADDED
Binary file (4.1 kB). View file
 
._tokenizer_config.json ADDED
Binary file (4.1 kB). View file
 
.cache/._huggingface ADDED
Binary file (4.1 kB). View file
 
config.json CHANGED
@@ -1,214 +1,236 @@
1
  {
2
  "architectures": [
3
- "Llama4ForCausalLM"
4
  ],
5
- "attention_bias": false,
6
- "attention_chunk_size": 8192,
7
- "attention_dropout": 0.0,
8
- "attn_scale": 0.1,
9
- "attn_temperature_tuning": 4,
10
- "bos_token_id": 200000,
11
- "cache_implementation": "hybrid",
12
  "dtype": "bfloat16",
13
- "eos_token_id": 200008,
14
- "floor_scale": 8192,
15
- "for_llm_compressor": false,
16
- "head_dim": 128,
17
- "hidden_act": "silu",
18
- "hidden_size": 5120,
19
- "initializer_range": 0.02,
20
- "interleave_moe_layer_step": 1,
21
- "intermediate_size": 8192,
22
- "intermediate_size_mlp": 16384,
23
- "layer_types": [
24
- "chunked_attention",
25
- "chunked_attention",
26
- "chunked_attention",
27
- "full_attention",
28
- "chunked_attention",
29
- "chunked_attention",
30
- "chunked_attention",
31
- "full_attention",
32
- "chunked_attention",
33
- "chunked_attention",
34
- "chunked_attention",
35
- "full_attention",
36
- "chunked_attention",
37
- "chunked_attention",
38
- "chunked_attention",
39
- "full_attention",
40
- "chunked_attention",
41
- "chunked_attention",
42
- "chunked_attention",
43
- "full_attention",
44
- "chunked_attention",
45
- "chunked_attention",
46
- "chunked_attention",
47
- "full_attention",
48
- "chunked_attention",
49
- "chunked_attention",
50
- "chunked_attention",
51
- "full_attention",
52
- "chunked_attention",
53
- "chunked_attention",
54
- "chunked_attention",
55
- "full_attention",
56
- "chunked_attention",
57
- "chunked_attention",
58
- "chunked_attention",
59
- "full_attention",
60
- "chunked_attention",
61
- "chunked_attention",
62
- "chunked_attention",
63
- "full_attention",
64
- "chunked_attention",
65
- "chunked_attention",
66
- "chunked_attention",
67
- "full_attention",
68
- "chunked_attention",
69
- "chunked_attention",
70
- "chunked_attention",
71
- "full_attention"
72
- ],
73
- "max_position_embeddings": 10485760,
74
- "model_type": "llama4_text",
75
- "moe_layers": [
76
- 0,
77
- 1,
78
- 2,
79
- 3,
80
- 4,
81
- 5,
82
- 6,
83
- 7,
84
- 8,
85
- 9,
86
- 10,
87
- 11,
88
- 12,
89
- 13,
90
- 14,
91
- 15,
92
- 16,
93
- 17,
94
- 18,
95
- 19,
96
- 20,
97
- 21,
98
- 22,
99
- 23,
100
- 24,
101
- 25,
102
- 26,
103
- 27,
104
- 28,
105
- 29,
106
- 30,
107
- 31,
108
- 32,
109
- 33,
110
- 34,
111
- 35,
112
- 36,
113
- 37,
114
- 38,
115
- 39,
116
- 40,
117
- 41,
118
- 42,
119
- 43,
120
- 44,
121
- 45,
122
- 46,
123
- 47
124
- ],
125
- "no_rope_layers": [
126
- 1,
127
- 1,
128
- 1,
129
- 0,
130
- 1,
131
- 1,
132
- 1,
133
- 0,
134
- 1,
135
- 1,
136
- 1,
137
- 0,
138
- 1,
139
- 1,
140
- 1,
141
- 0,
142
- 1,
143
- 1,
144
- 1,
145
- 0,
146
- 1,
147
- 1,
148
- 1,
149
- 0,
150
- 1,
151
- 1,
152
- 1,
153
- 0,
154
- 1,
155
- 1,
156
- 1,
157
- 0,
158
- 1,
159
- 1,
160
- 1,
161
- 0,
162
- 1,
163
- 1,
164
- 1,
165
- 0,
166
- 1,
167
- 1,
168
- 1,
169
- 0,
170
- 1,
171
- 1,
172
- 1,
173
- 0
174
- ],
175
- "num_attention_heads": 40,
176
- "num_experts_per_tok": 1,
177
- "num_hidden_layers": 48,
178
- "num_key_value_heads": 8,
179
- "num_local_experts": 16,
180
- "output_router_logits": false,
181
- "pad_token_id": 200018,
182
- "quantization_config": {
183
- "_load_in_4bit": true,
184
- "_load_in_8bit": false,
185
- "bnb_4bit_compute_dtype": "bfloat16",
186
- "bnb_4bit_quant_storage": "bfloat16",
187
- "bnb_4bit_quant_type": "nf4",
188
- "bnb_4bit_use_double_quant": false,
189
- "llm_int8_enable_fp32_cpu_offload": false,
190
- "llm_int8_has_fp16_weight": false,
191
- "llm_int8_skip_modules": null,
192
- "llm_int8_threshold": 6.0,
193
- "load_in_4bit": true,
194
- "load_in_8bit": false,
195
- "quant_method": "bitsandbytes"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
196
  },
197
- "rms_norm_eps": 1e-05,
198
- "rope_scaling": {
199
- "factor": 16.0,
200
- "high_freq_factor": 1.0,
201
- "low_freq_factor": 1.0,
202
- "original_max_position_embeddings": 8192,
203
- "rope_type": "llama3"
204
- },
205
- "rope_theta": 500000.0,
206
- "router_aux_loss_coef": 0.001,
207
- "router_jitter_noise": 0.0,
208
  "tie_word_embeddings": false,
209
  "transformers_version": "4.57.0",
210
- "use_cache": false,
211
- "use_qk_norm": true,
212
- "vocab_size": 202048,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
213
  "torch_dtype": "bfloat16"
214
  }
 
1
  {
2
  "architectures": [
3
+ "Llama4ForConditionalGeneration"
4
  ],
5
+ "boi_token_index": 200080,
 
 
 
 
 
 
6
  "dtype": "bfloat16",
7
+ "eoi_token_index": 200081,
8
+ "image_token_index": 200092,
9
+ "model_type": "llama4",
10
+ "text_config": {
11
+ "_attn_implementation_autoset": true,
12
+ "attention_bias": false,
13
+ "attention_chunk_size": 8192,
14
+ "attention_dropout": 0.0,
15
+ "attn_scale": 0.1,
16
+ "attn_temperature_tuning": true,
17
+ "bos_token_id": 200000,
18
+ "dtype": "bfloat16",
19
+ "eos_token_id": [
20
+ 200001,
21
+ 200007,
22
+ 200008
23
+ ],
24
+ "floor_scale": 8192,
25
+ "for_llm_compressor": false,
26
+ "head_dim": 128,
27
+ "hidden_act": "silu",
28
+ "hidden_size": 5120,
29
+ "initializer_range": 0.02,
30
+ "interleave_moe_layer_step": 1,
31
+ "intermediate_size": 8192,
32
+ "intermediate_size_mlp": 16384,
33
+ "layer_types": [
34
+ "chunked_attention",
35
+ "chunked_attention",
36
+ "chunked_attention",
37
+ "full_attention",
38
+ "chunked_attention",
39
+ "chunked_attention",
40
+ "chunked_attention",
41
+ "full_attention",
42
+ "chunked_attention",
43
+ "chunked_attention",
44
+ "chunked_attention",
45
+ "full_attention",
46
+ "chunked_attention",
47
+ "chunked_attention",
48
+ "chunked_attention",
49
+ "full_attention",
50
+ "chunked_attention",
51
+ "chunked_attention",
52
+ "chunked_attention",
53
+ "full_attention",
54
+ "chunked_attention",
55
+ "chunked_attention",
56
+ "chunked_attention",
57
+ "full_attention",
58
+ "chunked_attention",
59
+ "chunked_attention",
60
+ "chunked_attention",
61
+ "full_attention",
62
+ "chunked_attention",
63
+ "chunked_attention",
64
+ "chunked_attention",
65
+ "full_attention",
66
+ "chunked_attention",
67
+ "chunked_attention",
68
+ "chunked_attention",
69
+ "full_attention",
70
+ "chunked_attention",
71
+ "chunked_attention",
72
+ "chunked_attention",
73
+ "full_attention",
74
+ "chunked_attention",
75
+ "chunked_attention",
76
+ "chunked_attention",
77
+ "full_attention",
78
+ "chunked_attention",
79
+ "chunked_attention",
80
+ "chunked_attention",
81
+ "full_attention"
82
+ ],
83
+ "max_position_embeddings": 10485760,
84
+ "model_type": "llama4_text",
85
+ "moe_layers": [
86
+ 0,
87
+ 1,
88
+ 2,
89
+ 3,
90
+ 4,
91
+ 5,
92
+ 6,
93
+ 7,
94
+ 8,
95
+ 9,
96
+ 10,
97
+ 11,
98
+ 12,
99
+ 13,
100
+ 14,
101
+ 15,
102
+ 16,
103
+ 17,
104
+ 18,
105
+ 19,
106
+ 20,
107
+ 21,
108
+ 22,
109
+ 23,
110
+ 24,
111
+ 25,
112
+ 26,
113
+ 27,
114
+ 28,
115
+ 29,
116
+ 30,
117
+ 31,
118
+ 32,
119
+ 33,
120
+ 34,
121
+ 35,
122
+ 36,
123
+ 37,
124
+ 38,
125
+ 39,
126
+ 40,
127
+ 41,
128
+ 42,
129
+ 43,
130
+ 44,
131
+ 45,
132
+ 46,
133
+ 47
134
+ ],
135
+ "no_rope_layers": [
136
+ 1,
137
+ 1,
138
+ 1,
139
+ 0,
140
+ 1,
141
+ 1,
142
+ 1,
143
+ 0,
144
+ 1,
145
+ 1,
146
+ 1,
147
+ 0,
148
+ 1,
149
+ 1,
150
+ 1,
151
+ 0,
152
+ 1,
153
+ 1,
154
+ 1,
155
+ 0,
156
+ 1,
157
+ 1,
158
+ 1,
159
+ 0,
160
+ 1,
161
+ 1,
162
+ 1,
163
+ 0,
164
+ 1,
165
+ 1,
166
+ 1,
167
+ 0,
168
+ 1,
169
+ 1,
170
+ 1,
171
+ 0,
172
+ 1,
173
+ 1,
174
+ 1,
175
+ 0,
176
+ 1,
177
+ 1,
178
+ 1,
179
+ 0,
180
+ 1,
181
+ 1,
182
+ 1,
183
+ 0
184
+ ],
185
+ "num_attention_heads": 40,
186
+ "num_experts_per_tok": 1,
187
+ "num_hidden_layers": 48,
188
+ "num_key_value_heads": 8,
189
+ "num_local_experts": 16,
190
+ "output_router_logits": false,
191
+ "pad_token_id": 200018,
192
+ "rms_norm_eps": 1e-05,
193
+ "rope_scaling": {
194
+ "factor": 16.0,
195
+ "high_freq_factor": 1.0,
196
+ "low_freq_factor": 1.0,
197
+ "original_max_position_embeddings": 8192,
198
+ "rope_type": "llama3"
199
+ },
200
+ "rope_theta": 500000.0,
201
+ "router_aux_loss_coef": 0.001,
202
+ "router_jitter_noise": 0.0,
203
+ "use_cache": true,
204
+ "use_qk_norm": true,
205
+ "vocab_size": 202048,
206
+ "torch_dtype": "bfloat16"
207
  },
 
 
 
 
 
 
 
 
 
 
 
208
  "tie_word_embeddings": false,
209
  "transformers_version": "4.57.0",
210
+ "vision_config": {
211
+ "_attn_implementation_autoset": true,
212
+ "_vision_feature_layer": -1,
213
+ "attention_dropout": 0.0,
214
+ "hidden_act": "gelu",
215
+ "hidden_size": 1408,
216
+ "image_size": 336,
217
+ "initializer_range": 0.02,
218
+ "intermediate_size": 5632,
219
+ "model_type": "llama4_vision_model",
220
+ "multi_modal_projector_bias": false,
221
+ "norm_eps": 1e-05,
222
+ "num_attention_heads": 16,
223
+ "num_channels": 3,
224
+ "num_hidden_layers": 34,
225
+ "patch_size": 14,
226
+ "pixel_shuffle_ratio": 0.5,
227
+ "projector_dropout": 0.0,
228
+ "projector_input_dim": 4096,
229
+ "projector_output_dim": 4096,
230
+ "rope_theta": 10000,
231
+ "vision_feature_layer": -1,
232
+ "vision_feature_select_strategy": "default",
233
+ "vision_output_dim": 4096
234
+ },
235
  "torch_dtype": "bfloat16"
236
  }
generation_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 200000,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 200001,
6
+ 200007,
7
+ 200008
8
+ ],
9
+ "pad_token_id": 200018,
10
+ "temperature": 0.6,
11
+ "top_p": 0.9,
12
+ "transformers_version": "4.57.0"
13
+ }
model-00001-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12958641d3be49086efbd1c40d6cae79187eb63ffb6e1042baed2cb6331a8f2f
3
+ size 3938735392
model-00002-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5328ccd6704dfa5dfe7c2b32df8863a0f915ca3c55cd9ec5fa6bfd4645ba94ce
3
+ size 4404205216
model-00006-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e4924af687b2e08ec7ca19dfda39b6070089de3d0307a943d265e755bb27a4d
3
+ size 4404205216
model-00039-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:172965fa885a8fe67f3cfdcf88933aab8ea7d2c23846f5cd57eb7a7b63c087ad
3
+ size 4404205232
model-00040-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:681727fb33f8c614cc1d45b80186adda279f57e40a4b0ea2af07210150838254
3
+ size 4404205232
model-00041-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8a2de9f4d2c7c2d3fc95de0ab9575759f45544d1e3bf5c9cbb5e45af57f2e5d
3
+ size 4404205232
model-00042-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9fac37c0a86827f8b0b9fe741e530fe06a072eba3a517369c1dcbc409131255
3
+ size 4404205232
model-00043-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e496b89b05169f874b50514238d015ba265afbb690302cd93505d245e749709
3
+ size 4404205232
model-00044-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28aa132bb5bfb77ab02fb679c59a3f1fb975381b37cededa0e57207507289296
3
+ size 4404205232
model-00045-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14b360a83fbaf5a925451de903ec8a58ecfde9588a410efa4af1e1782569ced8
3
+ size 4404205232
model-00046-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21c4656bd266e9645520f11d31bb8be5b644f6c1b80af34f29616a534a81e0e0
3
+ size 4404205232
model-00047-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fb8a6e29c5e734cf1804f5a20ee44d987f6bf611d80e54ad57c0cd20c9142c3
3
+ size 4404205232
model-00048-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb4a1844d4aa52b24e00764370f3e6aea4528de42006c208613db809609899c4
3
+ size 4404205232
model-00049-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5271c1f3151827d09436bb51b5661725103356d639781c6214690a1248bc0961
3
+ size 4278385928
model-00050-of-00050.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f412ae287747cc86e6ee5d58600bbe5c574d6fa77f351367aae9ea3f4dbcc3a
3
+ size 2068971664
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
preprocessor_config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": null,
3
+ "data_format": "channels_first",
4
+ "default_to_square": true,
5
+ "device": null,
6
+ "disable_grouping": null,
7
+ "do_center_crop": null,
8
+ "do_convert_rgb": true,
9
+ "do_normalize": true,
10
+ "do_pad": null,
11
+ "do_rescale": true,
12
+ "do_resize": true,
13
+ "image_mean": [
14
+ 0.5,
15
+ 0.5,
16
+ 0.5
17
+ ],
18
+ "image_processor_type": "Llama4ImageProcessorFast",
19
+ "image_std": [
20
+ 0.5,
21
+ 0.5,
22
+ 0.5
23
+ ],
24
+ "input_data_format": null,
25
+ "max_patches": 16,
26
+ "pad_size": null,
27
+ "processor_class": "Llama4Processor",
28
+ "resample": 2,
29
+ "rescale_factor": 0.00392156862745098,
30
+ "resize_to_max_canvas": false,
31
+ "return_tensors": null,
32
+ "size": {
33
+ "height": 336,
34
+ "width": 336
35
+ }
36
+ }
processor_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "fake_image_token": "<|image|>",
3
+ "image_token": "<|image|>",
4
+ "patch_size": 14,
5
+ "processor_class": "Llama4Processor"
6
+ }
tokenizer_config.json CHANGED
@@ -9091,7 +9091,6 @@
9091
  ],
9092
  "model_max_length": 10485760,
9093
  "pad_token": "<|finetune_right_pad|>",
9094
- "padding_side": "right",
9095
  "processor_class": "Llama4Processor",
9096
  "tokenizer_class": "PreTrainedTokenizerFast"
9097
  }
 
9091
  ],
9092
  "model_max_length": 10485760,
9093
  "pad_token": "<|finetune_right_pad|>",
 
9094
  "processor_class": "Llama4Processor",
9095
  "tokenizer_class": "PreTrainedTokenizerFast"
9096
  }