Mingke977 commited on
Commit
facd22e
·
verified ·
1 Parent(s): bb94af7

Add files using upload-large-folder tool

Browse files
README.md CHANGED
@@ -1,378 +1,49 @@
1
  ---
2
- language:
3
- - zh
4
- - en
5
- pipeline_tag: text-generation
6
  library_name: transformers
7
- ---
8
- <div align="center">
9
- <picture>
10
- <img src="figures/joyai-logo.png" width="30%" alt="JoyAI-LLM Flash">
11
- </picture>
12
- </div>
13
- <hr>
14
-
15
- <div align="center" style="line-height: 1;">
16
- <a href="https://huggingface.co/jdopensource" target="_blank"><img alt="Hugging Face" src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-JD-ffc107?color=ffc107&logoColor=white"/></a>
17
- <a href="https://huggingface.co/jdopensource/JoyAI-LLM-Flash/blob/main/LICENSE"><img alt="License" src="https://img.shields.io/badge/License-Modified_MIT-f5de53?&color=f5de53"/></a>
18
- </div>
19
-
20
-
21
-
22
-
23
- ## 1. Model Introduction
24
-
25
- JoyAI-LLM-Flash is a state-of-the-art medium-sized instruct language model with 3 billion activated parameters and 48 billion total parameters. JoyAI-LLM-Flash was pretrained on 20 trillion text tokens using Muon optimizer, followed by large-scale supervised fine-tuning (SFT), direct preference optimization (DPO), and reinforcement learning (RL) across diverse environments. JoyAI-LLM-Flash achieves strong performance across frontier knowledge, reasoning, coding tasks and agentic capabilities.
26
-
27
- ### Key Features
28
-
29
- - Fiber Bundle RL: Introduces fiber bundle theory into reinforcement learning, proposing a novel optimization framework, FiberPO. This method is specifically designed to handle the challenges of large-scale and heterogeneous agent training, improving stability and robustness under complex data distributions.
30
- - Training-Inference Collaboration: apply Muon optimizer with dense MTP, develop novel optimization techniques to resolve instabilities while scaling up, delivering 1.3× to 1.7× the throughput of the non-MTP version.
31
- - Agentic Intelligence: designed for tool use, reasoning, and autonomous problem-solving.
32
-
33
- ## 2. Model Summary
34
-
35
- | | |
36
- | :-----------------------------------------: | :----------------------: |
37
- | **Architecture** | Mixture-of-Experts (MoE) |
38
- | **Total Parameters** | 48B |
39
- | **Activated Parameters** | 3B |
40
- | **Number of Layers** (Dense layer included) | 40 |
41
- | **Number of Dense Layers** | 1 |
42
- | **Attention Hidden Dimension** | 2048 |
43
- | **MoE Hidden Dimension** (per Expert) | 768 |
44
- | **Number of Attention Heads** | 32 |
45
- | **Number of Experts** | 256 |
46
- | **Selected Experts per Token** | 8 |
47
- | **Number of Shared Experts** | 1 |
48
- | **Vocabulary Size** | 129K |
49
- | **Context Length** | 128K |
50
- | **Attention Mechanism** | MLA |
51
- | **Activation Function** | SwiGLU |
52
- | </div> | |
53
-
54
-
55
- ## 3. Evaluation Results
56
-
57
- <table>
58
- <thead>
59
- <tr>
60
- <th align="center">Benchmark</th>
61
- <th align="center"><sup>JoyAI-LLM Flash</sup></th>
62
- <th align="center"><sup>Qwen3-30B-A3B-Instuct-2507</sup></th>
63
- <th align="center"><sup>GLM-4.7-Flash<br>(Non-thinking)</sup></th>
64
- </tr>
65
- </thead>
66
- <tbody>
67
-
68
-
69
- <tr>
70
- <td align="center" colspan=8><strong>Knowledge &amp; Alignment</strong></td>
71
- </tr>
72
- <tr>
73
- <td align="center" style="vertical-align: middle">MMLU</td>
74
- <td align="center" style="vertical-align: middle"><strong>89.50</strong></td>
75
- <td align="center" style="vertical-align: middle">86.87</td>
76
- <td align="center" style="vertical-align: middle">80.53</td>
77
- </tr>
78
- <tr>
79
- <td align="center" style="vertical-align: middle">MMLU-Pro</td>
80
- <td align="center" style="vertical-align: middle"><strong>81.02</strong></td>
81
- <td align="center" style="vertical-align: middle">73.88</td>
82
- <td align="center" style="vertical-align: middle">63.62</td>
83
- </tr>
84
- <tr>
85
- <td align="center" style="vertical-align: middle">CMMLU</td>
86
- <td align="center" style="vertical-align: middle"><strong>87.03</strong></td>
87
- <td align="center" style="vertical-align: middle">85.88</td>
88
- <td align="center" style="vertical-align: middle">75.85</td>
89
- </tr>
90
- <tr>
91
- <td align="center" style="vertical-align: middle">GPQA-Diamond</td>
92
- <td align="center" style="vertical-align: middle"><strong>74.43</strong></td>
93
- <td align="center" style="vertical-align: middle">68.69</td>
94
- <td align="center" style="vertical-align: middle">39.90</td>
95
- </tr>
96
- <tr>
97
- <td align="center" style="vertical-align: middle">SuperGPQA</td>
98
- <td align="center" style="vertical-align: middle"><strong>55.00</strong></td>
99
- <td align="center" style="vertical-align: middle">52.00</td>
100
- <td align="center" style="vertical-align: middle">32.00</td>
101
- </tr>
102
- <tr>
103
- <td align="center" style="vertical-align: middle">LiveBench</td>
104
- <td align="center" style="vertical-align: middle"><strong>72.90</strong></td>
105
- <td align="center" style="vertical-align: middle">59.70</td>
106
- <td align="center" style="vertical-align: middle">43.10</td>
107
- </tr>
108
- <tr>
109
- <td align="center" style="vertical-align: middle">IFEval</td>
110
- <td align="center" style="vertical-align: middle"><strong>86.69</strong></td>
111
- <td align="center" style="vertical-align: middle">83.18</td>
112
- <td align="center" style="vertical-align: middle">82.44</td>
113
- </tr>
114
- <tr>
115
- <td align="center" style="vertical-align: middle">AlignBench</td>
116
- <td align="center" style="vertical-align: middle"><strong>8.24</strong></td>
117
- <td align="center" style="vertical-align: middle">8.07</td>
118
- <td align="center" style="vertical-align: middle">6.85</td>
119
- </tr>
120
- <tr>
121
- <td align="center" style="vertical-align: middle">HellaSwag</td>
122
- <td align="center" style="vertical-align: middle"><strong>91.79</strong></td>
123
- <td align="center" style="vertical-align: middle">89.90</td>
124
- <td align="center" style="vertical-align: middle">60.84</td>
125
- </tr>
126
-
127
- <tr>
128
- <td align="center" colspan=8><strong>Coding</strong></td>
129
- </tr>
130
- <tr>
131
- <td align="center" style="vertical-align: middle">HumanEval</td>
132
- <td align="center" style="vertical-align: middle"><strong>96.34</strong></td>
133
- <td align="center" style="vertical-align: middle">95.12</td>
134
- <td align="center" style="vertical-align: middle">74.39</td>
135
- </tr>
136
- <tr>
137
- <td align="center" style="vertical-align: middle">LiveCodeBench</td>
138
- <td align="center" style="vertical-align: middle"><strong>65.60</strong></td>
139
- <td align="center" style="vertical-align: middle">39.71</td>
140
- <td align="center" style="vertical-align: middle">27.43</td>
141
- </tr>
142
- <tr>
143
- <td align="center" style="vertical-align: middle">SciCode</td>
144
- <td align="center" style="vertical-align: middle"><strong>3.08/22.92</strong></td>
145
- <td align="center" style="vertical-align: middle"><strong>3.08/22.92</strong></td>
146
- <td align="center" style="vertical-align: middle">3.08/15.11</td>
147
- </tr>
148
- <tr>
149
- <td align="center" colspan=8><strong>Mathematics</strong></td>
150
- </tr>
151
- <tr>
152
- <td align="center" style="vertical-align: middle">GSM8K</td>
153
- <td align="center" style="vertical-align: middle"><strong>95.83</strong></td>
154
- <td align="center" style="vertical-align: middle">79.83</td>
155
- <td align="center" style="vertical-align: middle">81.88</td>
156
- </tr>
157
- <tr>
158
- <td align="center" style="vertical-align: middle">AIME2025</td>
159
- <td align="center" style="vertical-align: middle"><strong>65.83</strong></td>
160
- <td align="center" style="vertical-align: middle">62.08</td>
161
- <td align="center" style="vertical-align: middle">24.17</td>
162
- </tr>
163
- <tr>
164
- <td align="center" style="vertical-align: middle">MATH 500</td>
165
- <td align="center" style="vertical-align: middle"><strong>97.10</strong></td>
166
- <td align="center" style="vertical-align: middle">89.80</td>
167
- <td align="center" style="vertical-align: middle">90.90</td>
168
- </tr>
169
-
170
- <tr>
171
- <td align="center" colspan=8><strong>Agentic</strong></td>
172
- </tr>
173
- <tr>
174
- <td align="center" style="vertical-align: middle">SWE-bench Verified</td>
175
- <td align="center" style="vertical-align: middle"><strong>60.60</strong></td>
176
- <td align="center" style="vertical-align: middle">24.44</td>
177
- <td align="center" style="vertical-align: middle">51.60</td>
178
- </tr>
179
- <tr>
180
- <td align="center" style="vertical-align: middle">Tau2-Retail</td>
181
- <td align="center" style="vertical-align: middle"><strong>67.55</strong></td>
182
- <td align="center" style="vertical-align: middle">53.51</td>
183
- <td align="center" style="vertical-align: middle">62.28</td>
184
- </tr>
185
- <tr>
186
- <td align="center" style="vertical-align: middle">Tau2-Airline</td>
187
- <td align="center" style="vertical-align: middle"><strong>54.00</strong></td>
188
- <td align="center" style="vertical-align: middle">32.00</td>
189
- <td align="center" style="vertical-align: middle">52.00</td>
190
- </tr>
191
- <tr>
192
- <td align="center" style="vertical-align: middle">Tau2-Telecom</td>
193
- <td align="center" style="vertical-align: middle">79.83</td>
194
- <td align="center" style="vertical-align: middle">4.39</td>
195
- <td align="center" style="vertical-align: middle"><strong>88.60</strong></td>
196
- </tr>
197
-
198
- <tr>
199
- <td align="center" colspan=8><strong>Long Context</strong></td>
200
- </tr>
201
- <tr>
202
- <td align="center" style="vertical-align: middle">RULER</td>
203
- <td align="center" style="vertical-align: middle"><strong>95.60</strong></td>
204
- <td align="center" style="vertical-align: middle">89.66</td>
205
- <td align="center" style="vertical-align: middle">56.12</td>
206
- </tr>
207
- </tbody>
208
- </table>
209
-
210
-
211
- ## 4. Deployment
212
-
213
- > [!Note]
214
- > You can access JoyAI-LLM Flash API on https://docs.jdcloud.com/cn/jdaip/chat and we provide OpenAI/Anthropic-compatible API for you.
215
- > Currently, JoyAI-LLM-Flash-Block-INT8 is recommended to run on the following inference engines:
216
-
217
- * SGLang
218
-
219
- Deployment examples can be found in the [Model Deployment Guide](docs/deploy_guidance.md).
220
-
221
-
222
-
223
- ## 5. Model Usage
224
-
225
- The usage demos below demonstrate how to call our official API.
226
-
227
- For third-party APIs deployed with vLLM or SGLang, please note that:
228
-
229
- > [!Note] Recommended sampling parameters: `temperature=0.6`, `top_p=1.0`
230
-
231
- ### Chat Completion
232
-
233
- This is a simple chat completion script which shows how to call JoyAI-Flash API.
234
-
235
- ```python
236
- from openai import OpenAI
237
-
238
- client = OpenAI(base_url="http://IP:PORT/v1", api_key="EMPTY")
239
-
240
-
241
- def simple_chat(client: OpenAI):
242
- messages = [
243
- {
244
- "role": "user",
245
- "content": [
246
- {
247
- "type": "text",
248
- "text": "which one is bigger, 9.11 or 9.9? think carefully.",
249
- }
250
- ],
251
- },
252
- ]
253
- model_name = client.models.list().data[0].id
254
- response = client.chat.completions.create(
255
- model=model_name, messages=messages, stream=False, max_tokens=4096
256
- )
257
- print(f"response: {response.choices[0].message.content}")
258
-
259
-
260
- if __name__ == "__main__":
261
- simple_chat(client)
262
- ```
263
-
264
-
265
- ### Tool call Completion
266
-
267
- This is a simple toll call completion script which shows how to call JoyAI-Flash API.
268
-
269
- ```python
270
- import json
271
-
272
- from openai import OpenAI
273
-
274
- client = OpenAI(base_url="http://IP:PORT/v1", api_key="EMPTY")
275
-
276
-
277
- def my_calculator(expression: str) -> str:
278
- return str(eval(expression))
279
-
280
-
281
- def rewrite(expression: str) -> str:
282
- return str(expression)
283
-
284
-
285
- def simple_tool_call(client: OpenAI):
286
- messages = [
287
- {
288
- "role": "user",
289
- "content": [
290
- {
291
- "type": "text",
292
- "text": "use my functions to compute the results for the equations: 6+1",
293
- },
294
- ],
295
- },
296
- ]
297
- tools = [
298
- {
299
- "type": "function",
300
- "function": {
301
- "name": "my_calculator",
302
- "description": "A calculator that can evaluate a mathematical equation and compute its results.",
303
- "parameters": {
304
- "type": "object",
305
- "properties": {
306
- "expression": {
307
- "type": "string",
308
- "description": "The mathematical expression to evaluate.",
309
- },
310
- },
311
- "required": ["expression"],
312
- },
313
- },
314
- },
315
- {
316
- "type": "function",
317
- "function": {
318
- "name": "rewrite",
319
- "description": "Rewrite a given text for improved clarity",
320
- "parameters": {
321
- "type": "object",
322
- "properties": {
323
- "text": {
324
- "type": "string",
325
- "description": "The input text to rewrite",
326
- }
327
- },
328
- },
329
- },
330
- },
331
- ]
332
- model_name = client.models.list().data[0].id
333
- response = client.chat.completions.create(
334
- model=model_name,
335
- messages=messages,
336
- temperature=1.0,
337
- max_tokens=1024,
338
- tools=tools,
339
- tool_choice="auto",
340
- )
341
- tool_calls = response.choices[0].message.tool_calls
342
-
343
- results = []
344
- for tool_call in tool_calls:
345
- function_name = tool_call.function.name
346
- function_args = tool_call.function.arguments
347
- if function_name == "my_calculator":
348
- result = my_calculator(**json.loads(function_args))
349
- results.append(result)
350
- messages.append({"role": "assistant", "tool_calls": tool_calls})
351
- for tool_call, result in zip(tool_calls, results):
352
- messages.append(
353
- {
354
- "role": "tool",
355
- "tool_call_id": tool_call.id,
356
- "name": tool_call.function.name,
357
- "content": result,
358
- }
359
- )
360
- response = client.chat.completions.create(
361
- model=model_name,
362
- messages=messages,
363
- temperature=1.0,
364
- max_tokens=1024,
365
- )
366
- print(response.choices[0].message.content)
367
-
368
-
369
- if __name__ == "__main__":
370
- simple_tool_call(client)
371
-
372
- ```
373
 
374
  ---
375
-
376
- ## 6. License
377
-
378
- Both the code repository and the model weights are released under the [Modified MIT License](LICENSE).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model: []
 
 
 
3
  library_name: transformers
4
+ tags:
5
+ - mergekit
6
+ - merge
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
 
8
  ---
9
+ # c362_step50_ta05
10
+
11
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
12
+
13
+ ## Merge Details
14
+ ### Merge Method
15
+
16
+ This model was merged using the [Linear DARE](https://arxiv.org/abs/2311.03099) merge method using /root/myCodeLab/host/downloads/models/40Bra as a base.
17
+
18
+ ### Models Merged
19
+
20
+ The following models were included in the merge:
21
+ * /root/myCodeLab/host/verl/ckpts/40bra_k8s_single_domain/40bra_k8s_16node_sd_c362_20260327_205644_unknown/global_step_50/actor/huggingface
22
+
23
+ ### Configuration
24
+
25
+ The following YAML configuration was used to produce this model:
26
+
27
+ ```yaml
28
+ base_model: /root/myCodeLab/host/downloads/models/40Bra
29
+ dtype: float32
30
+ merge_method: dare_linear
31
+ modules:
32
+ default:
33
+ slices:
34
+ - sources:
35
+ - layer_range: [0, 40]
36
+ model: /root/myCodeLab/host/downloads/models/40Bra
37
+ - layer_range: [0, 40]
38
+ model: /root/myCodeLab/host/verl/ckpts/40bra_k8s_single_domain/40bra_k8s_16node_sd_c362_20260327_205644_unknown/global_step_50/actor/huggingface
39
+ parameters:
40
+ density: 1.0
41
+ weight:
42
+ - filter: .mlp.gate.
43
+ value: 0.0
44
+ - value: 0.5
45
+ - sources:
46
+ - layer_range: [40, 41]
47
+ model: /root/myCodeLab/host/downloads/models/40Bra
48
+ out_dtype: bfloat16
49
+ ```
config.json CHANGED
@@ -10,16 +10,18 @@
10
  "AutoModelForCausalLM": "modeling_deepseek.DeepseekV3ForCausalLM"
11
  },
12
  "bos_token_id": 0,
 
13
  "eos_token_id": 1,
14
  "ep_size": 1,
15
  "first_k_dense_replace": 1,
 
16
  "hidden_act": "silu",
17
  "hidden_size": 2048,
18
  "initializer_range": 0.02,
19
  "intermediate_size": 7168,
20
  "kv_lora_rank": 512,
21
  "max_position_embeddings": 131072,
22
- "model_type": "joyai_llm_flash",
23
  "moe_intermediate_size": 768,
24
  "moe_layer_freq": 1,
25
  "n_group": 1,
@@ -31,7 +33,9 @@
31
  "num_hidden_layers": 40,
32
  "num_key_value_heads": 32,
33
  "num_nextn_predict_layers": 1,
 
34
  "q_lora_rank": 1536,
 
35
  "qk_nope_head_dim": 128,
36
  "qk_rope_head_dim": 64,
37
  "quantization_config": {
@@ -43,15 +47,16 @@
43
  ]
44
  },
45
  "rms_norm_eps": 1e-06,
 
 
46
  "rope_theta": 32000000,
47
  "routed_scaling_factor": 2.5,
48
  "scoring_func": "sigmoid",
49
  "tie_word_embeddings": false,
50
  "topk_group": 1,
51
  "topk_method": "noaux_tc",
52
- "torch_dtype": "bfloat16",
53
- "transformers_version": "4.44.2",
54
  "use_cache": true,
55
  "v_head_dim": 128,
56
  "vocab_size": 129280
57
- }
 
10
  "AutoModelForCausalLM": "modeling_deepseek.DeepseekV3ForCausalLM"
11
  },
12
  "bos_token_id": 0,
13
+ "dtype": "bfloat16",
14
  "eos_token_id": 1,
15
  "ep_size": 1,
16
  "first_k_dense_replace": 1,
17
+ "head_dim": 64,
18
  "hidden_act": "silu",
19
  "hidden_size": 2048,
20
  "initializer_range": 0.02,
21
  "intermediate_size": 7168,
22
  "kv_lora_rank": 512,
23
  "max_position_embeddings": 131072,
24
+ "model_type":"joyai_llm_flash",
25
  "moe_intermediate_size": 768,
26
  "moe_layer_freq": 1,
27
  "n_group": 1,
 
33
  "num_hidden_layers": 40,
34
  "num_key_value_heads": 32,
35
  "num_nextn_predict_layers": 1,
36
+ "pretraining_tp": 1,
37
  "q_lora_rank": 1536,
38
+ "qk_head_dim": 192,
39
  "qk_nope_head_dim": 128,
40
  "qk_rope_head_dim": 64,
41
  "quantization_config": {
 
47
  ]
48
  },
49
  "rms_norm_eps": 1e-06,
50
+ "rope_interleave": true,
51
+ "rope_scaling": null,
52
  "rope_theta": 32000000,
53
  "routed_scaling_factor": 2.5,
54
  "scoring_func": "sigmoid",
55
  "tie_word_embeddings": false,
56
  "topk_group": 1,
57
  "topk_method": "noaux_tc",
58
+ "transformers_version": "4.57.3",
 
59
  "use_cache": true,
60
  "v_head_dim": 128,
61
  "vocab_size": 129280
62
+ }
mergekit_config.yml ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ base_model: /root/myCodeLab/host/downloads/models/40Bra
2
+ dtype: float32
3
+ merge_method: dare_linear
4
+ modules:
5
+ default:
6
+ slices:
7
+ - sources:
8
+ - layer_range: [0, 40]
9
+ model: /root/myCodeLab/host/downloads/models/40Bra
10
+ - layer_range: [0, 40]
11
+ model: /root/myCodeLab/host/verl/ckpts/40bra_k8s_single_domain/40bra_k8s_16node_sd_c362_20260327_205644_unknown/global_step_50/actor/huggingface
12
+ parameters:
13
+ density: 1.0
14
+ weight:
15
+ - filter: .mlp.gate.
16
+ value: 0.0
17
+ - value: 0.5
18
+ - sources:
19
+ - layer_range: [40, 41]
20
+ model: /root/myCodeLab/host/downloads/models/40Bra
21
+ out_dtype: bfloat16
model-1-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f48ddf430827ffdb51975cfd39f9426dc705d549159e0b3031c9386ad60cd92
3
  size 70417424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3e78b8a79349b638ac5cb7139dbee0189474e2a41043e49189150bd46fd0b50
3
  size 70417424
model-10-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c91cf0ce6f06c7dd5ebf4ed4b416fb115e8dc16b96b907571f54af9c8b7addc
3
- size 1240575152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9da377dcb5bac74fd43f4983126f1cf84e7accc5e5dd26a6664fa47acb0443fc
3
+ size 1240574640
model-11-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c3395a6d10132a7965d3e2bebaac84a072624c3f0dde104c3392a7075d0bdab8
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19eafdc4c9b4bc748b0e8dea5e19d3dc718e73c0a4ddf8dc318c90ee32d5f4d0
3
+ size 1240576192
model-12-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ac71db3d3322e5fb3ba411105cbbcd773555cc929a9f59a68473a445030af67
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a418e1068b40ebb60f4b69124ab631c4a0f708172f3365b6c669689da945c1ab
3
+ size 1240576192
model-13-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a24e8ce9f29b0d3c9ff6e5b4d3493df5eb60ee8c52ed1a67cb2d8cf6db217dc5
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e11a81dd89b553ea4ec708fcc2a335b2428967d993ab8c81a473c9d01360e484
3
+ size 1240576192
model-14-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:184e58487f757f2e5fe05f51351f41f3c995578a3b152d406573ef46b85529ae
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c2f4d9e1f17289700b246b2059ec6eff978db14b4a51298925538471c544802
3
+ size 1240576192
model-15-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a50c38dd5267e2eb8470c9b0b6ce80c397a70def85147cab046e91f40bacf30
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acfd3c136db7e3419b22de25de2942bf35a78dc5a5d2fb253e02f766c4be33ea
3
+ size 1240576192
model-16-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:11864879ee5baf912acb086c736130cdeb40a48c9defb2b6ec2446c722bf81d6
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44a56cb3b937d388bddf3f40407b2705bbe91c0bb1af31dfdd43102c5df2f5d2
3
+ size 1240576192
model-17-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e03ce933ed6dda2dd9ced2fc05f6f9ff20f22bb5a924530a2ecc5fdc1d643a9
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7df20a82591ba541f59038b4f4e1cf7ae5e2998614ad954504d425a60b5c0ea4
3
+ size 1240576192
model-18-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bbe260ad95b9c788ea2a2b7f304fe0c765ddcfaca1e2ff344da558d7f86baef
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a305a86aa156997d8845359b907b2c7363909f37d59ddd6b532276e2d8c5f8c4
3
+ size 1240576192
model-19-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4569083a64418b328aa7b9083068391be62f46337c662a752ff93b4a501758c8
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:118013021992908478f2c7318cf57a22325769be520c262f86b385ab267678b7
3
+ size 1240576192
model-2-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b51f6e450426f9a1acca1daac7c28b15927c709612fb7b5df9bdd8b7d5cf7ef6
3
- size 1240575152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77dfe5162ff011af212696ac29b31273396121f0f08b17ad147bcf0592f87319
3
+ size 1240574640
model-20-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e7f08b5e331b4c094f76702e0b8ccfc5ffe540cfed309d6534a9d81a9ca36029
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab2c3b4fc1a6d3f0671782bef9beeb8766825d67213889938b05ff6a9551fa9b
3
+ size 1240576192
model-21-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c04abe8b76c5015f97526e79df0a85a9f17615d9fd3d8662e13b1209854c4a1
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:921c963a2a3948712da54396366ee7871c023d8473799f0d6bda2674cae7e6c0
3
+ size 1240576192
model-22-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c56faaa1cb68b0f89f2cd6afaa00afb2072de41dc56c85a7a1fe69aeca282eab
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:723d02655f20936a6efb99d81295832abab6081a735c36a29149a333d753d845
3
+ size 1240576192
model-23-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:76b32333d89529b8f406dddd0b16571eb3f07a6c055c4f63fa05e8bf1cb61822
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd583ecdbc08fb94ac2b701936b763e779669604d34e1608c41770d3fff0dc32
3
+ size 1240576192
model-24-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7dcf0fe97ee36d329d237c55b9231e28b2f5536678d7ba5deab4fae0c9737206
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1027648adeb8aeb6baa34f1f8fcc0ccde91f7684e0684cac1eacd6b78f11ef53
3
+ size 1240576192
model-25-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:09e63583cb2acb5d279ba8d61674aa8c0e031fda96664b8e833713278a4696a0
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26425adf5423eeb45dd469b455a0d86f1f07c4ac6b9fb2274c8bf57c173cb8bf
3
+ size 1240576192
model-26-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fba5920b459dbf18927202eccd87ef47013b0537a6863772b9466594d820f625
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf7d3674d89ad3e7b75cf2aa1bbafc4d4d59c188ef2ca813d27068a0f181102a
3
+ size 1240576192
model-27-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9c7ca88ee3f851b2a2fa2edba45bf94677d50554710535310536cbcdc21a4e4
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23325ad47f7967fbc63e53562cbfc27d7f0f8ad149a34fc224d9ca0faa588600
3
+ size 1240576192
model-28-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfcef06073d7eff69a26d30a298444eee9c17968b970a9af48bd9386860feec2
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0f0c1c41480fa1405a05777294d6925980d8b7157112b05a12aab35f0c007cb
3
+ size 1240576192
model-29-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a8cab07da30997cc7dfa269362b2f03c8ff8c8bc199557bd0d12aca7fdbcc22
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8c478ff960026dd7ee4ba9048b092b81cac6eb38e67be0f68959702601a9e0d
3
+ size 1240576192
model-3-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e41720e0ef2303621353674634234a406bde493f8e05d88b23084ffe112d67db
3
- size 1240575152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b82527e24104dcb69840afaf49cfbb16449c42465854d01fb73fa6827de7cf95
3
+ size 1240574640
model-30-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfa7161d6099ff72463fef10762583c290fe5682e3b938e4066c93e4ace5b71e
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba4e991f40e5bf3ffa2efc2f31ad31b05aea580cfda986fd14fc6f7fe2913cb6
3
+ size 1240576192
model-31-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ccdd3eb8e24ff9561e77068e3ac74eaea1746d8925e71965a1b2b8127f1972e6
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddb4c3211db92ff361c5ea2f04337a97aa67c36eab38de8f2cf18cb4811a2b94
3
+ size 1240576192
model-32-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:96810a9f732857b8fa8f1e08f957392ccf41abf1cfe2a38a95c01a9fd70b9502
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d99e57522ed93ba01e14b7d0d98b26f8f0e5de2711c97ae894628f380f310c7
3
+ size 1240576192
model-33-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d3f02a4e22221075d3a84521fe06a2d8e714b1d9bd2c2d07c1bc417a76d52e6d
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5551f0b231afa4606982c970a30e36b488cda4a9cde7084d3381c602a0f9f8d5
3
+ size 1240576192
model-34-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82982c5da6ec2217f169924c05f4868eb9353654fcbe52fc555cdf1e54e954ab
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8c862fc52f1f34046d81867df9dfbc320f16355b1072b7048484e7a94ee26b1
3
+ size 1240576192
model-35-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7cddd2ed5e46db1ce96c5ac9e6bca7a813498bc2be2872fd8d847261cee6a10
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0063484a9d45f05b435069c82b8e0f1630df7b41eb48d8ec0c3d34d5383d15d5
3
+ size 1240576192
model-36-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85bafba90f7517fcfc50de98af728dd7943c54e4bcc936ef28d3390391370a2b
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e964f833051abd324f4af376864e262a7dae09c83d30d57c292beab094c01b58
3
+ size 1240576192
model-37-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:932cf6ccecd8f8e1b6dea7ff99318d607432c25cd243278171009f706d3eca8f
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30d882ce806dd69ab3e1a6a1e8215c037a97dc2b8eebe39b1e73a5d597b10624
3
+ size 1240576192
model-38-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d02dce21d360f7a2390ec11da920c6516407f0abb86b8e08886e6829794541e
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01b9a194c9882d07a4fe0344ec9e44267cbbfe77efb37464ca94943ba5cb4894
3
+ size 1240576192
model-39-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:71d231d31a4da0c90c36cb95ffe091937b67199f27bd4da4f97c4511068e73c4
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b700ada0cd83229be4003b49ad99ebafcbb01de3a3a344be9033a79fb40272a4
3
+ size 1240576192
model-4-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6ebab19e7fbdeb44956dd3ef63c1a1900c36fd4e9b0799f2adddd3045eefadfc
3
- size 1240575152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e8cc7d10ab9a1e9f7858ad9d7be017e294d92c2aeab1d5285c7dc32e0171f09
3
+ size 1240574640
model-40-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8d9b8b7aec70e054c9da57d3f76bff33c059453bd7b24512ea980de1e36ffa27
3
- size 1240576704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2bfff3bca35a5b9a2f055f5169912d692629a7e94b402df99211185d520985f
3
+ size 1240576192
model-5-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:83f78c8d03ee6a20fcf3c8b1c50da169b9f5fe32cdf66961cf847470e3a76fd6
3
- size 1240575152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f053c4ad2283fbca4bc849d358c1aca1a01dfe879c2e415192fa0d730a5804c
3
+ size 1240574640
model-6-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:316b401e76e4db37a6ca37aba97fd91238cf8213cc631e43401c4c40b9298612
3
- size 1240575152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1c9591722e7fbfa9a914620a5b7e3b6768e019b249e48b5228700354adf8548
3
+ size 1240574640
model-7-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91f8cb67739b65bfc0417beca5b5174b3c2f0b3bf6e139c23e4ae91e361a0944
3
- size 1240575152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae77ea69752a2e2ea748a20646b640cbbd4cef91e5947f6f31e08c115fb5c780
3
+ size 1240574640
model-8-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75a4595159bacc23dc5b85f737a53c88d66c1e594c1a1642e9cc0e5e1a2e9402
3
- size 1240575152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09cf4e2fc6c07b51b48ee048a930a379331e509146dd24cba7f9ccce1844c235
3
+ size 1240574640
model-9-of-40.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9212c9d8925dba5e1367c32032afc05af97c8c3a6d96555aa38ece98ddde31ad
3
- size 1240575152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:131c1e377964f5f059532619e45416af849af7662675595f7d475023daba37cc
3
+ size 1240574640
model-non-layer.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:358810bbba6e784b3be7462f7398fed2ba159a0edcec4f586aa72e877f4187ca
3
  size 1059066184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b992b0a3e3ab06a9490d364bf942df3bcf69874dcb9f940e935f674672f09cd
3
  size 1059066184