jobs-git lcz75 commited on
Commit
99c0bb4
·
verified ·
0 Parent(s):

Duplicate from Skywork/SkyReels-V2-DF-14B-720P

Browse files

Co-authored-by: linchunze <lcz75@users.noreply.huggingface.co>

.gitattributes ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ google/umt5-xxl/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ assets/comp_effic.png filter=lfs diff=lfs merge=lfs -text
38
+ assets/data_for_diff_stage.jpg filter=lfs diff=lfs merge=lfs -text
39
+ assets/i2v_res.png filter=lfs diff=lfs merge=lfs -text
40
+ assets/logo.png filter=lfs diff=lfs merge=lfs -text
41
+ assets/t2v_res.jpg filter=lfs diff=lfs merge=lfs -text
42
+ assets/vben_vs_sota.png filter=lfs diff=lfs merge=lfs -text
43
+ assets/vben_vs_sota_t2i.jpg filter=lfs diff=lfs merge=lfs -text
44
+ assets/video_dit_arch.jpg filter=lfs diff=lfs merge=lfs -text
45
+ assets/video_vae_res.jpg filter=lfs diff=lfs merge=lfs -text
46
+ examples/i2v_input.JPG filter=lfs diff=lfs merge=lfs -text
47
+ assets/.DS_Store filter=lfs diff=lfs merge=lfs -text
48
+ assets/main_pipeline.jpg filter=lfs diff=lfs merge=lfs -text
LICENSE ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - zh
5
+ license: other
6
+ tasks:
7
+ - video-generation
8
+
9
+ ---
10
+
11
+ <!-- markdownlint-disable first-line-h1 -->
12
+ <!-- markdownlint-disable html -->
13
+
14
+ # <span id="Terms">声明与协议/Terms and Conditions</span>
15
+
16
+ ## 声明
17
+
18
+ 我们在此声明,不要利用Skywork模型进行任何危害国家社会安全或违法的活动。另外,我们也要求使用者不要将 Skywork 模型用于未经适当安全审查和备案的互联网服务。我们希望所有的使用者都能遵守这个原则,确保科技的发展能在规范和合法的环境下进行。
19
+
20
+ 我们已经尽我们所能,来确保模型训练过程中使用的数据的合规性。然而,尽管我们已经做出了巨大的努力,但由于模型和数据的复杂性,仍有可能存在一些无法预见的问题。因此,如果由于使用skywork开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。
21
+
22
+ We hereby declare that the Skywork model should not be used for any activities that pose a threat to national or societal security or engage in unlawful actions. Additionally, we request users not to deploy the Skywork model for internet services without appropriate security reviews and records. We hope that all users will adhere to this principle to ensure that technological advancements occur in a regulated and lawful environment.
23
+
24
+ We have done our utmost to ensure the compliance of the data used during the model's training process. However, despite our extensive efforts, due to the complexity of the model and data, there may still be unpredictable risks and issues. Therefore, if any problems arise as a result of using the Skywork open-source model, including but not limited to data security issues, public opinion risks, or any risks and problems arising from the model being misled, abused, disseminated, or improperly utilized, we will not assume any responsibility.
25
+
26
+ ## 协议
27
+
28
+ 社区使用Skywork模型需要遵循[《Skywork 模型社区许可协议》](https://github.com/SkyworkAI/Skywork/blob/main/Skywork%20模型社区许可协议.pdf)。Skywork模型支持商业用途,如果您计划将Skywork模型或其衍生品用于商业目的,无需再次申请, 但请您仔细阅读[《Skywork 模型社区许可协议》](https://github.com/SkyworkAI/Skywork/blob/main/Skywork%20模型社区许可协议.pdf)并严格遵守相关条款。
29
+
30
+
31
+ The community usage of Skywork model requires [Skywork Community License](https://github.com/SkyworkAI/Skywork/blob/main/Skywork%20Community%20License.pdf). The Skywork model supports commercial use. If you plan to use the Skywork model or its derivatives for commercial purposes, you must abide by terms and conditions within [Skywork Community License](https://github.com/SkyworkAI/Skywork/blob/main/Skywork%20Community%20License.pdf).
32
+
33
+
34
+
35
+ [《Skywork 模型社区许可协议》》]:https://github.com/SkyworkAI/Skywork/blob/main/Skywork%20模型社区许可协议.pdf
36
+
37
+
38
+ [skywork-opensource@kunlun-inc.com]: mailto:skywork-opensource@kunlun-inc.com
README.md ADDED
@@ -0,0 +1,723 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: skywork-license
4
+ license_link: LICENSE
5
+ pipeline_tag: text-to-video
6
+ ---
7
+ <p align="center">
8
+ <img src="assets/logo2.png" alt="SkyReels Logo" width="50%">
9
+ </p>
10
+
11
+ <h1 align="center">SkyReels V2: Infinite-Length Film Generative Model</h1>
12
+
13
+ <p align="center">
14
+ 📑 <a href="https://arxiv.org/pdf/2504.13074">Technical Report</a> · 👋 <a href="https://www.skyreels.ai/home?utm_campaign=huggingface_skyreels_v2" target="_blank">Playground</a> · 💬 <a href="https://discord.gg/PwM6NYtccQ" target="_blank">Discord</a> · 🤗 <a href="https://huggingface.co/collections/Skywork/skyreels-v2-6801b1b93df627d441d0d0d9" target="_blank">Hugging Face</a> · 🤖 <a href="https://www.modelscope.cn/collections/SkyReels-V2-f665650130b144" target="_blank">ModelScope</a> · 🌐 <a href="https://github.com/SkyworkAI/SkyReels-V2" target="_blank">GitHub</a>
15
+ </p>
16
+
17
+ ---
18
+ Welcome to the **SkyReels V2** repository! Here, you'll find the model weights for our infinite-length film generative models. To the best of our knowledge, it represents the first open-source video generative model employing **AutoRegressive Diffusion-Forcing architecture** that achieves the **SOTA performance** among publicly available models.
19
+
20
+
21
+ ## 🔥🔥🔥 News!!
22
+ * Apr 24, 2025: 🔥 We release the 720P models, [SkyReels-V2-DF-14B-720P](https://huggingface.co/Skywork/SkyReels-V2-DF-14B-720P) and [SkyReels-V2-I2V-14B-720P](https://huggingface.co/Skywork/SkyReels-V2-I2V-14B-720P). The former facilitates infinite-length autoregressive video generation, and the latter focuses on Image2Video synthesis.
23
+ * Apr 21, 2025: 👋 We release the inference code and model weights of [SkyReels-V2](https://huggingface.co/collections/Skywork/skyreels-v2-6801b1b93df627d441d0d0d9) Series Models and the video captioning model [SkyCaptioner-V1](https://huggingface.co/Skywork/SkyCaptioner-V1) .
24
+ * Apr 3, 2025: 🔥 We also release [SkyReels-A2](https://github.com/SkyworkAI/SkyReels-A2). This is an open-sourced controllable video generation framework capable of assembling arbitrary visual elements.
25
+ * Feb 18, 2025: 🔥 we released [SkyReels-A1](https://github.com/SkyworkAI/SkyReels-A1). This is an open-sourced and effective framework for portrait image animation.
26
+ * Feb 18, 2025: 🔥 We released [SkyReels-V1](https://github.com/SkyworkAI/SkyReels-V1). This is the first and most advanced open-source human-centric video foundation model.
27
+
28
+ ## 🎥 Demos
29
+ <table>
30
+ <tr>
31
+ <td align="center">
32
+ <video src="https://github.com/user-attachments/assets/f6f9f9a7-5d5f-433c-9d73-d8d593b7ad25" width="100%"></video>
33
+ </td>
34
+ <td align="center">
35
+ <video src="https://github.com/user-attachments/assets/0eb13415-f4d9-4aaf-bcd3-3031851109b9" width="100%"></video>
36
+ </td>
37
+ <td align="center">
38
+ <video src="https://github.com/user-attachments/assets/dcd16603-5bf4-4786-8e4d-1ed23889d07a" width="100%"></video>
39
+ </td>
40
+ </tr>
41
+ </table>
42
+ The demos above showcase 30-second videos generated using our SkyReels-V2 Diffusion Forcing model.
43
+
44
+
45
+ ## 📑 TODO List
46
+
47
+ - [x] <a href="https://arxiv.org/pdf/2504.13074">Technical Report</a>
48
+ - [x] Checkpoints of the 14B and 1.3B Models Series
49
+ - [x] Single-GPU & Multi-GPU Inference Code
50
+ - [x] <a href="https://huggingface.co/Skywork/SkyCaptioner-V1">SkyCaptioner-V1</a>: A Video Captioning Model
51
+ - [x] Prompt Enhancer
52
+ - [ ] Diffusers integration
53
+ - [ ] Checkpoints of the 5B Models Series
54
+ - [ ] Checkpoints of the Camera Director Models
55
+ - [ ] Checkpoints of the Step & Guidance Distill Model
56
+
57
+
58
+ ## 🚀 Quickstart
59
+
60
+ #### Installation
61
+ ```shell
62
+ # clone the repository.
63
+ git clone https://github.com/SkyworkAI/SkyReels-V2
64
+ cd SkyReels-V2
65
+ # Install dependencies. Test environment uses Python 3.10.12.
66
+ pip install -r requirements.txt
67
+ ```
68
+
69
+ #### Model Download
70
+ You can download our models from Hugging Face:
71
+ <table>
72
+ <thead>
73
+ <tr>
74
+ <th>Type</th>
75
+ <th>Model Variant</th>
76
+ <th>Recommended Height/Width/Frame</th>
77
+ <th>Link</th>
78
+ </tr>
79
+ </thead>
80
+ <tbody>
81
+ <tr>
82
+ <td rowspan="5">Diffusion Forcing</td>
83
+ <td>1.3B-540P</td>
84
+ <td>544 * 960 * 97f</td>
85
+ <td>🤗 <a href="https://huggingface.co/Skywork/SkyReels-V2-DF-1.3B-540P">Huggingface</a> 🤖 <a href="https://www.modelscope.cn/models/Skywork/SkyReels-V2-DF-1.3B-540P">ModelScope</a></td>
86
+ </tr>
87
+ <tr>
88
+ <td>5B-540P</td>
89
+ <td>544 * 960 * 97f</td>
90
+ <td>Coming Soon</td>
91
+ </tr>
92
+ <tr>
93
+ <td>5B-720P</td>
94
+ <td>720 * 1280 * 121f</td>
95
+ <td>Coming Soon</td>
96
+ </tr>
97
+ <tr>
98
+ <td>14B-540P</td>
99
+ <td>544 * 960 * 97f</td>
100
+ <td>🤗 <a href="https://huggingface.co/Skywork/SkyReels-V2-DF-14B-540P">Huggingface</a> 🤖 <a href="https://www.modelscope.cn/models/Skywork/SkyReels-V2-DF-14B-540P">ModelScope</a></td>
101
+ </tr>
102
+ <tr>
103
+ <td>14B-720P</td>
104
+ <td>720 * 1280 * 121f</td>
105
+ <td>🤗 <a href="https://huggingface.co/Skywork/SkyReels-V2-DF-14B-720P">Huggingface</a> 🤖 <a href="https://www.modelscope.cn/models/Skywork/SkyReels-V2-DF-14B-720P">ModelScope</a></td>
106
+ </tr>
107
+ <tr>
108
+ <td rowspan="5">Text-to-Video</td>
109
+ <td>1.3B-540P</td>
110
+ <td>544 * 960 * 97f</td>
111
+ <td>Coming Soon</td>
112
+ </tr>
113
+ <tr>
114
+ <td>5B-540P</td>
115
+ <td>544 * 960 * 97f</td>
116
+ <td>Coming Soon</td>
117
+ </tr>
118
+ <tr>
119
+ <td>5B-720P</td>
120
+ <td>720 * 1280 * 121f</td>
121
+ <td>Coming Soon</td>
122
+ </tr>
123
+ <tr>
124
+ <td>14B-540P</td>
125
+ <td>544 * 960 * 97f</td>
126
+ <td>🤗 <a href="https://huggingface.co/Skywork/SkyReels-V2-T2V-14B-540P">Huggingface</a> 🤖 <a href="https://www.modelscope.cn/models/Skywork/SkyReels-V2-T2V-14B-540P">ModelScope</a></td>
127
+ </tr>
128
+ <tr>
129
+ <td>14B-720P</td>
130
+ <td>720 * 1280 * 121f</td>
131
+ <td>🤗 <a href="https://huggingface.co/Skywork/SkyReels-V2-T2V-14B-720P">Huggingface</a> 🤖 <a href="https://www.modelscope.cn/models/Skywork/SkyReels-V2-T2V-14B-720P">ModelScope</a></td>
132
+ </tr>
133
+ <tr>
134
+ <td rowspan="5">Image-to-Video</td>
135
+ <td>1.3B-540P</td>
136
+ <td>544 * 960 * 97f</td>
137
+ <td>🤗 <a href="https://huggingface.co/Skywork/SkyReels-V2-I2V-1.3B-540P">Huggingface</a> 🤖 <a href="https://www.modelscope.cn/models/Skywork/SkyReels-V2-I2V-1.3B-540P">ModelScope</a></td>
138
+ </tr>
139
+ <tr>
140
+ <td>5B-540P</td>
141
+ <td>544 * 960 * 97f</td>
142
+ <td>Coming Soon</td>
143
+ </tr>
144
+ <tr>
145
+ <td>5B-720P</td>
146
+ <td>720 * 1280 * 121f</td>
147
+ <td>Coming Soon</td>
148
+ </tr>
149
+ <tr>
150
+ <td>14B-540P</td>
151
+ <td>544 * 960 * 97f</td>
152
+ <td>🤗 <a href="https://huggingface.co/Skywork/SkyReels-V2-I2V-14B-540P">Huggingface</a> 🤖 <a href="https://www.modelscope.cn/models/Skywork/SkyReels-V2-I2V-14B-540P">ModelScope</a></td>
153
+ </tr>
154
+ <tr>
155
+ <td>14B-720P</td>
156
+ <td>720 * 1280 * 121f</td>
157
+ <td>🤗 <a href="https://huggingface.co/Skywork/SkyReels-V2-I2V-14B-720P">Huggingface</a> 🤖 <a href="https://www.modelscope.cn/models/Skywork/SkyReels-V2-I2V-14B-720P">ModelScope</a></td>
158
+ </tr>
159
+ <tr>
160
+ <td rowspan="3">Camera Director</td>
161
+ <td>5B-540P</td>
162
+ <td>544 * 960 * 97f</td>
163
+ <td>Coming Soon</td>
164
+ </tr>
165
+ <tr>
166
+ <td>5B-720P</td>
167
+ <td>720 * 1280 * 121f</td>
168
+ <td>Coming Soon</td>
169
+ </tr>
170
+ <tr>
171
+ <td>14B-720P</td>
172
+ <td>720 * 1280 * 121f</td>
173
+ <td>Coming Soon</td>
174
+ </tr>
175
+ </tbody>
176
+ </table>
177
+
178
+ After downloading, set the model path in your generation commands:
179
+
180
+
181
+ #### Single GPU Inference
182
+
183
+ - **Diffusion Forcing for Long Video Generation**
184
+
185
+ The <a href="https://arxiv.org/abs/2407.01392">**Diffusion Forcing**</a> version model allows us to generate Infinite-Length videos. This model supports both **text-to-video (T2V)** and **image-to-video (I2V)** tasks, and it can perform inference in both synchronous and asynchronous modes. Here we demonstrate 2 running scripts as examples for long video generation. If you want to adjust the inference parameters, e.g., the duration of video, inference mode, read the Note below first.
186
+
187
+ synchronous generation for 10s video
188
+ ```shell
189
+ model_id=Skywork/SkyReels-V2-DF-14B-540P
190
+ # synchronous inference
191
+ python3 generate_video_df.py \
192
+ --model_id ${model_id} \
193
+ --resolution 540P \
194
+ --ar_step 0 \
195
+ --base_num_frames 97 \
196
+ --num_frames 257 \
197
+ --overlap_history 17 \
198
+ --prompt "A graceful white swan with a curved neck and delicate feathers swimming in a serene lake at dawn, its reflection perfectly mirrored in the still water as mist rises from the surface, with the swan occasionally dipping its head into the water to feed." \
199
+ --addnoise_condition 20 \
200
+ --offload \
201
+ --teacache \
202
+ --use_ret_steps \
203
+ --teacache_thresh 0.3
204
+ ```
205
+
206
+ asynchronous generation for 30s video
207
+ ```shell
208
+ model_id=Skywork/SkyReels-V2-DF-14B-540P
209
+ # asynchronous inference
210
+ python3 generate_video_df.py \
211
+ --model_id ${model_id} \
212
+ --resolution 540P \
213
+ --ar_step 5 \
214
+ --causal_block_size 5 \
215
+ --base_num_frames 97 \
216
+ --num_frames 737 \
217
+ --overlap_history 17 \
218
+ --prompt "A graceful white swan with a curved neck and delicate feathers swimming in a serene lake at dawn, its reflection perfectly mirrored in the still water as mist rises from the surface, with the swan occasionally dipping its head into the water to feed." \
219
+ --addnoise_condition 20 \
220
+ --offload
221
+ ```
222
+
223
+ > **Note**:
224
+ > - If you want to run the **image-to-video (I2V)** task, add `--image ${image_path}` to your command and it is also better to use **text-to-video (T2V)**-like prompt which includes some descriptions of the first-frame image.
225
+ > - For long video generation, you can just switch the `--num_frames`, e.g., `--num_frames 257` for 10s video, `--num_frames 377` for 15s video, `--num_frames 737` for 30s video, `--num_frames 1457` for 60s video. The number is not strictly aligned with the logical frame number for specified time duration, but it is aligned with some training parameters, which means it may perform better. When you use asynchronous inference with causal_block_size > 1, the `--num_frames` should be carefully set.
226
+ > - You can use `--ar_step 5` to enable asynchronous inference. When asynchronous inference, `--causal_block_size 5` is recommended while it is not supposed to be set for synchronous generation. REMEMBER that the frame latent number inputted into the model in every iteration, e.g., base frame latent number (e.g., (97-1)//4+1=25 for base_num_frames=97) and (e.g., (237-97-(97-17)x1+17-1)//4+1=20 for base_num_frames=97, num_frames=237, overlap_history=17) for the last iteration, MUST be divided by causal_block_size. If you find it too hard to calculate and set proper values, just use our recommended setting above :). Asynchronous inference will take more steps to diffuse the whole sequence which means it will be SLOWER than synchronous mode. In our experiments, asynchronous inference may improve the instruction following and visual consistent performance.
227
+ > - To reduce peak VRAM, just lower the `--base_num_frames`, e.g., to 77 or 57, while keeping the same generative length `--num_frames` you want to generate. This may slightly reduce video quality, and it should not be set too small.
228
+ > - `--addnoise_condition` is used to help smooth the long video generation by adding some noise to the clean condition. Too large noise can cause the inconsistency as well. 20 is a recommended value, and you may try larger ones, but it is recommended to not exceed 50.
229
+ > - Generating a 540P video using the 1.3B model requires approximately 14.7GB peak VRAM, while the same resolution video using the 14B model demands around 51.2GB peak VRAM.
230
+
231
+ - **Text To Video & Image To Video**
232
+
233
+ ```shell
234
+ # run Text-to-Video Generation
235
+ model_id=Skywork/SkyReels-V2-T2V-14B-540P
236
+ python3 generate_video.py \
237
+ --model_id ${model_id} \
238
+ --resolution 540P \
239
+ --num_frames 97 \
240
+ --guidance_scale 6.0 \
241
+ --shift 8.0 \
242
+ --fps 24 \
243
+ --prompt "A serene lake surrounded by towering mountains, with a few swans gracefully gliding across the water and sunlight dancing on the surface." \
244
+ --offload \
245
+ --teacache \
246
+ --use_ret_steps \
247
+ --teacache_thresh 0.3
248
+ ```
249
+ > **Note**:
250
+ > - When using an **image-to-video (I2V)** model, you must provide an input image using the `--image ${image_path}` parameter. The `--guidance_scale 5.0` and `--shift 3.0` is recommended for I2V model.
251
+ > - Generating a 540P video using the 1.3B model requires approximately 14.7GB peak VRAM, while the same resolution video using the 14B model demands around 43.4GB peak VRAM.
252
+
253
+
254
+ - **Prompt Enhancer**
255
+
256
+ The prompt enhancer is implemented based on <a href="https://huggingface.co/Qwen/Qwen2.5-32B-Instruct">Qwen2.5-32B-Instruct</a> and is utilized via the `--prompt_enhancer` parameter. It works ideally for short prompts, while for long prompts, it might generate an excessively lengthy prompt that could lead to over-saturation in the generative video. Note the peak memory of GPU is 64G+ if you use `--prompt_enhancer`. If you want to obtain the enhanced prompt separately, you can also run the prompt_enhancer script separately for testing. The steps are as follows:
257
+
258
+ ```shell
259
+ cd skyreels_v2_infer/pipelines
260
+ python3 prompt_enhancer.py --prompt "A serene lake surrounded by towering mountains, with a few swans gracefully gliding across the water and sunlight dancing on the surface."
261
+ ```
262
+ > **Note**:
263
+ > - `--prompt_enhancer` is not allowed if using `--use_usp`. We recommend running the skyreels_v2_infer/pipelines/prompt_enhancer.py script first to generate enhanced prompt before enabling the `--use_usp` parameter.
264
+
265
+
266
+ **Advanced Configuration Options**
267
+
268
+ Below are the key parameters you can customize for video generation:
269
+
270
+ | Parameter | Recommended Value | Description |
271
+ |:----------------------:|:---------:|:-----------------------------------------:|
272
+ | --prompt | | Text description for generating your video |
273
+ | --image | | Path to input image for image-to-video generation |
274
+ | --resolution | 540P or 720P | Output video resolution (select based on model type) |
275
+ | --num_frames | 97 or 121 | Total frames to generate (**97 for 540P models**, **121 for 720P models**) |
276
+ | --inference_steps | 50 | Number of denoising steps |
277
+ | --fps | 24 | Frames per second in the output video |
278
+ | --shift | 8.0 or 5.0 | Flow matching scheduler parameter (**8.0 for T2V**, **5.0 for I2V**) |
279
+ | --guidance_scale | 6.0 or 5.0 | Controls text adherence strength (**6.0 for T2V**, **5.0 for I2V**) |
280
+ | --seed | | Fixed seed for reproducible results (omit for random generation) |
281
+ | --offload | True | Offloads model components to CPU to reduce VRAM usage (recommended) |
282
+ | --use_usp | True | Enables multi-GPU acceleration with xDiT USP |
283
+ | --outdir | ./video_out | Directory where generated videos will be saved |
284
+ | --prompt_enhancer | True | Expand the prompt into a more detailed description |
285
+ | --teacache | False | Enables teacache for faster inference |
286
+ | --teacache_thresh | 0.2 | Higher speedup will cause to worse quality |
287
+ | --use_ret_steps | False | Retention Steps for teacache |
288
+
289
+ **Diffusion Forcing Additional Parameters**
290
+ | Parameter | Recommended Value | Description |
291
+ |:----------------------:|:---------:|:-----------------------------------------:|
292
+ | --ar_step | 0 | Controls asynchronous inference (0 for synchronous mode) |
293
+ | --base_num_frames | 97 or 121 | Base frame count (**97 for 540P**, **121 for 720P**) |
294
+ | --overlap_history | 17 | Number of frames to overlap for smooth transitions in long videos |
295
+ | --addnoise_condition | 20 | Improves consistency in long video generation |
296
+ | --causal_block_size | 5 | Recommended when using asynchronous inference (--ar_step > 0) |
297
+
298
+ #### Multi-GPU inference using xDiT USP
299
+
300
+ We use [xDiT](https://github.com/xdit-project/xDiT) USP to accelerate inference. For example, to generate a video with 2 GPUs, you can use the following command:
301
+ - **Diffusion Forcing**
302
+ ```shell
303
+ model_id=Skywork/SkyReels-V2-DF-14B-540P
304
+ # diffusion forcing synchronous inference
305
+ torchrun --nproc_per_node=2 generate_video_df.py \
306
+ --model_id ${model_id} \
307
+ --resolution 540P \
308
+ --ar_step 0 \
309
+ --base_num_frames 97 \
310
+ --num_frames 257 \
311
+ --overlap_history 17 \
312
+ --prompt "A graceful white swan with a curved neck and delicate feathers swimming in a serene lake at dawn, its reflection perfectly mirrored in the still water as mist rises from the surface, with the swan occasionally dipping its head into the water to feed." \
313
+ --addnoise_condition 20 \
314
+ --use_usp \
315
+ --offload \
316
+ --seed 42
317
+ ```
318
+ - **Text To Video & Image To Video**
319
+ ```shell
320
+ # run Text-to-Video Generation
321
+ model_id=Skywork/SkyReels-V2-T2V-14B-540P
322
+ torchrun --nproc_per_node=2 generate_video.py \
323
+ --model_id ${model_id} \
324
+ --resolution 540P \
325
+ --num_frames 97 \
326
+ --guidance_scale 6.0 \
327
+ --shift 8.0 \
328
+ --fps 24 \
329
+ --offload \
330
+ --prompt "A serene lake surrounded by towering mountains, with a few swans gracefully gliding across the water and sunlight dancing on the surface." \
331
+ --use_usp \
332
+ --seed 42
333
+ ```
334
+ > **Note**:
335
+ > - When using an **image-to-video (I2V)** model, you must provide an input image using the `--image ${image_path}` parameter. The `--guidance_scale 5.0` and `--shift 3.0` is recommended for I2V model.
336
+
337
+
338
+ ## Contents
339
+ - [Abstract](#abstract)
340
+ - [Methodology of SkyReels-V2](#methodology-of-skyreels-v2)
341
+ - [Key Contributions of SkyReels-V2](#key-contributions-of-skyreels-v2)
342
+ - [Video Captioner](#video-captioner)
343
+ - [Reinforcement Learning](#reinforcement-learning)
344
+ - [Diffusion Forcing](#diffusion-forcing)
345
+ - [High-Quality Supervised Fine-Tuning(SFT)](#high-quality-supervised-fine-tuning-sft)
346
+ - [Performance](#performance)
347
+ - [Acknowledgements](#acknowledgements)
348
+ - [Citation](#citation)
349
+ ---
350
+
351
+ ## Abstract
352
+ Recent advances in video generation have been driven by diffusion models and autoregressive frameworks, yet critical challenges persist in harmonizing prompt adherence, visual quality, motion dynamics, and duration: compromises in motion dynamics to enhance temporal visual quality, constrained video duration (5-10 seconds) to prioritize resolution, and inadequate shot-aware generation stemming from general-purpose MLLMs' inability to interpret cinematic grammar, such as shot composition, actor expressions, and camera motions. These intertwined limitations hinder realistic long-form synthesis and professional film-style generation.
353
+
354
+ To address these limitations, we introduce SkyReels-V2, the world's first infinite-length film generative model using a Diffusion Forcing framework. Our approach synergizes Multi-modal Large Language Models (MLLM), Multi-stage Pretraining, Reinforcement Learning, and Diffusion Forcing techniques to achieve comprehensive optimization. Beyond its technical innovations, SkyReels-V2 enables multiple practical applications, including Story Generation, Image-to-Video Synthesis, Camera Director functionality, and multi-subject consistent video generation through our <a href="https://github.com/SkyworkAI/SkyReels-A2">Skyreels-A2</a> system.
355
+
356
+ ## Methodology of SkyReels-V2
357
+
358
+ The SkyReels-V2 methodology consists of several interconnected components. It starts with a comprehensive data processing pipeline that prepares various quality training data. At its core is the Video Captioner architecture, which provides detailed annotations for video content. The system employs a multi-task pretraining strategy to build fundamental video generation capabilities. Post-training optimization includes Reinforcement Learning to enhance motion quality, Diffusion Forcing Training for generating extended videos, and High-quality Supervised Fine-Tuning (SFT) stages for visual refinement. The model runs on optimized computational infrastructure for efficient training and inference. SkyReels-V2 supports multiple applications, including Story Generation, Image-to-Video Synthesis, Camera Director functionality, and Elements-to-Video Generation.
359
+
360
+ <p align="center">
361
+ <img src="assets/main_pipeline.jpg" alt="mainpipeline" width="100%">
362
+ </p>
363
+
364
+ ## Key Contributions of SkyReels-V2
365
+
366
+ #### Video Captioner
367
+
368
+ <a href="https://huggingface.co/Skywork/SkyCaptioner-V1">SkyCaptioner-V1</a> serves as our video captioning model for data annotation. This model is trained on the captioning result from the base model <a href="https://huggingface.co/Qwen/Qwen2.5-VL-72B-Instruct">Qwen2.5-VL-72B-Instruct</a> and the sub-expert captioners on a balanced video data. The balanced video data is a carefully curated dataset of approximately 2 million videos to ensure conceptual balance and annotation quality. Built upon the <a href="https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct">Qwen2.5-VL-7B-Instruct</a> foundation model, <a href="https://huggingface.co/Skywork/SkyCaptioner-V1">SkyCaptioner-V1</a> is fine-tuned to enhance performance in domain-specific video captioning tasks. To compare the performance with the SOTA models, we conducted a manual assessment of accuracy across different captioning fields using a test set of 1,000 samples. The proposed <a href="https://huggingface.co/Skywork/SkyCaptioner-V1">SkyCaptioner-V1</a> achieves the highest average accuracy among the baseline models, and show a dramatic result in the shot related fields
369
+
370
+ <p align="center">
371
+ <table align="center">
372
+ <thead>
373
+ <tr>
374
+ <th>model</th>
375
+ <th><a href="https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct">Qwen2.5-VL-7B-Ins.</a></th>
376
+ <th><a href="https://huggingface.co/Qwen/Qwen2.5-VL-72B-Instruct">Qwen2.5-VL-72B-Ins.</a></th>
377
+ <th><a href="https://huggingface.co/omni-research/Tarsier2-Recap-7b">Tarsier2-Recap-7b</a></th>
378
+ <th><a href="https://huggingface.co/Skywork/SkyCaptioner-V1">SkyCaptioner-V1</th>
379
+ </tr>
380
+ </thead>
381
+ <tbody>
382
+ <tr>
383
+ <td>Avg accuracy</td>
384
+ <td>51.4%</td>
385
+ <td>58.7%</td>
386
+ <td>49.4%</td>
387
+ <td><strong>76.3%</strong></td>
388
+ </tr>
389
+ <tr>
390
+ <td>shot type</td>
391
+ <td>76.8%</td>
392
+ <td>82.5%</td>
393
+ <td>60.2%</td>
394
+ <td><strong>93.7%</strong></td>
395
+ </tr>
396
+ <tr>
397
+ <td>shot angle</td>
398
+ <td>60.0%</td>
399
+ <td>73.7%</td>
400
+ <td>52.4%</td>
401
+ <td><strong>89.8%</strong></td>
402
+ </tr>
403
+ <tr>
404
+ <td>shot position</td>
405
+ <td>28.4%</td>
406
+ <td>32.7%</td>
407
+ <td>23.6%</td>
408
+ <td><strong>83.1%</strong></td>
409
+ </tr>
410
+ <tr>
411
+ <td>camera motion</td>
412
+ <td>62.0%</td>
413
+ <td>61.2%</td>
414
+ <td>45.3%</td>
415
+ <td><strong>85.3%</strong></td>
416
+ </tr>
417
+ <tr>
418
+ <td>expression</td>
419
+ <td>43.6%</td>
420
+ <td>51.5%</td>
421
+ <td>54.3%</td>
422
+ <td><strong>68.8%</strong></td>
423
+ </tr>
424
+ <tr>
425
+ <td colspan="5" style="text-align: center; border-bottom: 1px solid #ddd; padding: 8px;"></td>
426
+ </tr>
427
+ <tr>
428
+ <td>TYPES_type</td>
429
+ <td>43.5%</td>
430
+ <td>49.7%</td>
431
+ <td>47.6%</td>
432
+ <td><strong>82.5%</strong></td>
433
+ </tr>
434
+ <tr>
435
+ <td>TYPES_sub_type</td>
436
+ <td>38.9%</td>
437
+ <td>44.9%</td>
438
+ <td>45.9%</td>
439
+ <td><strong>75.4%</strong></td>
440
+ </tr>
441
+ <tr>
442
+ <td>appearance</td>
443
+ <td>40.9%</td>
444
+ <td>52.0%</td>
445
+ <td>45.6%</td>
446
+ <td><strong>59.3%</strong></td>
447
+ </tr>
448
+ <tr>
449
+ <td>action</td>
450
+ <td>32.4%</td>
451
+ <td>52.0%</td>
452
+ <td><strong>69.8%</strong></td>
453
+ <td>68.8%</td>
454
+ </tr>
455
+ <tr>
456
+ <td>position</td>
457
+ <td>35.4%</td>
458
+ <td>48.6%</td>
459
+ <td>45.5%</td>
460
+ <td><strong>57.5%</strong></td>
461
+ </tr>
462
+ <tr>
463
+ <td>is_main_subject</td>
464
+ <td>58.5%</td>
465
+ <td>68.7%</td>
466
+ <td>69.7%</td>
467
+ <td><strong>80.9%</strong></td>
468
+ </tr>
469
+ <tr>
470
+ <td>environment</td>
471
+ <td>70.4%</td>
472
+ <td><strong>72.7%</strong></td>
473
+ <td>61.4%</td>
474
+ <td>70.5%</td>
475
+ </tr>
476
+ <tr>
477
+ <td>lighting</td>
478
+ <td>77.1%</td>
479
+ <td><strong>80.0%</strong></td>
480
+ <td>21.2%</td>
481
+ <td>76.5%</td>
482
+ </tr>
483
+ </tbody>
484
+ </table>
485
+ </p>
486
+
487
+ #### Reinforcement Learning
488
+ Inspired by the previous success in LLM, we propose to enhance the performance of the generative model by Reinforcement Learning. Specifically, we focus on the motion quality because we find that the main drawback of our generative model is:
489
+
490
+ - the generative model does not handle well with large, deformable motions.
491
+ - the generated videos may violate the physical law.
492
+
493
+ To avoid the degradation in other metrics, such as text alignment and video quality, we ensure the preference data pairs have comparable text alignment and video quality, while only the motion quality varies. This requirement poses greater challenges in obtaining preference annotations due to the inherently higher costs of human annotation. To address this challenge, we propose a semi-automatic pipeline that strategically combines automatically generated motion pairs and human annotation results. This hybrid approach not only enhances the data scale but also improves alignment with human preferences through curated quality control. Leveraging this enhanced dataset, we first train a specialized reward model to capture the generic motion quality differences between paired samples. This learned reward function subsequently guides the sample selection process for Direct Preference Optimization (DPO), enhancing the motion quality of the generative model.
494
+
495
+ #### Diffusion Forcing
496
+
497
+ We introduce the Diffusion Forcing Transformer to unlock our model’s ability to generate long videos. Diffusion Forcing is a training and sampling strategy where each token is assigned an independent noise level. This allows tokens to be denoised according to arbitrary, per-token schedules. Conceptually, this approach functions as a form of partial masking: a token with zero noise is fully unmasked, while complete noise fully masks it. Diffusion Forcing trains the model to "unmask" any combination of variably noised tokens, using the cleaner tokens as conditional information to guide the recovery of noisy ones. Building on this, our Diffusion Forcing Transformer can extend video generation indefinitely based on the last frames of the previous segment. Note that the synchronous full sequence diffusion is a special case of Diffusion Forcing, where all tokens share the same noise level. This relationship allows us to fine-tune the Diffusion Forcing Transformer from a full-sequence diffusion model.
498
+
499
+ #### High-Quality Supervised Fine-Tuning (SFT)
500
+
501
+ We implement two sequential high-quality supervised fine-tuning (SFT) stages at 540p and 720p resolutions respectively, with the initial SFT phase conducted immediately after pretraining but prior to reinforcement learning (RL) stage.This first-stage SFT serves as a conceptual equilibrium trainer, building upon the foundation model’s pretraining outcomes that utilized only fps24 video data, while strategically removing FPS embedding components to streamline thearchitecture. Trained with the high-quality concept-balanced samples, this phase establishes optimized initialization parameters for subsequent training processes. Following this, we execute a secondary high-resolution SFT at 720p after completing the diffusion forcing stage, incorporating identical loss formulations and the higher-quality concept-balanced datasets by the manually filter. This final refinement phase focuses on resolution increase such that the overall video quality will be further enhanced.
502
+
503
+ ## Performance
504
+
505
+ To comprehensively evaluate our proposed method, we construct the SkyReels-Bench for human assessment and leveraged the open-source <a href="https://github.com/Vchitect/VBench">V-Bench</a> for automated evaluation. This allows us to compare our model with the state-of-the-art (SOTA) baselines, including both open-source and proprietary models.
506
+
507
+ #### Human Evaluation
508
+
509
+ For human evaluation, we design SkyReels-Bench with 1,020 text prompts, systematically assessing three dimensions: Instruction Adherence, Motion Quality, Consistency and Visual Quality. This benchmark is designed to evaluate both text-to-video (T2V) and image-to-video (I2V) generation models, providing comprehensive assessment across different generation paradigms. To ensure fairness, all models were evaluated under default settings with consistent resolutions, and no post-generation filtering was applied.
510
+
511
+ - Text To Video Models
512
+
513
+ <p align="center">
514
+ <table align="center">
515
+ <thead>
516
+ <tr>
517
+ <th>Model Name</th>
518
+ <th>Average</th>
519
+ <th>Instruction Adherence</th>
520
+ <th>Consistency</th>
521
+ <th>Visual Quality</th>
522
+ <th>Motion Quality</th>
523
+ </tr>
524
+ </thead>
525
+ <tbody>
526
+ <tr>
527
+ <td><a href="https://runwayml.com/research/introducing-gen-3-alpha">Runway-Gen3 Alpha</a></td>
528
+ <td>2.53</td>
529
+ <td>2.19</td>
530
+ <td>2.57</td>
531
+ <td>3.23</td>
532
+ <td>2.11</td>
533
+ </tr>
534
+ <tr>
535
+ <td><a href="https://github.com/Tencent/HunyuanVideo">HunyuanVideo-13B</a></td>
536
+ <td>2.82</td>
537
+ <td>2.64</td>
538
+ <td>2.81</td>
539
+ <td>3.20</td>
540
+ <td>2.61</td>
541
+ </tr>
542
+ <tr>
543
+ <td><a href="https://klingai.com">Kling-1.6 STD Mode</a></td>
544
+ <td>2.99</td>
545
+ <td>2.77</td>
546
+ <td>3.05</td>
547
+ <td>3.39</td>
548
+ <td><strong>2.76</strong></td>
549
+ </tr>
550
+ <tr>
551
+ <td><a href="https://hailuoai.video">Hailuo-01</a></td>
552
+ <td>3.0</td>
553
+ <td>2.8</td>
554
+ <td>3.08</td>
555
+ <td>3.29</td>
556
+ <td>2.74</td>
557
+ </tr>
558
+ <tr>
559
+ <td><a href="https://github.com/Wan-Video/Wan2.1">Wan2.1-14B</a></td>
560
+ <td>3.12</td>
561
+ <td>2.91</td>
562
+ <td>3.31</td>
563
+ <td><strong>3.54</strong></td>
564
+ <td>2.71</td>
565
+ </tr>
566
+ <tr>
567
+ <td>SkyReels-V2</td>
568
+ <td><strong>3.14</strong></td>
569
+ <td><strong>3.15</strong></td>
570
+ <td><strong>3.35</strong></td>
571
+ <td>3.34</td>
572
+ <td>2.74</td>
573
+ </tr>
574
+ </tbody>
575
+ </table>
576
+ </p>
577
+
578
+ The evaluation demonstrates that our model achieves significant advancements in **instruction adherence (3.15)** compared to baseline methods, while maintaining competitive performance in **motion quality (2.74)** without sacrificing the **consistency (3.35)**.
579
+
580
+ - Image To Video Models
581
+
582
+ <p align="center">
583
+ <table align="center">
584
+ <thead>
585
+ <tr>
586
+ <th>Model</th>
587
+ <th>Average</th>
588
+ <th>Instruction Adherence</th>
589
+ <th>Consistency</th>
590
+ <th>Visual Quality</th>
591
+ <th>Motion Quality</th>
592
+ </tr>
593
+ </thead>
594
+ <tbody>
595
+ <tr>
596
+ <td><a href="https://github.com/Tencent/HunyuanVideo">HunyuanVideo-13B</a></td>
597
+ <td>2.84</td>
598
+ <td>2.97</td>
599
+ <td>2.95</td>
600
+ <td>2.87</td>
601
+ <td>2.56</td>
602
+ </tr>
603
+ <tr>
604
+ <td><a href="https://github.com/Wan-Video/Wan2.1">Wan2.1-14B</a></td>
605
+ <td>2.85</td>
606
+ <td>3.10</td>
607
+ <td>2.81</td>
608
+ <td>3.00</td>
609
+ <td>2.48</td>
610
+ </tr>
611
+ <tr>
612
+ <td><a href="https://hailuoai.video">Hailuo-01</a></td>
613
+ <td>3.05</td>
614
+ <td>3.31</td>
615
+ <td>2.58</td>
616
+ <td>3.55</td>
617
+ <td>2.74</td>
618
+ </tr>
619
+ <tr>
620
+ <td><a href="https://klingai.com">Kling-1.6 Pro Mode</a></td>
621
+ <td>3.4</td>
622
+ <td>3.56</td>
623
+ <td>3.03</td>
624
+ <td>3.58</td>
625
+ <td>3.41</td>
626
+ </tr>
627
+ <tr>
628
+ <td><a href="https://runwayml.com/research/introducing-runway-gen-4">Runway-Gen4</a></td>
629
+ <td>3.39</td>
630
+ <td>3.75</td>
631
+ <td>3.2</td>
632
+ <td>3.4</td>
633
+ <td>3.37</td>
634
+ </tr>
635
+ <tr>
636
+ <td>SkyReels-V2-DF</td>
637
+ <td>3.24</td>
638
+ <td>3.64</td>
639
+ <td>3.21</td>
640
+ <td>3.18</td>
641
+ <td>2.93</td>
642
+ </tr>
643
+ <tr>
644
+ <td>SkyReels-V2-I2V</td>
645
+ <td>3.29</td>
646
+ <td>3.42</td>
647
+ <td>3.18</td>
648
+ <td>3.56</td>
649
+ <td>3.01</td>
650
+ </tr>
651
+ </tbody>
652
+ </table>
653
+ </p>
654
+
655
+ Our results demonstrate that both **SkyReels-V2-I2V (3.29)** and **SkyReels-V2-DF (3.24)** achieve state-of-the-art performance among open-source models, significantly outperforming HunyuanVideo-13B (2.84) and Wan2.1-14B (2.85) across all quality dimensions. With an average score of 3.29, SkyReels-V2-I2V demonstrates comparable performance to proprietary models Kling-1.6 (3.4) and Runway-Gen4 (3.39).
656
+
657
+
658
+ #### VBench
659
+ To objectively compare SkyReels-V2 Model against other leading open-source Text-To-Video models, we conduct comprehensive evaluations using the public benchmark <a href="https://github.com/Vchitect/VBench">V-Bench</a>. Our evaluation specifically leverages the benchmark’s longer version prompt. For fair comparison with baseline models, we strictly follow their recommended setting for inference.
660
+
661
+ <p align="center">
662
+ <table align="center">
663
+ <thead>
664
+ <tr>
665
+ <th>Model</th>
666
+ <th>Total Score</th>
667
+ <th>Quality Score</th>
668
+ <th>Semantic Score</th>
669
+ </tr>
670
+ </thead>
671
+ <tbody>
672
+ <tr>
673
+ <td><a href="https://github.com/hpcaitech/Open-Sora">OpenSora 2.0</a></td>
674
+ <td>81.5 %</td>
675
+ <td>82.1 %</td>
676
+ <td>78.2 %</td>
677
+ </tr>
678
+ <tr>
679
+ <td><a href="https://github.com/THUDM/CogVideo">CogVideoX1.5-5B</a></td>
680
+ <td>80.3 %</td>
681
+ <td>80.9 %</td>
682
+ <td>77.9 %</td>
683
+ </tr>
684
+ <tr>
685
+ <td><a href="https://github.com/Tencent/HunyuanVideo">HunyuanVideo-13B</a></td>
686
+ <td>82.7 %</td>
687
+ <td>84.4 %</td>
688
+ <td>76.2 %</td>
689
+ </tr>
690
+ <tr>
691
+ <td><a href="https://github.com/Wan-Video/Wan2.1">Wan2.1-14B</a></td>
692
+ <td>83.7 %</td>
693
+ <td>84.2 %</td>
694
+ <td><strong>81.4 %</strong></td>
695
+ </tr>
696
+ <tr>
697
+ <td>SkyReels-V2</td>
698
+ <td><strong>83.9 %</strong></td>
699
+ <td><strong>84.7 %</strong></td>
700
+ <td>80.8 %</td>
701
+ </tr>
702
+ </tbody>
703
+ </table>
704
+ </p>
705
+
706
+ The VBench results demonstrate that SkyReels-V2 outperforms all compared models including HunyuanVideo-13B and Wan2.1-14B, With the highest **total score (83.9%)** and **quality score (84.7%)**. In this evaluation, the semantic score is slightly lower than Wan2.1-14B, while we outperform Wan2.1-14B in human evaluations, with the primary gap attributed to V-Bench’s insufficient evaluation of shot-scenario semantic adherence.
707
+
708
+ ## Acknowledgements
709
+ We would like to thank the contributors of <a href="https://github.com/Wan-Video/Wan2.1">Wan 2.1</a>, <a href="https://github.com/xdit-project/xDiT">XDit</a> and <a href="https://qwenlm.github.io/blog/qwen2.5/">Qwen 2.5</a> repositories, for their open research and contributions.
710
+
711
+ ## Citation
712
+
713
+ ```bibtex
714
+ @misc{chen2025skyreelsv2infinitelengthfilmgenerative,
715
+ title={SkyReels-V2: Infinite-length Film Generative Model},
716
+ author={Guibin Chen and Dixuan Lin and Jiangping Yang and Chunze Lin and Junchen Zhu and Mingyuan Fan and Hao Zhang and Sheng Chen and Zheng Chen and Chengcheng Ma and Weiming Xiong and Wei Wang and Nuo Pang and Kang Kang and Zhiheng Xu and Yuzhe Jin and Yupeng Liang and Yubing Song and Peng Zhao and Boyuan Xu and Di Qiu and Debang Li and Zhengcong Fei and Yang Li and Yahui Zhou},
717
+ year={2025},
718
+ eprint={2504.13074},
719
+ archivePrefix={arXiv},
720
+ primaryClass={cs.CV},
721
+ url={https://arxiv.org/abs/2504.13074},
722
+ }
723
+ ```
Wan2.1_VAE.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38071ab59bd94681c686fa51d75a1968f64e470262043be31f7a094e442fd981
3
+ size 507609880
assets/logo2.png ADDED
assets/main_pipeline.jpg ADDED

Git LFS Details

  • SHA256: e8fd982dd51a3edd0a1ce451b391526c1a51ea94ae68f5ed79380173dbcce7fb
  • Pointer size: 131 Bytes
  • Size of remote file: 183 kB
config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.29.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "inject_sample_info": false,
10
+ "model_type": "t2v",
11
+ "num_heads": 40,
12
+ "num_layers": 40,
13
+ "out_dim": 16,
14
+ "text_len": 512
15
+ }
diffusion_pytorch_model-00001-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8854859dccac2b6f52a7df4546ba527fc94dab8cf69a0a690cc5b7f1af3ebded
3
+ size 9992481544
diffusion_pytorch_model-00002-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e6d65dffc8e737a59c5a97748f8b2813e89e10c7cdc731371c80c70c9af5f70
3
+ size 9943937936
diffusion_pytorch_model-00003-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a8180edba11a344ac1e4d8f3b00c4b2ce3d5803239024483614e807f7340e24
3
+ size 9943979184
diffusion_pytorch_model-00004-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eed67acbbaa98237af3ed5f64efe3ebbc5c715786af3c01e1118809fa6b582af
3
+ size 9839059744
diffusion_pytorch_model-00005-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41471752ff72b3a74d948286e6f03134139358e686bfd91c3ef0cf2290b5b065
3
+ size 9839059744
diffusion_pytorch_model-00006-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24180a105b65a77c08550de4488bfbea6cbce19d906a63bd2d7607c3217b456f
3
+ size 7595559224
diffusion_pytorch_model.safetensors.index.json ADDED
@@ -0,0 +1,1102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 57153966336
4
+ },
5
+ "weight_map": {
6
+ "blocks.0.cross_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
7
+ "blocks.0.cross_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
8
+ "blocks.0.cross_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
9
+ "blocks.0.cross_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
10
+ "blocks.0.cross_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
11
+ "blocks.0.cross_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
12
+ "blocks.0.cross_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
13
+ "blocks.0.cross_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
14
+ "blocks.0.cross_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
15
+ "blocks.0.cross_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
16
+ "blocks.0.ffn.0.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
17
+ "blocks.0.ffn.0.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
18
+ "blocks.0.ffn.2.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
19
+ "blocks.0.ffn.2.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
20
+ "blocks.0.modulation": "diffusion_pytorch_model-00001-of-00006.safetensors",
21
+ "blocks.0.norm3.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
22
+ "blocks.0.norm3.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
23
+ "blocks.0.self_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
24
+ "blocks.0.self_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
25
+ "blocks.0.self_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
26
+ "blocks.0.self_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
27
+ "blocks.0.self_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
28
+ "blocks.0.self_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
29
+ "blocks.0.self_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
30
+ "blocks.0.self_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
31
+ "blocks.0.self_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
32
+ "blocks.0.self_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
33
+ "blocks.1.cross_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
34
+ "blocks.1.cross_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
35
+ "blocks.1.cross_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
36
+ "blocks.1.cross_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
37
+ "blocks.1.cross_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
38
+ "blocks.1.cross_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
39
+ "blocks.1.cross_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
40
+ "blocks.1.cross_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
41
+ "blocks.1.cross_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
42
+ "blocks.1.cross_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
43
+ "blocks.1.ffn.0.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
44
+ "blocks.1.ffn.0.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
45
+ "blocks.1.ffn.2.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
46
+ "blocks.1.ffn.2.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
47
+ "blocks.1.modulation": "diffusion_pytorch_model-00001-of-00006.safetensors",
48
+ "blocks.1.norm3.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
49
+ "blocks.1.norm3.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
50
+ "blocks.1.self_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
51
+ "blocks.1.self_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
52
+ "blocks.1.self_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
53
+ "blocks.1.self_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
54
+ "blocks.1.self_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
55
+ "blocks.1.self_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
56
+ "blocks.1.self_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
57
+ "blocks.1.self_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
58
+ "blocks.1.self_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
59
+ "blocks.1.self_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
60
+ "blocks.10.cross_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
61
+ "blocks.10.cross_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
62
+ "blocks.10.cross_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
63
+ "blocks.10.cross_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
64
+ "blocks.10.cross_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
65
+ "blocks.10.cross_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
66
+ "blocks.10.cross_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
67
+ "blocks.10.cross_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
68
+ "blocks.10.cross_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
69
+ "blocks.10.cross_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
70
+ "blocks.10.ffn.0.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
71
+ "blocks.10.ffn.0.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
72
+ "blocks.10.ffn.2.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
73
+ "blocks.10.ffn.2.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
74
+ "blocks.10.modulation": "diffusion_pytorch_model-00002-of-00006.safetensors",
75
+ "blocks.10.norm3.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
76
+ "blocks.10.norm3.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
77
+ "blocks.10.self_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
78
+ "blocks.10.self_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
79
+ "blocks.10.self_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
80
+ "blocks.10.self_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
81
+ "blocks.10.self_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
82
+ "blocks.10.self_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
83
+ "blocks.10.self_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
84
+ "blocks.10.self_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
85
+ "blocks.10.self_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
86
+ "blocks.10.self_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
87
+ "blocks.11.cross_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
88
+ "blocks.11.cross_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
89
+ "blocks.11.cross_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
90
+ "blocks.11.cross_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
91
+ "blocks.11.cross_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
92
+ "blocks.11.cross_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
93
+ "blocks.11.cross_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
94
+ "blocks.11.cross_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
95
+ "blocks.11.cross_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
96
+ "blocks.11.cross_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
97
+ "blocks.11.ffn.0.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
98
+ "blocks.11.ffn.0.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
99
+ "blocks.11.ffn.2.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
100
+ "blocks.11.ffn.2.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
101
+ "blocks.11.modulation": "diffusion_pytorch_model-00002-of-00006.safetensors",
102
+ "blocks.11.norm3.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
103
+ "blocks.11.norm3.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
104
+ "blocks.11.self_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
105
+ "blocks.11.self_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
106
+ "blocks.11.self_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
107
+ "blocks.11.self_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
108
+ "blocks.11.self_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
109
+ "blocks.11.self_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
110
+ "blocks.11.self_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
111
+ "blocks.11.self_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
112
+ "blocks.11.self_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
113
+ "blocks.11.self_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
114
+ "blocks.12.cross_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
115
+ "blocks.12.cross_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
116
+ "blocks.12.cross_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
117
+ "blocks.12.cross_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
118
+ "blocks.12.cross_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
119
+ "blocks.12.cross_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
120
+ "blocks.12.cross_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
121
+ "blocks.12.cross_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
122
+ "blocks.12.cross_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
123
+ "blocks.12.cross_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
124
+ "blocks.12.ffn.0.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
125
+ "blocks.12.ffn.0.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
126
+ "blocks.12.ffn.2.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
127
+ "blocks.12.ffn.2.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
128
+ "blocks.12.modulation": "diffusion_pytorch_model-00002-of-00006.safetensors",
129
+ "blocks.12.norm3.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
130
+ "blocks.12.norm3.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
131
+ "blocks.12.self_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
132
+ "blocks.12.self_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
133
+ "blocks.12.self_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
134
+ "blocks.12.self_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
135
+ "blocks.12.self_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
136
+ "blocks.12.self_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
137
+ "blocks.12.self_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
138
+ "blocks.12.self_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
139
+ "blocks.12.self_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
140
+ "blocks.12.self_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
141
+ "blocks.13.cross_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
142
+ "blocks.13.cross_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
143
+ "blocks.13.cross_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
144
+ "blocks.13.cross_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
145
+ "blocks.13.cross_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
146
+ "blocks.13.cross_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
147
+ "blocks.13.cross_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
148
+ "blocks.13.cross_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
149
+ "blocks.13.cross_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
150
+ "blocks.13.cross_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
151
+ "blocks.13.ffn.0.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
152
+ "blocks.13.ffn.0.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
153
+ "blocks.13.ffn.2.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
154
+ "blocks.13.ffn.2.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
155
+ "blocks.13.modulation": "diffusion_pytorch_model-00002-of-00006.safetensors",
156
+ "blocks.13.norm3.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
157
+ "blocks.13.norm3.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
158
+ "blocks.13.self_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
159
+ "blocks.13.self_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
160
+ "blocks.13.self_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
161
+ "blocks.13.self_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
162
+ "blocks.13.self_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
163
+ "blocks.13.self_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
164
+ "blocks.13.self_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
165
+ "blocks.13.self_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
166
+ "blocks.13.self_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
167
+ "blocks.13.self_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
168
+ "blocks.14.cross_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
169
+ "blocks.14.cross_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
170
+ "blocks.14.cross_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
171
+ "blocks.14.cross_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
172
+ "blocks.14.cross_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
173
+ "blocks.14.cross_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
174
+ "blocks.14.cross_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
175
+ "blocks.14.cross_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
176
+ "blocks.14.cross_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
177
+ "blocks.14.cross_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
178
+ "blocks.14.ffn.0.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
179
+ "blocks.14.ffn.0.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
180
+ "blocks.14.ffn.2.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
181
+ "blocks.14.ffn.2.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
182
+ "blocks.14.modulation": "diffusion_pytorch_model-00003-of-00006.safetensors",
183
+ "blocks.14.norm3.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
184
+ "blocks.14.norm3.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
185
+ "blocks.14.self_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
186
+ "blocks.14.self_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
187
+ "blocks.14.self_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
188
+ "blocks.14.self_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
189
+ "blocks.14.self_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
190
+ "blocks.14.self_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
191
+ "blocks.14.self_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
192
+ "blocks.14.self_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
193
+ "blocks.14.self_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
194
+ "blocks.14.self_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
195
+ "blocks.15.cross_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
196
+ "blocks.15.cross_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
197
+ "blocks.15.cross_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
198
+ "blocks.15.cross_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
199
+ "blocks.15.cross_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
200
+ "blocks.15.cross_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
201
+ "blocks.15.cross_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
202
+ "blocks.15.cross_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
203
+ "blocks.15.cross_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
204
+ "blocks.15.cross_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
205
+ "blocks.15.ffn.0.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
206
+ "blocks.15.ffn.0.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
207
+ "blocks.15.ffn.2.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
208
+ "blocks.15.ffn.2.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
209
+ "blocks.15.modulation": "diffusion_pytorch_model-00003-of-00006.safetensors",
210
+ "blocks.15.norm3.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
211
+ "blocks.15.norm3.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
212
+ "blocks.15.self_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
213
+ "blocks.15.self_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
214
+ "blocks.15.self_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
215
+ "blocks.15.self_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
216
+ "blocks.15.self_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
217
+ "blocks.15.self_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
218
+ "blocks.15.self_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
219
+ "blocks.15.self_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
220
+ "blocks.15.self_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
221
+ "blocks.15.self_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
222
+ "blocks.16.cross_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
223
+ "blocks.16.cross_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
224
+ "blocks.16.cross_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
225
+ "blocks.16.cross_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
226
+ "blocks.16.cross_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
227
+ "blocks.16.cross_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
228
+ "blocks.16.cross_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
229
+ "blocks.16.cross_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
230
+ "blocks.16.cross_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
231
+ "blocks.16.cross_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
232
+ "blocks.16.ffn.0.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
233
+ "blocks.16.ffn.0.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
234
+ "blocks.16.ffn.2.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
235
+ "blocks.16.ffn.2.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
236
+ "blocks.16.modulation": "diffusion_pytorch_model-00003-of-00006.safetensors",
237
+ "blocks.16.norm3.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
238
+ "blocks.16.norm3.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
239
+ "blocks.16.self_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
240
+ "blocks.16.self_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
241
+ "blocks.16.self_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
242
+ "blocks.16.self_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
243
+ "blocks.16.self_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
244
+ "blocks.16.self_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
245
+ "blocks.16.self_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
246
+ "blocks.16.self_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
247
+ "blocks.16.self_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
248
+ "blocks.16.self_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
249
+ "blocks.17.cross_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
250
+ "blocks.17.cross_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
251
+ "blocks.17.cross_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
252
+ "blocks.17.cross_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
253
+ "blocks.17.cross_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
254
+ "blocks.17.cross_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
255
+ "blocks.17.cross_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
256
+ "blocks.17.cross_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
257
+ "blocks.17.cross_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
258
+ "blocks.17.cross_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
259
+ "blocks.17.ffn.0.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
260
+ "blocks.17.ffn.0.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
261
+ "blocks.17.ffn.2.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
262
+ "blocks.17.ffn.2.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
263
+ "blocks.17.modulation": "diffusion_pytorch_model-00003-of-00006.safetensors",
264
+ "blocks.17.norm3.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
265
+ "blocks.17.norm3.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
266
+ "blocks.17.self_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
267
+ "blocks.17.self_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
268
+ "blocks.17.self_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
269
+ "blocks.17.self_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
270
+ "blocks.17.self_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
271
+ "blocks.17.self_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
272
+ "blocks.17.self_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
273
+ "blocks.17.self_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
274
+ "blocks.17.self_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
275
+ "blocks.17.self_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
276
+ "blocks.18.cross_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
277
+ "blocks.18.cross_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
278
+ "blocks.18.cross_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
279
+ "blocks.18.cross_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
280
+ "blocks.18.cross_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
281
+ "blocks.18.cross_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
282
+ "blocks.18.cross_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
283
+ "blocks.18.cross_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
284
+ "blocks.18.cross_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
285
+ "blocks.18.cross_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
286
+ "blocks.18.ffn.0.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
287
+ "blocks.18.ffn.0.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
288
+ "blocks.18.ffn.2.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
289
+ "blocks.18.ffn.2.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
290
+ "blocks.18.modulation": "diffusion_pytorch_model-00003-of-00006.safetensors",
291
+ "blocks.18.norm3.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
292
+ "blocks.18.norm3.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
293
+ "blocks.18.self_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
294
+ "blocks.18.self_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
295
+ "blocks.18.self_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
296
+ "blocks.18.self_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
297
+ "blocks.18.self_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
298
+ "blocks.18.self_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
299
+ "blocks.18.self_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
300
+ "blocks.18.self_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
301
+ "blocks.18.self_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
302
+ "blocks.18.self_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
303
+ "blocks.19.cross_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
304
+ "blocks.19.cross_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
305
+ "blocks.19.cross_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
306
+ "blocks.19.cross_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
307
+ "blocks.19.cross_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
308
+ "blocks.19.cross_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
309
+ "blocks.19.cross_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
310
+ "blocks.19.cross_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
311
+ "blocks.19.cross_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
312
+ "blocks.19.cross_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
313
+ "blocks.19.ffn.0.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
314
+ "blocks.19.ffn.0.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
315
+ "blocks.19.ffn.2.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
316
+ "blocks.19.ffn.2.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
317
+ "blocks.19.modulation": "diffusion_pytorch_model-00003-of-00006.safetensors",
318
+ "blocks.19.norm3.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
319
+ "blocks.19.norm3.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
320
+ "blocks.19.self_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
321
+ "blocks.19.self_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
322
+ "blocks.19.self_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
323
+ "blocks.19.self_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
324
+ "blocks.19.self_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
325
+ "blocks.19.self_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
326
+ "blocks.19.self_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
327
+ "blocks.19.self_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
328
+ "blocks.19.self_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
329
+ "blocks.19.self_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
330
+ "blocks.2.cross_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
331
+ "blocks.2.cross_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
332
+ "blocks.2.cross_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
333
+ "blocks.2.cross_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
334
+ "blocks.2.cross_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
335
+ "blocks.2.cross_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
336
+ "blocks.2.cross_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
337
+ "blocks.2.cross_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
338
+ "blocks.2.cross_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
339
+ "blocks.2.cross_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
340
+ "blocks.2.ffn.0.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
341
+ "blocks.2.ffn.0.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
342
+ "blocks.2.ffn.2.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
343
+ "blocks.2.ffn.2.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
344
+ "blocks.2.modulation": "diffusion_pytorch_model-00001-of-00006.safetensors",
345
+ "blocks.2.norm3.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
346
+ "blocks.2.norm3.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
347
+ "blocks.2.self_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
348
+ "blocks.2.self_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
349
+ "blocks.2.self_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
350
+ "blocks.2.self_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
351
+ "blocks.2.self_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
352
+ "blocks.2.self_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
353
+ "blocks.2.self_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
354
+ "blocks.2.self_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
355
+ "blocks.2.self_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
356
+ "blocks.2.self_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
357
+ "blocks.20.cross_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
358
+ "blocks.20.cross_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
359
+ "blocks.20.cross_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
360
+ "blocks.20.cross_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
361
+ "blocks.20.cross_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
362
+ "blocks.20.cross_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
363
+ "blocks.20.cross_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
364
+ "blocks.20.cross_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
365
+ "blocks.20.cross_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
366
+ "blocks.20.cross_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
367
+ "blocks.20.ffn.0.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
368
+ "blocks.20.ffn.0.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
369
+ "blocks.20.ffn.2.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
370
+ "blocks.20.ffn.2.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
371
+ "blocks.20.modulation": "diffusion_pytorch_model-00003-of-00006.safetensors",
372
+ "blocks.20.norm3.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
373
+ "blocks.20.norm3.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
374
+ "blocks.20.self_attn.k.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
375
+ "blocks.20.self_attn.k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
376
+ "blocks.20.self_attn.norm_k.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
377
+ "blocks.20.self_attn.norm_q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
378
+ "blocks.20.self_attn.o.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
379
+ "blocks.20.self_attn.o.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
380
+ "blocks.20.self_attn.q.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
381
+ "blocks.20.self_attn.q.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
382
+ "blocks.20.self_attn.v.bias": "diffusion_pytorch_model-00003-of-00006.safetensors",
383
+ "blocks.20.self_attn.v.weight": "diffusion_pytorch_model-00003-of-00006.safetensors",
384
+ "blocks.21.cross_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
385
+ "blocks.21.cross_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
386
+ "blocks.21.cross_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
387
+ "blocks.21.cross_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
388
+ "blocks.21.cross_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
389
+ "blocks.21.cross_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
390
+ "blocks.21.cross_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
391
+ "blocks.21.cross_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
392
+ "blocks.21.cross_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
393
+ "blocks.21.cross_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
394
+ "blocks.21.ffn.0.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
395
+ "blocks.21.ffn.0.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
396
+ "blocks.21.ffn.2.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
397
+ "blocks.21.ffn.2.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
398
+ "blocks.21.modulation": "diffusion_pytorch_model-00004-of-00006.safetensors",
399
+ "blocks.21.norm3.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
400
+ "blocks.21.norm3.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
401
+ "blocks.21.self_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
402
+ "blocks.21.self_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
403
+ "blocks.21.self_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
404
+ "blocks.21.self_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
405
+ "blocks.21.self_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
406
+ "blocks.21.self_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
407
+ "blocks.21.self_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
408
+ "blocks.21.self_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
409
+ "blocks.21.self_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
410
+ "blocks.21.self_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
411
+ "blocks.22.cross_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
412
+ "blocks.22.cross_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
413
+ "blocks.22.cross_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
414
+ "blocks.22.cross_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
415
+ "blocks.22.cross_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
416
+ "blocks.22.cross_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
417
+ "blocks.22.cross_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
418
+ "blocks.22.cross_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
419
+ "blocks.22.cross_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
420
+ "blocks.22.cross_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
421
+ "blocks.22.ffn.0.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
422
+ "blocks.22.ffn.0.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
423
+ "blocks.22.ffn.2.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
424
+ "blocks.22.ffn.2.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
425
+ "blocks.22.modulation": "diffusion_pytorch_model-00004-of-00006.safetensors",
426
+ "blocks.22.norm3.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
427
+ "blocks.22.norm3.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
428
+ "blocks.22.self_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
429
+ "blocks.22.self_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
430
+ "blocks.22.self_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
431
+ "blocks.22.self_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
432
+ "blocks.22.self_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
433
+ "blocks.22.self_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
434
+ "blocks.22.self_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
435
+ "blocks.22.self_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
436
+ "blocks.22.self_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
437
+ "blocks.22.self_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
438
+ "blocks.23.cross_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
439
+ "blocks.23.cross_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
440
+ "blocks.23.cross_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
441
+ "blocks.23.cross_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
442
+ "blocks.23.cross_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
443
+ "blocks.23.cross_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
444
+ "blocks.23.cross_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
445
+ "blocks.23.cross_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
446
+ "blocks.23.cross_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
447
+ "blocks.23.cross_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
448
+ "blocks.23.ffn.0.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
449
+ "blocks.23.ffn.0.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
450
+ "blocks.23.ffn.2.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
451
+ "blocks.23.ffn.2.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
452
+ "blocks.23.modulation": "diffusion_pytorch_model-00004-of-00006.safetensors",
453
+ "blocks.23.norm3.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
454
+ "blocks.23.norm3.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
455
+ "blocks.23.self_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
456
+ "blocks.23.self_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
457
+ "blocks.23.self_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
458
+ "blocks.23.self_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
459
+ "blocks.23.self_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
460
+ "blocks.23.self_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
461
+ "blocks.23.self_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
462
+ "blocks.23.self_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
463
+ "blocks.23.self_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
464
+ "blocks.23.self_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
465
+ "blocks.24.cross_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
466
+ "blocks.24.cross_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
467
+ "blocks.24.cross_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
468
+ "blocks.24.cross_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
469
+ "blocks.24.cross_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
470
+ "blocks.24.cross_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
471
+ "blocks.24.cross_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
472
+ "blocks.24.cross_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
473
+ "blocks.24.cross_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
474
+ "blocks.24.cross_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
475
+ "blocks.24.ffn.0.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
476
+ "blocks.24.ffn.0.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
477
+ "blocks.24.ffn.2.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
478
+ "blocks.24.ffn.2.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
479
+ "blocks.24.modulation": "diffusion_pytorch_model-00004-of-00006.safetensors",
480
+ "blocks.24.norm3.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
481
+ "blocks.24.norm3.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
482
+ "blocks.24.self_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
483
+ "blocks.24.self_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
484
+ "blocks.24.self_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
485
+ "blocks.24.self_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
486
+ "blocks.24.self_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
487
+ "blocks.24.self_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
488
+ "blocks.24.self_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
489
+ "blocks.24.self_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
490
+ "blocks.24.self_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
491
+ "blocks.24.self_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
492
+ "blocks.25.cross_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
493
+ "blocks.25.cross_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
494
+ "blocks.25.cross_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
495
+ "blocks.25.cross_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
496
+ "blocks.25.cross_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
497
+ "blocks.25.cross_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
498
+ "blocks.25.cross_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
499
+ "blocks.25.cross_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
500
+ "blocks.25.cross_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
501
+ "blocks.25.cross_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
502
+ "blocks.25.ffn.0.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
503
+ "blocks.25.ffn.0.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
504
+ "blocks.25.ffn.2.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
505
+ "blocks.25.ffn.2.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
506
+ "blocks.25.modulation": "diffusion_pytorch_model-00004-of-00006.safetensors",
507
+ "blocks.25.norm3.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
508
+ "blocks.25.norm3.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
509
+ "blocks.25.self_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
510
+ "blocks.25.self_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
511
+ "blocks.25.self_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
512
+ "blocks.25.self_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
513
+ "blocks.25.self_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
514
+ "blocks.25.self_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
515
+ "blocks.25.self_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
516
+ "blocks.25.self_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
517
+ "blocks.25.self_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
518
+ "blocks.25.self_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
519
+ "blocks.26.cross_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
520
+ "blocks.26.cross_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
521
+ "blocks.26.cross_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
522
+ "blocks.26.cross_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
523
+ "blocks.26.cross_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
524
+ "blocks.26.cross_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
525
+ "blocks.26.cross_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
526
+ "blocks.26.cross_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
527
+ "blocks.26.cross_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
528
+ "blocks.26.cross_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
529
+ "blocks.26.ffn.0.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
530
+ "blocks.26.ffn.0.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
531
+ "blocks.26.ffn.2.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
532
+ "blocks.26.ffn.2.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
533
+ "blocks.26.modulation": "diffusion_pytorch_model-00004-of-00006.safetensors",
534
+ "blocks.26.norm3.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
535
+ "blocks.26.norm3.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
536
+ "blocks.26.self_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
537
+ "blocks.26.self_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
538
+ "blocks.26.self_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
539
+ "blocks.26.self_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
540
+ "blocks.26.self_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
541
+ "blocks.26.self_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
542
+ "blocks.26.self_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
543
+ "blocks.26.self_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
544
+ "blocks.26.self_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
545
+ "blocks.26.self_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
546
+ "blocks.27.cross_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
547
+ "blocks.27.cross_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
548
+ "blocks.27.cross_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
549
+ "blocks.27.cross_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
550
+ "blocks.27.cross_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
551
+ "blocks.27.cross_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
552
+ "blocks.27.cross_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
553
+ "blocks.27.cross_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
554
+ "blocks.27.cross_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
555
+ "blocks.27.cross_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
556
+ "blocks.27.ffn.0.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
557
+ "blocks.27.ffn.0.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
558
+ "blocks.27.ffn.2.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
559
+ "blocks.27.ffn.2.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
560
+ "blocks.27.modulation": "diffusion_pytorch_model-00004-of-00006.safetensors",
561
+ "blocks.27.norm3.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
562
+ "blocks.27.norm3.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
563
+ "blocks.27.self_attn.k.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
564
+ "blocks.27.self_attn.k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
565
+ "blocks.27.self_attn.norm_k.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
566
+ "blocks.27.self_attn.norm_q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
567
+ "blocks.27.self_attn.o.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
568
+ "blocks.27.self_attn.o.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
569
+ "blocks.27.self_attn.q.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
570
+ "blocks.27.self_attn.q.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
571
+ "blocks.27.self_attn.v.bias": "diffusion_pytorch_model-00004-of-00006.safetensors",
572
+ "blocks.27.self_attn.v.weight": "diffusion_pytorch_model-00004-of-00006.safetensors",
573
+ "blocks.28.cross_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
574
+ "blocks.28.cross_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
575
+ "blocks.28.cross_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
576
+ "blocks.28.cross_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
577
+ "blocks.28.cross_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
578
+ "blocks.28.cross_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
579
+ "blocks.28.cross_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
580
+ "blocks.28.cross_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
581
+ "blocks.28.cross_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
582
+ "blocks.28.cross_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
583
+ "blocks.28.ffn.0.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
584
+ "blocks.28.ffn.0.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
585
+ "blocks.28.ffn.2.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
586
+ "blocks.28.ffn.2.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
587
+ "blocks.28.modulation": "diffusion_pytorch_model-00005-of-00006.safetensors",
588
+ "blocks.28.norm3.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
589
+ "blocks.28.norm3.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
590
+ "blocks.28.self_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
591
+ "blocks.28.self_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
592
+ "blocks.28.self_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
593
+ "blocks.28.self_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
594
+ "blocks.28.self_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
595
+ "blocks.28.self_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
596
+ "blocks.28.self_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
597
+ "blocks.28.self_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
598
+ "blocks.28.self_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
599
+ "blocks.28.self_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
600
+ "blocks.29.cross_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
601
+ "blocks.29.cross_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
602
+ "blocks.29.cross_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
603
+ "blocks.29.cross_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
604
+ "blocks.29.cross_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
605
+ "blocks.29.cross_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
606
+ "blocks.29.cross_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
607
+ "blocks.29.cross_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
608
+ "blocks.29.cross_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
609
+ "blocks.29.cross_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
610
+ "blocks.29.ffn.0.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
611
+ "blocks.29.ffn.0.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
612
+ "blocks.29.ffn.2.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
613
+ "blocks.29.ffn.2.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
614
+ "blocks.29.modulation": "diffusion_pytorch_model-00005-of-00006.safetensors",
615
+ "blocks.29.norm3.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
616
+ "blocks.29.norm3.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
617
+ "blocks.29.self_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
618
+ "blocks.29.self_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
619
+ "blocks.29.self_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
620
+ "blocks.29.self_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
621
+ "blocks.29.self_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
622
+ "blocks.29.self_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
623
+ "blocks.29.self_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
624
+ "blocks.29.self_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
625
+ "blocks.29.self_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
626
+ "blocks.29.self_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
627
+ "blocks.3.cross_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
628
+ "blocks.3.cross_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
629
+ "blocks.3.cross_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
630
+ "blocks.3.cross_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
631
+ "blocks.3.cross_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
632
+ "blocks.3.cross_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
633
+ "blocks.3.cross_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
634
+ "blocks.3.cross_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
635
+ "blocks.3.cross_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
636
+ "blocks.3.cross_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
637
+ "blocks.3.ffn.0.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
638
+ "blocks.3.ffn.0.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
639
+ "blocks.3.ffn.2.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
640
+ "blocks.3.ffn.2.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
641
+ "blocks.3.modulation": "diffusion_pytorch_model-00001-of-00006.safetensors",
642
+ "blocks.3.norm3.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
643
+ "blocks.3.norm3.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
644
+ "blocks.3.self_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
645
+ "blocks.3.self_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
646
+ "blocks.3.self_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
647
+ "blocks.3.self_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
648
+ "blocks.3.self_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
649
+ "blocks.3.self_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
650
+ "blocks.3.self_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
651
+ "blocks.3.self_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
652
+ "blocks.3.self_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
653
+ "blocks.3.self_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
654
+ "blocks.30.cross_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
655
+ "blocks.30.cross_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
656
+ "blocks.30.cross_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
657
+ "blocks.30.cross_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
658
+ "blocks.30.cross_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
659
+ "blocks.30.cross_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
660
+ "blocks.30.cross_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
661
+ "blocks.30.cross_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
662
+ "blocks.30.cross_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
663
+ "blocks.30.cross_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
664
+ "blocks.30.ffn.0.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
665
+ "blocks.30.ffn.0.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
666
+ "blocks.30.ffn.2.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
667
+ "blocks.30.ffn.2.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
668
+ "blocks.30.modulation": "diffusion_pytorch_model-00005-of-00006.safetensors",
669
+ "blocks.30.norm3.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
670
+ "blocks.30.norm3.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
671
+ "blocks.30.self_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
672
+ "blocks.30.self_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
673
+ "blocks.30.self_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
674
+ "blocks.30.self_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
675
+ "blocks.30.self_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
676
+ "blocks.30.self_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
677
+ "blocks.30.self_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
678
+ "blocks.30.self_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
679
+ "blocks.30.self_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
680
+ "blocks.30.self_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
681
+ "blocks.31.cross_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
682
+ "blocks.31.cross_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
683
+ "blocks.31.cross_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
684
+ "blocks.31.cross_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
685
+ "blocks.31.cross_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
686
+ "blocks.31.cross_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
687
+ "blocks.31.cross_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
688
+ "blocks.31.cross_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
689
+ "blocks.31.cross_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
690
+ "blocks.31.cross_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
691
+ "blocks.31.ffn.0.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
692
+ "blocks.31.ffn.0.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
693
+ "blocks.31.ffn.2.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
694
+ "blocks.31.ffn.2.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
695
+ "blocks.31.modulation": "diffusion_pytorch_model-00005-of-00006.safetensors",
696
+ "blocks.31.norm3.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
697
+ "blocks.31.norm3.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
698
+ "blocks.31.self_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
699
+ "blocks.31.self_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
700
+ "blocks.31.self_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
701
+ "blocks.31.self_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
702
+ "blocks.31.self_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
703
+ "blocks.31.self_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
704
+ "blocks.31.self_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
705
+ "blocks.31.self_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
706
+ "blocks.31.self_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
707
+ "blocks.31.self_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
708
+ "blocks.32.cross_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
709
+ "blocks.32.cross_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
710
+ "blocks.32.cross_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
711
+ "blocks.32.cross_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
712
+ "blocks.32.cross_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
713
+ "blocks.32.cross_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
714
+ "blocks.32.cross_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
715
+ "blocks.32.cross_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
716
+ "blocks.32.cross_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
717
+ "blocks.32.cross_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
718
+ "blocks.32.ffn.0.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
719
+ "blocks.32.ffn.0.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
720
+ "blocks.32.ffn.2.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
721
+ "blocks.32.ffn.2.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
722
+ "blocks.32.modulation": "diffusion_pytorch_model-00005-of-00006.safetensors",
723
+ "blocks.32.norm3.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
724
+ "blocks.32.norm3.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
725
+ "blocks.32.self_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
726
+ "blocks.32.self_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
727
+ "blocks.32.self_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
728
+ "blocks.32.self_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
729
+ "blocks.32.self_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
730
+ "blocks.32.self_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
731
+ "blocks.32.self_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
732
+ "blocks.32.self_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
733
+ "blocks.32.self_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
734
+ "blocks.32.self_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
735
+ "blocks.33.cross_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
736
+ "blocks.33.cross_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
737
+ "blocks.33.cross_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
738
+ "blocks.33.cross_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
739
+ "blocks.33.cross_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
740
+ "blocks.33.cross_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
741
+ "blocks.33.cross_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
742
+ "blocks.33.cross_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
743
+ "blocks.33.cross_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
744
+ "blocks.33.cross_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
745
+ "blocks.33.ffn.0.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
746
+ "blocks.33.ffn.0.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
747
+ "blocks.33.ffn.2.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
748
+ "blocks.33.ffn.2.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
749
+ "blocks.33.modulation": "diffusion_pytorch_model-00005-of-00006.safetensors",
750
+ "blocks.33.norm3.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
751
+ "blocks.33.norm3.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
752
+ "blocks.33.self_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
753
+ "blocks.33.self_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
754
+ "blocks.33.self_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
755
+ "blocks.33.self_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
756
+ "blocks.33.self_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
757
+ "blocks.33.self_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
758
+ "blocks.33.self_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
759
+ "blocks.33.self_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
760
+ "blocks.33.self_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
761
+ "blocks.33.self_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
762
+ "blocks.34.cross_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
763
+ "blocks.34.cross_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
764
+ "blocks.34.cross_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
765
+ "blocks.34.cross_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
766
+ "blocks.34.cross_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
767
+ "blocks.34.cross_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
768
+ "blocks.34.cross_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
769
+ "blocks.34.cross_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
770
+ "blocks.34.cross_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
771
+ "blocks.34.cross_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
772
+ "blocks.34.ffn.0.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
773
+ "blocks.34.ffn.0.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
774
+ "blocks.34.ffn.2.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
775
+ "blocks.34.ffn.2.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
776
+ "blocks.34.modulation": "diffusion_pytorch_model-00005-of-00006.safetensors",
777
+ "blocks.34.norm3.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
778
+ "blocks.34.norm3.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
779
+ "blocks.34.self_attn.k.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
780
+ "blocks.34.self_attn.k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
781
+ "blocks.34.self_attn.norm_k.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
782
+ "blocks.34.self_attn.norm_q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
783
+ "blocks.34.self_attn.o.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
784
+ "blocks.34.self_attn.o.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
785
+ "blocks.34.self_attn.q.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
786
+ "blocks.34.self_attn.q.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
787
+ "blocks.34.self_attn.v.bias": "diffusion_pytorch_model-00005-of-00006.safetensors",
788
+ "blocks.34.self_attn.v.weight": "diffusion_pytorch_model-00005-of-00006.safetensors",
789
+ "blocks.35.cross_attn.k.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
790
+ "blocks.35.cross_attn.k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
791
+ "blocks.35.cross_attn.norm_k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
792
+ "blocks.35.cross_attn.norm_q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
793
+ "blocks.35.cross_attn.o.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
794
+ "blocks.35.cross_attn.o.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
795
+ "blocks.35.cross_attn.q.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
796
+ "blocks.35.cross_attn.q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
797
+ "blocks.35.cross_attn.v.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
798
+ "blocks.35.cross_attn.v.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
799
+ "blocks.35.ffn.0.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
800
+ "blocks.35.ffn.0.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
801
+ "blocks.35.ffn.2.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
802
+ "blocks.35.ffn.2.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
803
+ "blocks.35.modulation": "diffusion_pytorch_model-00006-of-00006.safetensors",
804
+ "blocks.35.norm3.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
805
+ "blocks.35.norm3.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
806
+ "blocks.35.self_attn.k.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
807
+ "blocks.35.self_attn.k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
808
+ "blocks.35.self_attn.norm_k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
809
+ "blocks.35.self_attn.norm_q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
810
+ "blocks.35.self_attn.o.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
811
+ "blocks.35.self_attn.o.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
812
+ "blocks.35.self_attn.q.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
813
+ "blocks.35.self_attn.q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
814
+ "blocks.35.self_attn.v.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
815
+ "blocks.35.self_attn.v.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
816
+ "blocks.36.cross_attn.k.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
817
+ "blocks.36.cross_attn.k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
818
+ "blocks.36.cross_attn.norm_k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
819
+ "blocks.36.cross_attn.norm_q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
820
+ "blocks.36.cross_attn.o.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
821
+ "blocks.36.cross_attn.o.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
822
+ "blocks.36.cross_attn.q.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
823
+ "blocks.36.cross_attn.q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
824
+ "blocks.36.cross_attn.v.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
825
+ "blocks.36.cross_attn.v.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
826
+ "blocks.36.ffn.0.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
827
+ "blocks.36.ffn.0.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
828
+ "blocks.36.ffn.2.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
829
+ "blocks.36.ffn.2.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
830
+ "blocks.36.modulation": "diffusion_pytorch_model-00006-of-00006.safetensors",
831
+ "blocks.36.norm3.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
832
+ "blocks.36.norm3.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
833
+ "blocks.36.self_attn.k.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
834
+ "blocks.36.self_attn.k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
835
+ "blocks.36.self_attn.norm_k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
836
+ "blocks.36.self_attn.norm_q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
837
+ "blocks.36.self_attn.o.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
838
+ "blocks.36.self_attn.o.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
839
+ "blocks.36.self_attn.q.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
840
+ "blocks.36.self_attn.q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
841
+ "blocks.36.self_attn.v.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
842
+ "blocks.36.self_attn.v.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
843
+ "blocks.37.cross_attn.k.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
844
+ "blocks.37.cross_attn.k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
845
+ "blocks.37.cross_attn.norm_k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
846
+ "blocks.37.cross_attn.norm_q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
847
+ "blocks.37.cross_attn.o.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
848
+ "blocks.37.cross_attn.o.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
849
+ "blocks.37.cross_attn.q.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
850
+ "blocks.37.cross_attn.q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
851
+ "blocks.37.cross_attn.v.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
852
+ "blocks.37.cross_attn.v.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
853
+ "blocks.37.ffn.0.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
854
+ "blocks.37.ffn.0.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
855
+ "blocks.37.ffn.2.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
856
+ "blocks.37.ffn.2.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
857
+ "blocks.37.modulation": "diffusion_pytorch_model-00006-of-00006.safetensors",
858
+ "blocks.37.norm3.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
859
+ "blocks.37.norm3.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
860
+ "blocks.37.self_attn.k.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
861
+ "blocks.37.self_attn.k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
862
+ "blocks.37.self_attn.norm_k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
863
+ "blocks.37.self_attn.norm_q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
864
+ "blocks.37.self_attn.o.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
865
+ "blocks.37.self_attn.o.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
866
+ "blocks.37.self_attn.q.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
867
+ "blocks.37.self_attn.q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
868
+ "blocks.37.self_attn.v.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
869
+ "blocks.37.self_attn.v.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
870
+ "blocks.38.cross_attn.k.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
871
+ "blocks.38.cross_attn.k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
872
+ "blocks.38.cross_attn.norm_k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
873
+ "blocks.38.cross_attn.norm_q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
874
+ "blocks.38.cross_attn.o.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
875
+ "blocks.38.cross_attn.o.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
876
+ "blocks.38.cross_attn.q.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
877
+ "blocks.38.cross_attn.q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
878
+ "blocks.38.cross_attn.v.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
879
+ "blocks.38.cross_attn.v.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
880
+ "blocks.38.ffn.0.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
881
+ "blocks.38.ffn.0.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
882
+ "blocks.38.ffn.2.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
883
+ "blocks.38.ffn.2.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
884
+ "blocks.38.modulation": "diffusion_pytorch_model-00006-of-00006.safetensors",
885
+ "blocks.38.norm3.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
886
+ "blocks.38.norm3.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
887
+ "blocks.38.self_attn.k.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
888
+ "blocks.38.self_attn.k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
889
+ "blocks.38.self_attn.norm_k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
890
+ "blocks.38.self_attn.norm_q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
891
+ "blocks.38.self_attn.o.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
892
+ "blocks.38.self_attn.o.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
893
+ "blocks.38.self_attn.q.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
894
+ "blocks.38.self_attn.q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
895
+ "blocks.38.self_attn.v.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
896
+ "blocks.38.self_attn.v.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
897
+ "blocks.39.cross_attn.k.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
898
+ "blocks.39.cross_attn.k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
899
+ "blocks.39.cross_attn.norm_k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
900
+ "blocks.39.cross_attn.norm_q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
901
+ "blocks.39.cross_attn.o.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
902
+ "blocks.39.cross_attn.o.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
903
+ "blocks.39.cross_attn.q.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
904
+ "blocks.39.cross_attn.q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
905
+ "blocks.39.cross_attn.v.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
906
+ "blocks.39.cross_attn.v.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
907
+ "blocks.39.ffn.0.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
908
+ "blocks.39.ffn.0.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
909
+ "blocks.39.ffn.2.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
910
+ "blocks.39.ffn.2.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
911
+ "blocks.39.modulation": "diffusion_pytorch_model-00006-of-00006.safetensors",
912
+ "blocks.39.norm3.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
913
+ "blocks.39.norm3.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
914
+ "blocks.39.self_attn.k.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
915
+ "blocks.39.self_attn.k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
916
+ "blocks.39.self_attn.norm_k.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
917
+ "blocks.39.self_attn.norm_q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
918
+ "blocks.39.self_attn.o.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
919
+ "blocks.39.self_attn.o.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
920
+ "blocks.39.self_attn.q.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
921
+ "blocks.39.self_attn.q.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
922
+ "blocks.39.self_attn.v.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
923
+ "blocks.39.self_attn.v.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
924
+ "blocks.4.cross_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
925
+ "blocks.4.cross_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
926
+ "blocks.4.cross_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
927
+ "blocks.4.cross_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
928
+ "blocks.4.cross_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
929
+ "blocks.4.cross_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
930
+ "blocks.4.cross_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
931
+ "blocks.4.cross_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
932
+ "blocks.4.cross_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
933
+ "blocks.4.cross_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
934
+ "blocks.4.ffn.0.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
935
+ "blocks.4.ffn.0.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
936
+ "blocks.4.ffn.2.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
937
+ "blocks.4.ffn.2.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
938
+ "blocks.4.modulation": "diffusion_pytorch_model-00001-of-00006.safetensors",
939
+ "blocks.4.norm3.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
940
+ "blocks.4.norm3.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
941
+ "blocks.4.self_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
942
+ "blocks.4.self_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
943
+ "blocks.4.self_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
944
+ "blocks.4.self_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
945
+ "blocks.4.self_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
946
+ "blocks.4.self_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
947
+ "blocks.4.self_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
948
+ "blocks.4.self_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
949
+ "blocks.4.self_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
950
+ "blocks.4.self_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
951
+ "blocks.5.cross_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
952
+ "blocks.5.cross_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
953
+ "blocks.5.cross_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
954
+ "blocks.5.cross_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
955
+ "blocks.5.cross_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
956
+ "blocks.5.cross_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
957
+ "blocks.5.cross_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
958
+ "blocks.5.cross_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
959
+ "blocks.5.cross_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
960
+ "blocks.5.cross_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
961
+ "blocks.5.ffn.0.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
962
+ "blocks.5.ffn.0.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
963
+ "blocks.5.ffn.2.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
964
+ "blocks.5.ffn.2.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
965
+ "blocks.5.modulation": "diffusion_pytorch_model-00001-of-00006.safetensors",
966
+ "blocks.5.norm3.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
967
+ "blocks.5.norm3.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
968
+ "blocks.5.self_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
969
+ "blocks.5.self_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
970
+ "blocks.5.self_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
971
+ "blocks.5.self_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
972
+ "blocks.5.self_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
973
+ "blocks.5.self_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
974
+ "blocks.5.self_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
975
+ "blocks.5.self_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
976
+ "blocks.5.self_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
977
+ "blocks.5.self_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
978
+ "blocks.6.cross_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
979
+ "blocks.6.cross_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
980
+ "blocks.6.cross_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
981
+ "blocks.6.cross_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
982
+ "blocks.6.cross_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
983
+ "blocks.6.cross_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
984
+ "blocks.6.cross_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
985
+ "blocks.6.cross_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
986
+ "blocks.6.cross_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
987
+ "blocks.6.cross_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
988
+ "blocks.6.ffn.0.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
989
+ "blocks.6.ffn.0.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
990
+ "blocks.6.ffn.2.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
991
+ "blocks.6.ffn.2.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
992
+ "blocks.6.modulation": "diffusion_pytorch_model-00001-of-00006.safetensors",
993
+ "blocks.6.norm3.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
994
+ "blocks.6.norm3.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
995
+ "blocks.6.self_attn.k.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
996
+ "blocks.6.self_attn.k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
997
+ "blocks.6.self_attn.norm_k.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
998
+ "blocks.6.self_attn.norm_q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
999
+ "blocks.6.self_attn.o.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
1000
+ "blocks.6.self_attn.o.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
1001
+ "blocks.6.self_attn.q.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
1002
+ "blocks.6.self_attn.q.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
1003
+ "blocks.6.self_attn.v.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
1004
+ "blocks.6.self_attn.v.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
1005
+ "blocks.7.cross_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1006
+ "blocks.7.cross_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1007
+ "blocks.7.cross_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1008
+ "blocks.7.cross_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1009
+ "blocks.7.cross_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1010
+ "blocks.7.cross_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1011
+ "blocks.7.cross_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1012
+ "blocks.7.cross_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1013
+ "blocks.7.cross_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1014
+ "blocks.7.cross_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1015
+ "blocks.7.ffn.0.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1016
+ "blocks.7.ffn.0.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1017
+ "blocks.7.ffn.2.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1018
+ "blocks.7.ffn.2.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1019
+ "blocks.7.modulation": "diffusion_pytorch_model-00002-of-00006.safetensors",
1020
+ "blocks.7.norm3.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1021
+ "blocks.7.norm3.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1022
+ "blocks.7.self_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1023
+ "blocks.7.self_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1024
+ "blocks.7.self_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1025
+ "blocks.7.self_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1026
+ "blocks.7.self_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1027
+ "blocks.7.self_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1028
+ "blocks.7.self_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1029
+ "blocks.7.self_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1030
+ "blocks.7.self_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1031
+ "blocks.7.self_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1032
+ "blocks.8.cross_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1033
+ "blocks.8.cross_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1034
+ "blocks.8.cross_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1035
+ "blocks.8.cross_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1036
+ "blocks.8.cross_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1037
+ "blocks.8.cross_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1038
+ "blocks.8.cross_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1039
+ "blocks.8.cross_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1040
+ "blocks.8.cross_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1041
+ "blocks.8.cross_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1042
+ "blocks.8.ffn.0.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1043
+ "blocks.8.ffn.0.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1044
+ "blocks.8.ffn.2.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1045
+ "blocks.8.ffn.2.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1046
+ "blocks.8.modulation": "diffusion_pytorch_model-00002-of-00006.safetensors",
1047
+ "blocks.8.norm3.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1048
+ "blocks.8.norm3.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1049
+ "blocks.8.self_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1050
+ "blocks.8.self_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1051
+ "blocks.8.self_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1052
+ "blocks.8.self_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1053
+ "blocks.8.self_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1054
+ "blocks.8.self_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1055
+ "blocks.8.self_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1056
+ "blocks.8.self_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1057
+ "blocks.8.self_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1058
+ "blocks.8.self_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1059
+ "blocks.9.cross_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1060
+ "blocks.9.cross_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1061
+ "blocks.9.cross_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1062
+ "blocks.9.cross_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1063
+ "blocks.9.cross_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1064
+ "blocks.9.cross_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1065
+ "blocks.9.cross_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1066
+ "blocks.9.cross_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1067
+ "blocks.9.cross_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1068
+ "blocks.9.cross_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1069
+ "blocks.9.ffn.0.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1070
+ "blocks.9.ffn.0.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1071
+ "blocks.9.ffn.2.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1072
+ "blocks.9.ffn.2.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1073
+ "blocks.9.modulation": "diffusion_pytorch_model-00002-of-00006.safetensors",
1074
+ "blocks.9.norm3.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1075
+ "blocks.9.norm3.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1076
+ "blocks.9.self_attn.k.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1077
+ "blocks.9.self_attn.k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1078
+ "blocks.9.self_attn.norm_k.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1079
+ "blocks.9.self_attn.norm_q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1080
+ "blocks.9.self_attn.o.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1081
+ "blocks.9.self_attn.o.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1082
+ "blocks.9.self_attn.q.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1083
+ "blocks.9.self_attn.q.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1084
+ "blocks.9.self_attn.v.bias": "diffusion_pytorch_model-00002-of-00006.safetensors",
1085
+ "blocks.9.self_attn.v.weight": "diffusion_pytorch_model-00002-of-00006.safetensors",
1086
+ "head.head.bias": "diffusion_pytorch_model-00006-of-00006.safetensors",
1087
+ "head.head.weight": "diffusion_pytorch_model-00006-of-00006.safetensors",
1088
+ "head.modulation": "diffusion_pytorch_model-00006-of-00006.safetensors",
1089
+ "patch_embedding.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
1090
+ "patch_embedding.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
1091
+ "text_embedding.0.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
1092
+ "text_embedding.0.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
1093
+ "text_embedding.2.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
1094
+ "text_embedding.2.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
1095
+ "time_embedding.0.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
1096
+ "time_embedding.0.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
1097
+ "time_embedding.2.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
1098
+ "time_embedding.2.weight": "diffusion_pytorch_model-00001-of-00006.safetensors",
1099
+ "time_projection.1.bias": "diffusion_pytorch_model-00001-of-00006.safetensors",
1100
+ "time_projection.1.weight": "diffusion_pytorch_model-00001-of-00006.safetensors"
1101
+ }
1102
+ }
google/umt5-xxl/special_tokens_map.json ADDED
@@ -0,0 +1,308 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<extra_id_0>",
4
+ "<extra_id_1>",
5
+ "<extra_id_2>",
6
+ "<extra_id_3>",
7
+ "<extra_id_4>",
8
+ "<extra_id_5>",
9
+ "<extra_id_6>",
10
+ "<extra_id_7>",
11
+ "<extra_id_8>",
12
+ "<extra_id_9>",
13
+ "<extra_id_10>",
14
+ "<extra_id_11>",
15
+ "<extra_id_12>",
16
+ "<extra_id_13>",
17
+ "<extra_id_14>",
18
+ "<extra_id_15>",
19
+ "<extra_id_16>",
20
+ "<extra_id_17>",
21
+ "<extra_id_18>",
22
+ "<extra_id_19>",
23
+ "<extra_id_20>",
24
+ "<extra_id_21>",
25
+ "<extra_id_22>",
26
+ "<extra_id_23>",
27
+ "<extra_id_24>",
28
+ "<extra_id_25>",
29
+ "<extra_id_26>",
30
+ "<extra_id_27>",
31
+ "<extra_id_28>",
32
+ "<extra_id_29>",
33
+ "<extra_id_30>",
34
+ "<extra_id_31>",
35
+ "<extra_id_32>",
36
+ "<extra_id_33>",
37
+ "<extra_id_34>",
38
+ "<extra_id_35>",
39
+ "<extra_id_36>",
40
+ "<extra_id_37>",
41
+ "<extra_id_38>",
42
+ "<extra_id_39>",
43
+ "<extra_id_40>",
44
+ "<extra_id_41>",
45
+ "<extra_id_42>",
46
+ "<extra_id_43>",
47
+ "<extra_id_44>",
48
+ "<extra_id_45>",
49
+ "<extra_id_46>",
50
+ "<extra_id_47>",
51
+ "<extra_id_48>",
52
+ "<extra_id_49>",
53
+ "<extra_id_50>",
54
+ "<extra_id_51>",
55
+ "<extra_id_52>",
56
+ "<extra_id_53>",
57
+ "<extra_id_54>",
58
+ "<extra_id_55>",
59
+ "<extra_id_56>",
60
+ "<extra_id_57>",
61
+ "<extra_id_58>",
62
+ "<extra_id_59>",
63
+ "<extra_id_60>",
64
+ "<extra_id_61>",
65
+ "<extra_id_62>",
66
+ "<extra_id_63>",
67
+ "<extra_id_64>",
68
+ "<extra_id_65>",
69
+ "<extra_id_66>",
70
+ "<extra_id_67>",
71
+ "<extra_id_68>",
72
+ "<extra_id_69>",
73
+ "<extra_id_70>",
74
+ "<extra_id_71>",
75
+ "<extra_id_72>",
76
+ "<extra_id_73>",
77
+ "<extra_id_74>",
78
+ "<extra_id_75>",
79
+ "<extra_id_76>",
80
+ "<extra_id_77>",
81
+ "<extra_id_78>",
82
+ "<extra_id_79>",
83
+ "<extra_id_80>",
84
+ "<extra_id_81>",
85
+ "<extra_id_82>",
86
+ "<extra_id_83>",
87
+ "<extra_id_84>",
88
+ "<extra_id_85>",
89
+ "<extra_id_86>",
90
+ "<extra_id_87>",
91
+ "<extra_id_88>",
92
+ "<extra_id_89>",
93
+ "<extra_id_90>",
94
+ "<extra_id_91>",
95
+ "<extra_id_92>",
96
+ "<extra_id_93>",
97
+ "<extra_id_94>",
98
+ "<extra_id_95>",
99
+ "<extra_id_96>",
100
+ "<extra_id_97>",
101
+ "<extra_id_98>",
102
+ "<extra_id_99>",
103
+ "<extra_id_100>",
104
+ "<extra_id_101>",
105
+ "<extra_id_102>",
106
+ "<extra_id_103>",
107
+ "<extra_id_104>",
108
+ "<extra_id_105>",
109
+ "<extra_id_106>",
110
+ "<extra_id_107>",
111
+ "<extra_id_108>",
112
+ "<extra_id_109>",
113
+ "<extra_id_110>",
114
+ "<extra_id_111>",
115
+ "<extra_id_112>",
116
+ "<extra_id_113>",
117
+ "<extra_id_114>",
118
+ "<extra_id_115>",
119
+ "<extra_id_116>",
120
+ "<extra_id_117>",
121
+ "<extra_id_118>",
122
+ "<extra_id_119>",
123
+ "<extra_id_120>",
124
+ "<extra_id_121>",
125
+ "<extra_id_122>",
126
+ "<extra_id_123>",
127
+ "<extra_id_124>",
128
+ "<extra_id_125>",
129
+ "<extra_id_126>",
130
+ "<extra_id_127>",
131
+ "<extra_id_128>",
132
+ "<extra_id_129>",
133
+ "<extra_id_130>",
134
+ "<extra_id_131>",
135
+ "<extra_id_132>",
136
+ "<extra_id_133>",
137
+ "<extra_id_134>",
138
+ "<extra_id_135>",
139
+ "<extra_id_136>",
140
+ "<extra_id_137>",
141
+ "<extra_id_138>",
142
+ "<extra_id_139>",
143
+ "<extra_id_140>",
144
+ "<extra_id_141>",
145
+ "<extra_id_142>",
146
+ "<extra_id_143>",
147
+ "<extra_id_144>",
148
+ "<extra_id_145>",
149
+ "<extra_id_146>",
150
+ "<extra_id_147>",
151
+ "<extra_id_148>",
152
+ "<extra_id_149>",
153
+ "<extra_id_150>",
154
+ "<extra_id_151>",
155
+ "<extra_id_152>",
156
+ "<extra_id_153>",
157
+ "<extra_id_154>",
158
+ "<extra_id_155>",
159
+ "<extra_id_156>",
160
+ "<extra_id_157>",
161
+ "<extra_id_158>",
162
+ "<extra_id_159>",
163
+ "<extra_id_160>",
164
+ "<extra_id_161>",
165
+ "<extra_id_162>",
166
+ "<extra_id_163>",
167
+ "<extra_id_164>",
168
+ "<extra_id_165>",
169
+ "<extra_id_166>",
170
+ "<extra_id_167>",
171
+ "<extra_id_168>",
172
+ "<extra_id_169>",
173
+ "<extra_id_170>",
174
+ "<extra_id_171>",
175
+ "<extra_id_172>",
176
+ "<extra_id_173>",
177
+ "<extra_id_174>",
178
+ "<extra_id_175>",
179
+ "<extra_id_176>",
180
+ "<extra_id_177>",
181
+ "<extra_id_178>",
182
+ "<extra_id_179>",
183
+ "<extra_id_180>",
184
+ "<extra_id_181>",
185
+ "<extra_id_182>",
186
+ "<extra_id_183>",
187
+ "<extra_id_184>",
188
+ "<extra_id_185>",
189
+ "<extra_id_186>",
190
+ "<extra_id_187>",
191
+ "<extra_id_188>",
192
+ "<extra_id_189>",
193
+ "<extra_id_190>",
194
+ "<extra_id_191>",
195
+ "<extra_id_192>",
196
+ "<extra_id_193>",
197
+ "<extra_id_194>",
198
+ "<extra_id_195>",
199
+ "<extra_id_196>",
200
+ "<extra_id_197>",
201
+ "<extra_id_198>",
202
+ "<extra_id_199>",
203
+ "<extra_id_200>",
204
+ "<extra_id_201>",
205
+ "<extra_id_202>",
206
+ "<extra_id_203>",
207
+ "<extra_id_204>",
208
+ "<extra_id_205>",
209
+ "<extra_id_206>",
210
+ "<extra_id_207>",
211
+ "<extra_id_208>",
212
+ "<extra_id_209>",
213
+ "<extra_id_210>",
214
+ "<extra_id_211>",
215
+ "<extra_id_212>",
216
+ "<extra_id_213>",
217
+ "<extra_id_214>",
218
+ "<extra_id_215>",
219
+ "<extra_id_216>",
220
+ "<extra_id_217>",
221
+ "<extra_id_218>",
222
+ "<extra_id_219>",
223
+ "<extra_id_220>",
224
+ "<extra_id_221>",
225
+ "<extra_id_222>",
226
+ "<extra_id_223>",
227
+ "<extra_id_224>",
228
+ "<extra_id_225>",
229
+ "<extra_id_226>",
230
+ "<extra_id_227>",
231
+ "<extra_id_228>",
232
+ "<extra_id_229>",
233
+ "<extra_id_230>",
234
+ "<extra_id_231>",
235
+ "<extra_id_232>",
236
+ "<extra_id_233>",
237
+ "<extra_id_234>",
238
+ "<extra_id_235>",
239
+ "<extra_id_236>",
240
+ "<extra_id_237>",
241
+ "<extra_id_238>",
242
+ "<extra_id_239>",
243
+ "<extra_id_240>",
244
+ "<extra_id_241>",
245
+ "<extra_id_242>",
246
+ "<extra_id_243>",
247
+ "<extra_id_244>",
248
+ "<extra_id_245>",
249
+ "<extra_id_246>",
250
+ "<extra_id_247>",
251
+ "<extra_id_248>",
252
+ "<extra_id_249>",
253
+ "<extra_id_250>",
254
+ "<extra_id_251>",
255
+ "<extra_id_252>",
256
+ "<extra_id_253>",
257
+ "<extra_id_254>",
258
+ "<extra_id_255>",
259
+ "<extra_id_256>",
260
+ "<extra_id_257>",
261
+ "<extra_id_258>",
262
+ "<extra_id_259>",
263
+ "<extra_id_260>",
264
+ "<extra_id_261>",
265
+ "<extra_id_262>",
266
+ "<extra_id_263>",
267
+ "<extra_id_264>",
268
+ "<extra_id_265>",
269
+ "<extra_id_266>",
270
+ "<extra_id_267>",
271
+ "<extra_id_268>",
272
+ "<extra_id_269>",
273
+ "<extra_id_270>",
274
+ "<extra_id_271>",
275
+ "<extra_id_272>",
276
+ "<extra_id_273>",
277
+ "<extra_id_274>",
278
+ "<extra_id_275>",
279
+ "<extra_id_276>",
280
+ "<extra_id_277>",
281
+ "<extra_id_278>",
282
+ "<extra_id_279>",
283
+ "<extra_id_280>",
284
+ "<extra_id_281>",
285
+ "<extra_id_282>",
286
+ "<extra_id_283>",
287
+ "<extra_id_284>",
288
+ "<extra_id_285>",
289
+ "<extra_id_286>",
290
+ "<extra_id_287>",
291
+ "<extra_id_288>",
292
+ "<extra_id_289>",
293
+ "<extra_id_290>",
294
+ "<extra_id_291>",
295
+ "<extra_id_292>",
296
+ "<extra_id_293>",
297
+ "<extra_id_294>",
298
+ "<extra_id_295>",
299
+ "<extra_id_296>",
300
+ "<extra_id_297>",
301
+ "<extra_id_298>",
302
+ "<extra_id_299>"
303
+ ],
304
+ "bos_token": "<s>",
305
+ "eos_token": "</s>",
306
+ "pad_token": "<pad>",
307
+ "unk_token": "<unk>"
308
+ }
google/umt5-xxl/spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3909a67b780650b35cf529ac782ad2b6b26e6d1f849d3fbb6a872905f452458
3
+ size 4548313
google/umt5-xxl/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e197b4d3dbd71da14b4eb255f4fa91c9c1f2068b20a2de2472967ca3d22602b
3
+ size 16837417
google/umt5-xxl/tokenizer_config.json ADDED
@@ -0,0 +1,2748 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<pad>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "</s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "<s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "256000": {
36
+ "content": "<extra_id_299>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "256001": {
44
+ "content": "<extra_id_298>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "256002": {
52
+ "content": "<extra_id_297>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "256003": {
60
+ "content": "<extra_id_296>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "256004": {
68
+ "content": "<extra_id_295>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "256005": {
76
+ "content": "<extra_id_294>",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "256006": {
84
+ "content": "<extra_id_293>",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "256007": {
92
+ "content": "<extra_id_292>",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "256008": {
100
+ "content": "<extra_id_291>",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "256009": {
108
+ "content": "<extra_id_290>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "256010": {
116
+ "content": "<extra_id_289>",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "256011": {
124
+ "content": "<extra_id_288>",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "256012": {
132
+ "content": "<extra_id_287>",
133
+ "lstrip": false,
134
+ "normalized": false,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": true
138
+ },
139
+ "256013": {
140
+ "content": "<extra_id_286>",
141
+ "lstrip": false,
142
+ "normalized": false,
143
+ "rstrip": false,
144
+ "single_word": false,
145
+ "special": true
146
+ },
147
+ "256014": {
148
+ "content": "<extra_id_285>",
149
+ "lstrip": false,
150
+ "normalized": false,
151
+ "rstrip": false,
152
+ "single_word": false,
153
+ "special": true
154
+ },
155
+ "256015": {
156
+ "content": "<extra_id_284>",
157
+ "lstrip": false,
158
+ "normalized": false,
159
+ "rstrip": false,
160
+ "single_word": false,
161
+ "special": true
162
+ },
163
+ "256016": {
164
+ "content": "<extra_id_283>",
165
+ "lstrip": false,
166
+ "normalized": false,
167
+ "rstrip": false,
168
+ "single_word": false,
169
+ "special": true
170
+ },
171
+ "256017": {
172
+ "content": "<extra_id_282>",
173
+ "lstrip": false,
174
+ "normalized": false,
175
+ "rstrip": false,
176
+ "single_word": false,
177
+ "special": true
178
+ },
179
+ "256018": {
180
+ "content": "<extra_id_281>",
181
+ "lstrip": false,
182
+ "normalized": false,
183
+ "rstrip": false,
184
+ "single_word": false,
185
+ "special": true
186
+ },
187
+ "256019": {
188
+ "content": "<extra_id_280>",
189
+ "lstrip": false,
190
+ "normalized": false,
191
+ "rstrip": false,
192
+ "single_word": false,
193
+ "special": true
194
+ },
195
+ "256020": {
196
+ "content": "<extra_id_279>",
197
+ "lstrip": false,
198
+ "normalized": false,
199
+ "rstrip": false,
200
+ "single_word": false,
201
+ "special": true
202
+ },
203
+ "256021": {
204
+ "content": "<extra_id_278>",
205
+ "lstrip": false,
206
+ "normalized": false,
207
+ "rstrip": false,
208
+ "single_word": false,
209
+ "special": true
210
+ },
211
+ "256022": {
212
+ "content": "<extra_id_277>",
213
+ "lstrip": false,
214
+ "normalized": false,
215
+ "rstrip": false,
216
+ "single_word": false,
217
+ "special": true
218
+ },
219
+ "256023": {
220
+ "content": "<extra_id_276>",
221
+ "lstrip": false,
222
+ "normalized": false,
223
+ "rstrip": false,
224
+ "single_word": false,
225
+ "special": true
226
+ },
227
+ "256024": {
228
+ "content": "<extra_id_275>",
229
+ "lstrip": false,
230
+ "normalized": false,
231
+ "rstrip": false,
232
+ "single_word": false,
233
+ "special": true
234
+ },
235
+ "256025": {
236
+ "content": "<extra_id_274>",
237
+ "lstrip": false,
238
+ "normalized": false,
239
+ "rstrip": false,
240
+ "single_word": false,
241
+ "special": true
242
+ },
243
+ "256026": {
244
+ "content": "<extra_id_273>",
245
+ "lstrip": false,
246
+ "normalized": false,
247
+ "rstrip": false,
248
+ "single_word": false,
249
+ "special": true
250
+ },
251
+ "256027": {
252
+ "content": "<extra_id_272>",
253
+ "lstrip": false,
254
+ "normalized": false,
255
+ "rstrip": false,
256
+ "single_word": false,
257
+ "special": true
258
+ },
259
+ "256028": {
260
+ "content": "<extra_id_271>",
261
+ "lstrip": false,
262
+ "normalized": false,
263
+ "rstrip": false,
264
+ "single_word": false,
265
+ "special": true
266
+ },
267
+ "256029": {
268
+ "content": "<extra_id_270>",
269
+ "lstrip": false,
270
+ "normalized": false,
271
+ "rstrip": false,
272
+ "single_word": false,
273
+ "special": true
274
+ },
275
+ "256030": {
276
+ "content": "<extra_id_269>",
277
+ "lstrip": false,
278
+ "normalized": false,
279
+ "rstrip": false,
280
+ "single_word": false,
281
+ "special": true
282
+ },
283
+ "256031": {
284
+ "content": "<extra_id_268>",
285
+ "lstrip": false,
286
+ "normalized": false,
287
+ "rstrip": false,
288
+ "single_word": false,
289
+ "special": true
290
+ },
291
+ "256032": {
292
+ "content": "<extra_id_267>",
293
+ "lstrip": false,
294
+ "normalized": false,
295
+ "rstrip": false,
296
+ "single_word": false,
297
+ "special": true
298
+ },
299
+ "256033": {
300
+ "content": "<extra_id_266>",
301
+ "lstrip": false,
302
+ "normalized": false,
303
+ "rstrip": false,
304
+ "single_word": false,
305
+ "special": true
306
+ },
307
+ "256034": {
308
+ "content": "<extra_id_265>",
309
+ "lstrip": false,
310
+ "normalized": false,
311
+ "rstrip": false,
312
+ "single_word": false,
313
+ "special": true
314
+ },
315
+ "256035": {
316
+ "content": "<extra_id_264>",
317
+ "lstrip": false,
318
+ "normalized": false,
319
+ "rstrip": false,
320
+ "single_word": false,
321
+ "special": true
322
+ },
323
+ "256036": {
324
+ "content": "<extra_id_263>",
325
+ "lstrip": false,
326
+ "normalized": false,
327
+ "rstrip": false,
328
+ "single_word": false,
329
+ "special": true
330
+ },
331
+ "256037": {
332
+ "content": "<extra_id_262>",
333
+ "lstrip": false,
334
+ "normalized": false,
335
+ "rstrip": false,
336
+ "single_word": false,
337
+ "special": true
338
+ },
339
+ "256038": {
340
+ "content": "<extra_id_261>",
341
+ "lstrip": false,
342
+ "normalized": false,
343
+ "rstrip": false,
344
+ "single_word": false,
345
+ "special": true
346
+ },
347
+ "256039": {
348
+ "content": "<extra_id_260>",
349
+ "lstrip": false,
350
+ "normalized": false,
351
+ "rstrip": false,
352
+ "single_word": false,
353
+ "special": true
354
+ },
355
+ "256040": {
356
+ "content": "<extra_id_259>",
357
+ "lstrip": false,
358
+ "normalized": false,
359
+ "rstrip": false,
360
+ "single_word": false,
361
+ "special": true
362
+ },
363
+ "256041": {
364
+ "content": "<extra_id_258>",
365
+ "lstrip": false,
366
+ "normalized": false,
367
+ "rstrip": false,
368
+ "single_word": false,
369
+ "special": true
370
+ },
371
+ "256042": {
372
+ "content": "<extra_id_257>",
373
+ "lstrip": false,
374
+ "normalized": false,
375
+ "rstrip": false,
376
+ "single_word": false,
377
+ "special": true
378
+ },
379
+ "256043": {
380
+ "content": "<extra_id_256>",
381
+ "lstrip": false,
382
+ "normalized": false,
383
+ "rstrip": false,
384
+ "single_word": false,
385
+ "special": true
386
+ },
387
+ "256044": {
388
+ "content": "<extra_id_255>",
389
+ "lstrip": false,
390
+ "normalized": false,
391
+ "rstrip": false,
392
+ "single_word": false,
393
+ "special": true
394
+ },
395
+ "256045": {
396
+ "content": "<extra_id_254>",
397
+ "lstrip": false,
398
+ "normalized": false,
399
+ "rstrip": false,
400
+ "single_word": false,
401
+ "special": true
402
+ },
403
+ "256046": {
404
+ "content": "<extra_id_253>",
405
+ "lstrip": false,
406
+ "normalized": false,
407
+ "rstrip": false,
408
+ "single_word": false,
409
+ "special": true
410
+ },
411
+ "256047": {
412
+ "content": "<extra_id_252>",
413
+ "lstrip": false,
414
+ "normalized": false,
415
+ "rstrip": false,
416
+ "single_word": false,
417
+ "special": true
418
+ },
419
+ "256048": {
420
+ "content": "<extra_id_251>",
421
+ "lstrip": false,
422
+ "normalized": false,
423
+ "rstrip": false,
424
+ "single_word": false,
425
+ "special": true
426
+ },
427
+ "256049": {
428
+ "content": "<extra_id_250>",
429
+ "lstrip": false,
430
+ "normalized": false,
431
+ "rstrip": false,
432
+ "single_word": false,
433
+ "special": true
434
+ },
435
+ "256050": {
436
+ "content": "<extra_id_249>",
437
+ "lstrip": false,
438
+ "normalized": false,
439
+ "rstrip": false,
440
+ "single_word": false,
441
+ "special": true
442
+ },
443
+ "256051": {
444
+ "content": "<extra_id_248>",
445
+ "lstrip": false,
446
+ "normalized": false,
447
+ "rstrip": false,
448
+ "single_word": false,
449
+ "special": true
450
+ },
451
+ "256052": {
452
+ "content": "<extra_id_247>",
453
+ "lstrip": false,
454
+ "normalized": false,
455
+ "rstrip": false,
456
+ "single_word": false,
457
+ "special": true
458
+ },
459
+ "256053": {
460
+ "content": "<extra_id_246>",
461
+ "lstrip": false,
462
+ "normalized": false,
463
+ "rstrip": false,
464
+ "single_word": false,
465
+ "special": true
466
+ },
467
+ "256054": {
468
+ "content": "<extra_id_245>",
469
+ "lstrip": false,
470
+ "normalized": false,
471
+ "rstrip": false,
472
+ "single_word": false,
473
+ "special": true
474
+ },
475
+ "256055": {
476
+ "content": "<extra_id_244>",
477
+ "lstrip": false,
478
+ "normalized": false,
479
+ "rstrip": false,
480
+ "single_word": false,
481
+ "special": true
482
+ },
483
+ "256056": {
484
+ "content": "<extra_id_243>",
485
+ "lstrip": false,
486
+ "normalized": false,
487
+ "rstrip": false,
488
+ "single_word": false,
489
+ "special": true
490
+ },
491
+ "256057": {
492
+ "content": "<extra_id_242>",
493
+ "lstrip": false,
494
+ "normalized": false,
495
+ "rstrip": false,
496
+ "single_word": false,
497
+ "special": true
498
+ },
499
+ "256058": {
500
+ "content": "<extra_id_241>",
501
+ "lstrip": false,
502
+ "normalized": false,
503
+ "rstrip": false,
504
+ "single_word": false,
505
+ "special": true
506
+ },
507
+ "256059": {
508
+ "content": "<extra_id_240>",
509
+ "lstrip": false,
510
+ "normalized": false,
511
+ "rstrip": false,
512
+ "single_word": false,
513
+ "special": true
514
+ },
515
+ "256060": {
516
+ "content": "<extra_id_239>",
517
+ "lstrip": false,
518
+ "normalized": false,
519
+ "rstrip": false,
520
+ "single_word": false,
521
+ "special": true
522
+ },
523
+ "256061": {
524
+ "content": "<extra_id_238>",
525
+ "lstrip": false,
526
+ "normalized": false,
527
+ "rstrip": false,
528
+ "single_word": false,
529
+ "special": true
530
+ },
531
+ "256062": {
532
+ "content": "<extra_id_237>",
533
+ "lstrip": false,
534
+ "normalized": false,
535
+ "rstrip": false,
536
+ "single_word": false,
537
+ "special": true
538
+ },
539
+ "256063": {
540
+ "content": "<extra_id_236>",
541
+ "lstrip": false,
542
+ "normalized": false,
543
+ "rstrip": false,
544
+ "single_word": false,
545
+ "special": true
546
+ },
547
+ "256064": {
548
+ "content": "<extra_id_235>",
549
+ "lstrip": false,
550
+ "normalized": false,
551
+ "rstrip": false,
552
+ "single_word": false,
553
+ "special": true
554
+ },
555
+ "256065": {
556
+ "content": "<extra_id_234>",
557
+ "lstrip": false,
558
+ "normalized": false,
559
+ "rstrip": false,
560
+ "single_word": false,
561
+ "special": true
562
+ },
563
+ "256066": {
564
+ "content": "<extra_id_233>",
565
+ "lstrip": false,
566
+ "normalized": false,
567
+ "rstrip": false,
568
+ "single_word": false,
569
+ "special": true
570
+ },
571
+ "256067": {
572
+ "content": "<extra_id_232>",
573
+ "lstrip": false,
574
+ "normalized": false,
575
+ "rstrip": false,
576
+ "single_word": false,
577
+ "special": true
578
+ },
579
+ "256068": {
580
+ "content": "<extra_id_231>",
581
+ "lstrip": false,
582
+ "normalized": false,
583
+ "rstrip": false,
584
+ "single_word": false,
585
+ "special": true
586
+ },
587
+ "256069": {
588
+ "content": "<extra_id_230>",
589
+ "lstrip": false,
590
+ "normalized": false,
591
+ "rstrip": false,
592
+ "single_word": false,
593
+ "special": true
594
+ },
595
+ "256070": {
596
+ "content": "<extra_id_229>",
597
+ "lstrip": false,
598
+ "normalized": false,
599
+ "rstrip": false,
600
+ "single_word": false,
601
+ "special": true
602
+ },
603
+ "256071": {
604
+ "content": "<extra_id_228>",
605
+ "lstrip": false,
606
+ "normalized": false,
607
+ "rstrip": false,
608
+ "single_word": false,
609
+ "special": true
610
+ },
611
+ "256072": {
612
+ "content": "<extra_id_227>",
613
+ "lstrip": false,
614
+ "normalized": false,
615
+ "rstrip": false,
616
+ "single_word": false,
617
+ "special": true
618
+ },
619
+ "256073": {
620
+ "content": "<extra_id_226>",
621
+ "lstrip": false,
622
+ "normalized": false,
623
+ "rstrip": false,
624
+ "single_word": false,
625
+ "special": true
626
+ },
627
+ "256074": {
628
+ "content": "<extra_id_225>",
629
+ "lstrip": false,
630
+ "normalized": false,
631
+ "rstrip": false,
632
+ "single_word": false,
633
+ "special": true
634
+ },
635
+ "256075": {
636
+ "content": "<extra_id_224>",
637
+ "lstrip": false,
638
+ "normalized": false,
639
+ "rstrip": false,
640
+ "single_word": false,
641
+ "special": true
642
+ },
643
+ "256076": {
644
+ "content": "<extra_id_223>",
645
+ "lstrip": false,
646
+ "normalized": false,
647
+ "rstrip": false,
648
+ "single_word": false,
649
+ "special": true
650
+ },
651
+ "256077": {
652
+ "content": "<extra_id_222>",
653
+ "lstrip": false,
654
+ "normalized": false,
655
+ "rstrip": false,
656
+ "single_word": false,
657
+ "special": true
658
+ },
659
+ "256078": {
660
+ "content": "<extra_id_221>",
661
+ "lstrip": false,
662
+ "normalized": false,
663
+ "rstrip": false,
664
+ "single_word": false,
665
+ "special": true
666
+ },
667
+ "256079": {
668
+ "content": "<extra_id_220>",
669
+ "lstrip": false,
670
+ "normalized": false,
671
+ "rstrip": false,
672
+ "single_word": false,
673
+ "special": true
674
+ },
675
+ "256080": {
676
+ "content": "<extra_id_219>",
677
+ "lstrip": false,
678
+ "normalized": false,
679
+ "rstrip": false,
680
+ "single_word": false,
681
+ "special": true
682
+ },
683
+ "256081": {
684
+ "content": "<extra_id_218>",
685
+ "lstrip": false,
686
+ "normalized": false,
687
+ "rstrip": false,
688
+ "single_word": false,
689
+ "special": true
690
+ },
691
+ "256082": {
692
+ "content": "<extra_id_217>",
693
+ "lstrip": false,
694
+ "normalized": false,
695
+ "rstrip": false,
696
+ "single_word": false,
697
+ "special": true
698
+ },
699
+ "256083": {
700
+ "content": "<extra_id_216>",
701
+ "lstrip": false,
702
+ "normalized": false,
703
+ "rstrip": false,
704
+ "single_word": false,
705
+ "special": true
706
+ },
707
+ "256084": {
708
+ "content": "<extra_id_215>",
709
+ "lstrip": false,
710
+ "normalized": false,
711
+ "rstrip": false,
712
+ "single_word": false,
713
+ "special": true
714
+ },
715
+ "256085": {
716
+ "content": "<extra_id_214>",
717
+ "lstrip": false,
718
+ "normalized": false,
719
+ "rstrip": false,
720
+ "single_word": false,
721
+ "special": true
722
+ },
723
+ "256086": {
724
+ "content": "<extra_id_213>",
725
+ "lstrip": false,
726
+ "normalized": false,
727
+ "rstrip": false,
728
+ "single_word": false,
729
+ "special": true
730
+ },
731
+ "256087": {
732
+ "content": "<extra_id_212>",
733
+ "lstrip": false,
734
+ "normalized": false,
735
+ "rstrip": false,
736
+ "single_word": false,
737
+ "special": true
738
+ },
739
+ "256088": {
740
+ "content": "<extra_id_211>",
741
+ "lstrip": false,
742
+ "normalized": false,
743
+ "rstrip": false,
744
+ "single_word": false,
745
+ "special": true
746
+ },
747
+ "256089": {
748
+ "content": "<extra_id_210>",
749
+ "lstrip": false,
750
+ "normalized": false,
751
+ "rstrip": false,
752
+ "single_word": false,
753
+ "special": true
754
+ },
755
+ "256090": {
756
+ "content": "<extra_id_209>",
757
+ "lstrip": false,
758
+ "normalized": false,
759
+ "rstrip": false,
760
+ "single_word": false,
761
+ "special": true
762
+ },
763
+ "256091": {
764
+ "content": "<extra_id_208>",
765
+ "lstrip": false,
766
+ "normalized": false,
767
+ "rstrip": false,
768
+ "single_word": false,
769
+ "special": true
770
+ },
771
+ "256092": {
772
+ "content": "<extra_id_207>",
773
+ "lstrip": false,
774
+ "normalized": false,
775
+ "rstrip": false,
776
+ "single_word": false,
777
+ "special": true
778
+ },
779
+ "256093": {
780
+ "content": "<extra_id_206>",
781
+ "lstrip": false,
782
+ "normalized": false,
783
+ "rstrip": false,
784
+ "single_word": false,
785
+ "special": true
786
+ },
787
+ "256094": {
788
+ "content": "<extra_id_205>",
789
+ "lstrip": false,
790
+ "normalized": false,
791
+ "rstrip": false,
792
+ "single_word": false,
793
+ "special": true
794
+ },
795
+ "256095": {
796
+ "content": "<extra_id_204>",
797
+ "lstrip": false,
798
+ "normalized": false,
799
+ "rstrip": false,
800
+ "single_word": false,
801
+ "special": true
802
+ },
803
+ "256096": {
804
+ "content": "<extra_id_203>",
805
+ "lstrip": false,
806
+ "normalized": false,
807
+ "rstrip": false,
808
+ "single_word": false,
809
+ "special": true
810
+ },
811
+ "256097": {
812
+ "content": "<extra_id_202>",
813
+ "lstrip": false,
814
+ "normalized": false,
815
+ "rstrip": false,
816
+ "single_word": false,
817
+ "special": true
818
+ },
819
+ "256098": {
820
+ "content": "<extra_id_201>",
821
+ "lstrip": false,
822
+ "normalized": false,
823
+ "rstrip": false,
824
+ "single_word": false,
825
+ "special": true
826
+ },
827
+ "256099": {
828
+ "content": "<extra_id_200>",
829
+ "lstrip": false,
830
+ "normalized": false,
831
+ "rstrip": false,
832
+ "single_word": false,
833
+ "special": true
834
+ },
835
+ "256100": {
836
+ "content": "<extra_id_199>",
837
+ "lstrip": false,
838
+ "normalized": false,
839
+ "rstrip": false,
840
+ "single_word": false,
841
+ "special": true
842
+ },
843
+ "256101": {
844
+ "content": "<extra_id_198>",
845
+ "lstrip": false,
846
+ "normalized": false,
847
+ "rstrip": false,
848
+ "single_word": false,
849
+ "special": true
850
+ },
851
+ "256102": {
852
+ "content": "<extra_id_197>",
853
+ "lstrip": false,
854
+ "normalized": false,
855
+ "rstrip": false,
856
+ "single_word": false,
857
+ "special": true
858
+ },
859
+ "256103": {
860
+ "content": "<extra_id_196>",
861
+ "lstrip": false,
862
+ "normalized": false,
863
+ "rstrip": false,
864
+ "single_word": false,
865
+ "special": true
866
+ },
867
+ "256104": {
868
+ "content": "<extra_id_195>",
869
+ "lstrip": false,
870
+ "normalized": false,
871
+ "rstrip": false,
872
+ "single_word": false,
873
+ "special": true
874
+ },
875
+ "256105": {
876
+ "content": "<extra_id_194>",
877
+ "lstrip": false,
878
+ "normalized": false,
879
+ "rstrip": false,
880
+ "single_word": false,
881
+ "special": true
882
+ },
883
+ "256106": {
884
+ "content": "<extra_id_193>",
885
+ "lstrip": false,
886
+ "normalized": false,
887
+ "rstrip": false,
888
+ "single_word": false,
889
+ "special": true
890
+ },
891
+ "256107": {
892
+ "content": "<extra_id_192>",
893
+ "lstrip": false,
894
+ "normalized": false,
895
+ "rstrip": false,
896
+ "single_word": false,
897
+ "special": true
898
+ },
899
+ "256108": {
900
+ "content": "<extra_id_191>",
901
+ "lstrip": false,
902
+ "normalized": false,
903
+ "rstrip": false,
904
+ "single_word": false,
905
+ "special": true
906
+ },
907
+ "256109": {
908
+ "content": "<extra_id_190>",
909
+ "lstrip": false,
910
+ "normalized": false,
911
+ "rstrip": false,
912
+ "single_word": false,
913
+ "special": true
914
+ },
915
+ "256110": {
916
+ "content": "<extra_id_189>",
917
+ "lstrip": false,
918
+ "normalized": false,
919
+ "rstrip": false,
920
+ "single_word": false,
921
+ "special": true
922
+ },
923
+ "256111": {
924
+ "content": "<extra_id_188>",
925
+ "lstrip": false,
926
+ "normalized": false,
927
+ "rstrip": false,
928
+ "single_word": false,
929
+ "special": true
930
+ },
931
+ "256112": {
932
+ "content": "<extra_id_187>",
933
+ "lstrip": false,
934
+ "normalized": false,
935
+ "rstrip": false,
936
+ "single_word": false,
937
+ "special": true
938
+ },
939
+ "256113": {
940
+ "content": "<extra_id_186>",
941
+ "lstrip": false,
942
+ "normalized": false,
943
+ "rstrip": false,
944
+ "single_word": false,
945
+ "special": true
946
+ },
947
+ "256114": {
948
+ "content": "<extra_id_185>",
949
+ "lstrip": false,
950
+ "normalized": false,
951
+ "rstrip": false,
952
+ "single_word": false,
953
+ "special": true
954
+ },
955
+ "256115": {
956
+ "content": "<extra_id_184>",
957
+ "lstrip": false,
958
+ "normalized": false,
959
+ "rstrip": false,
960
+ "single_word": false,
961
+ "special": true
962
+ },
963
+ "256116": {
964
+ "content": "<extra_id_183>",
965
+ "lstrip": false,
966
+ "normalized": false,
967
+ "rstrip": false,
968
+ "single_word": false,
969
+ "special": true
970
+ },
971
+ "256117": {
972
+ "content": "<extra_id_182>",
973
+ "lstrip": false,
974
+ "normalized": false,
975
+ "rstrip": false,
976
+ "single_word": false,
977
+ "special": true
978
+ },
979
+ "256118": {
980
+ "content": "<extra_id_181>",
981
+ "lstrip": false,
982
+ "normalized": false,
983
+ "rstrip": false,
984
+ "single_word": false,
985
+ "special": true
986
+ },
987
+ "256119": {
988
+ "content": "<extra_id_180>",
989
+ "lstrip": false,
990
+ "normalized": false,
991
+ "rstrip": false,
992
+ "single_word": false,
993
+ "special": true
994
+ },
995
+ "256120": {
996
+ "content": "<extra_id_179>",
997
+ "lstrip": false,
998
+ "normalized": false,
999
+ "rstrip": false,
1000
+ "single_word": false,
1001
+ "special": true
1002
+ },
1003
+ "256121": {
1004
+ "content": "<extra_id_178>",
1005
+ "lstrip": false,
1006
+ "normalized": false,
1007
+ "rstrip": false,
1008
+ "single_word": false,
1009
+ "special": true
1010
+ },
1011
+ "256122": {
1012
+ "content": "<extra_id_177>",
1013
+ "lstrip": false,
1014
+ "normalized": false,
1015
+ "rstrip": false,
1016
+ "single_word": false,
1017
+ "special": true
1018
+ },
1019
+ "256123": {
1020
+ "content": "<extra_id_176>",
1021
+ "lstrip": false,
1022
+ "normalized": false,
1023
+ "rstrip": false,
1024
+ "single_word": false,
1025
+ "special": true
1026
+ },
1027
+ "256124": {
1028
+ "content": "<extra_id_175>",
1029
+ "lstrip": false,
1030
+ "normalized": false,
1031
+ "rstrip": false,
1032
+ "single_word": false,
1033
+ "special": true
1034
+ },
1035
+ "256125": {
1036
+ "content": "<extra_id_174>",
1037
+ "lstrip": false,
1038
+ "normalized": false,
1039
+ "rstrip": false,
1040
+ "single_word": false,
1041
+ "special": true
1042
+ },
1043
+ "256126": {
1044
+ "content": "<extra_id_173>",
1045
+ "lstrip": false,
1046
+ "normalized": false,
1047
+ "rstrip": false,
1048
+ "single_word": false,
1049
+ "special": true
1050
+ },
1051
+ "256127": {
1052
+ "content": "<extra_id_172>",
1053
+ "lstrip": false,
1054
+ "normalized": false,
1055
+ "rstrip": false,
1056
+ "single_word": false,
1057
+ "special": true
1058
+ },
1059
+ "256128": {
1060
+ "content": "<extra_id_171>",
1061
+ "lstrip": false,
1062
+ "normalized": false,
1063
+ "rstrip": false,
1064
+ "single_word": false,
1065
+ "special": true
1066
+ },
1067
+ "256129": {
1068
+ "content": "<extra_id_170>",
1069
+ "lstrip": false,
1070
+ "normalized": false,
1071
+ "rstrip": false,
1072
+ "single_word": false,
1073
+ "special": true
1074
+ },
1075
+ "256130": {
1076
+ "content": "<extra_id_169>",
1077
+ "lstrip": false,
1078
+ "normalized": false,
1079
+ "rstrip": false,
1080
+ "single_word": false,
1081
+ "special": true
1082
+ },
1083
+ "256131": {
1084
+ "content": "<extra_id_168>",
1085
+ "lstrip": false,
1086
+ "normalized": false,
1087
+ "rstrip": false,
1088
+ "single_word": false,
1089
+ "special": true
1090
+ },
1091
+ "256132": {
1092
+ "content": "<extra_id_167>",
1093
+ "lstrip": false,
1094
+ "normalized": false,
1095
+ "rstrip": false,
1096
+ "single_word": false,
1097
+ "special": true
1098
+ },
1099
+ "256133": {
1100
+ "content": "<extra_id_166>",
1101
+ "lstrip": false,
1102
+ "normalized": false,
1103
+ "rstrip": false,
1104
+ "single_word": false,
1105
+ "special": true
1106
+ },
1107
+ "256134": {
1108
+ "content": "<extra_id_165>",
1109
+ "lstrip": false,
1110
+ "normalized": false,
1111
+ "rstrip": false,
1112
+ "single_word": false,
1113
+ "special": true
1114
+ },
1115
+ "256135": {
1116
+ "content": "<extra_id_164>",
1117
+ "lstrip": false,
1118
+ "normalized": false,
1119
+ "rstrip": false,
1120
+ "single_word": false,
1121
+ "special": true
1122
+ },
1123
+ "256136": {
1124
+ "content": "<extra_id_163>",
1125
+ "lstrip": false,
1126
+ "normalized": false,
1127
+ "rstrip": false,
1128
+ "single_word": false,
1129
+ "special": true
1130
+ },
1131
+ "256137": {
1132
+ "content": "<extra_id_162>",
1133
+ "lstrip": false,
1134
+ "normalized": false,
1135
+ "rstrip": false,
1136
+ "single_word": false,
1137
+ "special": true
1138
+ },
1139
+ "256138": {
1140
+ "content": "<extra_id_161>",
1141
+ "lstrip": false,
1142
+ "normalized": false,
1143
+ "rstrip": false,
1144
+ "single_word": false,
1145
+ "special": true
1146
+ },
1147
+ "256139": {
1148
+ "content": "<extra_id_160>",
1149
+ "lstrip": false,
1150
+ "normalized": false,
1151
+ "rstrip": false,
1152
+ "single_word": false,
1153
+ "special": true
1154
+ },
1155
+ "256140": {
1156
+ "content": "<extra_id_159>",
1157
+ "lstrip": false,
1158
+ "normalized": false,
1159
+ "rstrip": false,
1160
+ "single_word": false,
1161
+ "special": true
1162
+ },
1163
+ "256141": {
1164
+ "content": "<extra_id_158>",
1165
+ "lstrip": false,
1166
+ "normalized": false,
1167
+ "rstrip": false,
1168
+ "single_word": false,
1169
+ "special": true
1170
+ },
1171
+ "256142": {
1172
+ "content": "<extra_id_157>",
1173
+ "lstrip": false,
1174
+ "normalized": false,
1175
+ "rstrip": false,
1176
+ "single_word": false,
1177
+ "special": true
1178
+ },
1179
+ "256143": {
1180
+ "content": "<extra_id_156>",
1181
+ "lstrip": false,
1182
+ "normalized": false,
1183
+ "rstrip": false,
1184
+ "single_word": false,
1185
+ "special": true
1186
+ },
1187
+ "256144": {
1188
+ "content": "<extra_id_155>",
1189
+ "lstrip": false,
1190
+ "normalized": false,
1191
+ "rstrip": false,
1192
+ "single_word": false,
1193
+ "special": true
1194
+ },
1195
+ "256145": {
1196
+ "content": "<extra_id_154>",
1197
+ "lstrip": false,
1198
+ "normalized": false,
1199
+ "rstrip": false,
1200
+ "single_word": false,
1201
+ "special": true
1202
+ },
1203
+ "256146": {
1204
+ "content": "<extra_id_153>",
1205
+ "lstrip": false,
1206
+ "normalized": false,
1207
+ "rstrip": false,
1208
+ "single_word": false,
1209
+ "special": true
1210
+ },
1211
+ "256147": {
1212
+ "content": "<extra_id_152>",
1213
+ "lstrip": false,
1214
+ "normalized": false,
1215
+ "rstrip": false,
1216
+ "single_word": false,
1217
+ "special": true
1218
+ },
1219
+ "256148": {
1220
+ "content": "<extra_id_151>",
1221
+ "lstrip": false,
1222
+ "normalized": false,
1223
+ "rstrip": false,
1224
+ "single_word": false,
1225
+ "special": true
1226
+ },
1227
+ "256149": {
1228
+ "content": "<extra_id_150>",
1229
+ "lstrip": false,
1230
+ "normalized": false,
1231
+ "rstrip": false,
1232
+ "single_word": false,
1233
+ "special": true
1234
+ },
1235
+ "256150": {
1236
+ "content": "<extra_id_149>",
1237
+ "lstrip": false,
1238
+ "normalized": false,
1239
+ "rstrip": false,
1240
+ "single_word": false,
1241
+ "special": true
1242
+ },
1243
+ "256151": {
1244
+ "content": "<extra_id_148>",
1245
+ "lstrip": false,
1246
+ "normalized": false,
1247
+ "rstrip": false,
1248
+ "single_word": false,
1249
+ "special": true
1250
+ },
1251
+ "256152": {
1252
+ "content": "<extra_id_147>",
1253
+ "lstrip": false,
1254
+ "normalized": false,
1255
+ "rstrip": false,
1256
+ "single_word": false,
1257
+ "special": true
1258
+ },
1259
+ "256153": {
1260
+ "content": "<extra_id_146>",
1261
+ "lstrip": false,
1262
+ "normalized": false,
1263
+ "rstrip": false,
1264
+ "single_word": false,
1265
+ "special": true
1266
+ },
1267
+ "256154": {
1268
+ "content": "<extra_id_145>",
1269
+ "lstrip": false,
1270
+ "normalized": false,
1271
+ "rstrip": false,
1272
+ "single_word": false,
1273
+ "special": true
1274
+ },
1275
+ "256155": {
1276
+ "content": "<extra_id_144>",
1277
+ "lstrip": false,
1278
+ "normalized": false,
1279
+ "rstrip": false,
1280
+ "single_word": false,
1281
+ "special": true
1282
+ },
1283
+ "256156": {
1284
+ "content": "<extra_id_143>",
1285
+ "lstrip": false,
1286
+ "normalized": false,
1287
+ "rstrip": false,
1288
+ "single_word": false,
1289
+ "special": true
1290
+ },
1291
+ "256157": {
1292
+ "content": "<extra_id_142>",
1293
+ "lstrip": false,
1294
+ "normalized": false,
1295
+ "rstrip": false,
1296
+ "single_word": false,
1297
+ "special": true
1298
+ },
1299
+ "256158": {
1300
+ "content": "<extra_id_141>",
1301
+ "lstrip": false,
1302
+ "normalized": false,
1303
+ "rstrip": false,
1304
+ "single_word": false,
1305
+ "special": true
1306
+ },
1307
+ "256159": {
1308
+ "content": "<extra_id_140>",
1309
+ "lstrip": false,
1310
+ "normalized": false,
1311
+ "rstrip": false,
1312
+ "single_word": false,
1313
+ "special": true
1314
+ },
1315
+ "256160": {
1316
+ "content": "<extra_id_139>",
1317
+ "lstrip": false,
1318
+ "normalized": false,
1319
+ "rstrip": false,
1320
+ "single_word": false,
1321
+ "special": true
1322
+ },
1323
+ "256161": {
1324
+ "content": "<extra_id_138>",
1325
+ "lstrip": false,
1326
+ "normalized": false,
1327
+ "rstrip": false,
1328
+ "single_word": false,
1329
+ "special": true
1330
+ },
1331
+ "256162": {
1332
+ "content": "<extra_id_137>",
1333
+ "lstrip": false,
1334
+ "normalized": false,
1335
+ "rstrip": false,
1336
+ "single_word": false,
1337
+ "special": true
1338
+ },
1339
+ "256163": {
1340
+ "content": "<extra_id_136>",
1341
+ "lstrip": false,
1342
+ "normalized": false,
1343
+ "rstrip": false,
1344
+ "single_word": false,
1345
+ "special": true
1346
+ },
1347
+ "256164": {
1348
+ "content": "<extra_id_135>",
1349
+ "lstrip": false,
1350
+ "normalized": false,
1351
+ "rstrip": false,
1352
+ "single_word": false,
1353
+ "special": true
1354
+ },
1355
+ "256165": {
1356
+ "content": "<extra_id_134>",
1357
+ "lstrip": false,
1358
+ "normalized": false,
1359
+ "rstrip": false,
1360
+ "single_word": false,
1361
+ "special": true
1362
+ },
1363
+ "256166": {
1364
+ "content": "<extra_id_133>",
1365
+ "lstrip": false,
1366
+ "normalized": false,
1367
+ "rstrip": false,
1368
+ "single_word": false,
1369
+ "special": true
1370
+ },
1371
+ "256167": {
1372
+ "content": "<extra_id_132>",
1373
+ "lstrip": false,
1374
+ "normalized": false,
1375
+ "rstrip": false,
1376
+ "single_word": false,
1377
+ "special": true
1378
+ },
1379
+ "256168": {
1380
+ "content": "<extra_id_131>",
1381
+ "lstrip": false,
1382
+ "normalized": false,
1383
+ "rstrip": false,
1384
+ "single_word": false,
1385
+ "special": true
1386
+ },
1387
+ "256169": {
1388
+ "content": "<extra_id_130>",
1389
+ "lstrip": false,
1390
+ "normalized": false,
1391
+ "rstrip": false,
1392
+ "single_word": false,
1393
+ "special": true
1394
+ },
1395
+ "256170": {
1396
+ "content": "<extra_id_129>",
1397
+ "lstrip": false,
1398
+ "normalized": false,
1399
+ "rstrip": false,
1400
+ "single_word": false,
1401
+ "special": true
1402
+ },
1403
+ "256171": {
1404
+ "content": "<extra_id_128>",
1405
+ "lstrip": false,
1406
+ "normalized": false,
1407
+ "rstrip": false,
1408
+ "single_word": false,
1409
+ "special": true
1410
+ },
1411
+ "256172": {
1412
+ "content": "<extra_id_127>",
1413
+ "lstrip": false,
1414
+ "normalized": false,
1415
+ "rstrip": false,
1416
+ "single_word": false,
1417
+ "special": true
1418
+ },
1419
+ "256173": {
1420
+ "content": "<extra_id_126>",
1421
+ "lstrip": false,
1422
+ "normalized": false,
1423
+ "rstrip": false,
1424
+ "single_word": false,
1425
+ "special": true
1426
+ },
1427
+ "256174": {
1428
+ "content": "<extra_id_125>",
1429
+ "lstrip": false,
1430
+ "normalized": false,
1431
+ "rstrip": false,
1432
+ "single_word": false,
1433
+ "special": true
1434
+ },
1435
+ "256175": {
1436
+ "content": "<extra_id_124>",
1437
+ "lstrip": false,
1438
+ "normalized": false,
1439
+ "rstrip": false,
1440
+ "single_word": false,
1441
+ "special": true
1442
+ },
1443
+ "256176": {
1444
+ "content": "<extra_id_123>",
1445
+ "lstrip": false,
1446
+ "normalized": false,
1447
+ "rstrip": false,
1448
+ "single_word": false,
1449
+ "special": true
1450
+ },
1451
+ "256177": {
1452
+ "content": "<extra_id_122>",
1453
+ "lstrip": false,
1454
+ "normalized": false,
1455
+ "rstrip": false,
1456
+ "single_word": false,
1457
+ "special": true
1458
+ },
1459
+ "256178": {
1460
+ "content": "<extra_id_121>",
1461
+ "lstrip": false,
1462
+ "normalized": false,
1463
+ "rstrip": false,
1464
+ "single_word": false,
1465
+ "special": true
1466
+ },
1467
+ "256179": {
1468
+ "content": "<extra_id_120>",
1469
+ "lstrip": false,
1470
+ "normalized": false,
1471
+ "rstrip": false,
1472
+ "single_word": false,
1473
+ "special": true
1474
+ },
1475
+ "256180": {
1476
+ "content": "<extra_id_119>",
1477
+ "lstrip": false,
1478
+ "normalized": false,
1479
+ "rstrip": false,
1480
+ "single_word": false,
1481
+ "special": true
1482
+ },
1483
+ "256181": {
1484
+ "content": "<extra_id_118>",
1485
+ "lstrip": false,
1486
+ "normalized": false,
1487
+ "rstrip": false,
1488
+ "single_word": false,
1489
+ "special": true
1490
+ },
1491
+ "256182": {
1492
+ "content": "<extra_id_117>",
1493
+ "lstrip": false,
1494
+ "normalized": false,
1495
+ "rstrip": false,
1496
+ "single_word": false,
1497
+ "special": true
1498
+ },
1499
+ "256183": {
1500
+ "content": "<extra_id_116>",
1501
+ "lstrip": false,
1502
+ "normalized": false,
1503
+ "rstrip": false,
1504
+ "single_word": false,
1505
+ "special": true
1506
+ },
1507
+ "256184": {
1508
+ "content": "<extra_id_115>",
1509
+ "lstrip": false,
1510
+ "normalized": false,
1511
+ "rstrip": false,
1512
+ "single_word": false,
1513
+ "special": true
1514
+ },
1515
+ "256185": {
1516
+ "content": "<extra_id_114>",
1517
+ "lstrip": false,
1518
+ "normalized": false,
1519
+ "rstrip": false,
1520
+ "single_word": false,
1521
+ "special": true
1522
+ },
1523
+ "256186": {
1524
+ "content": "<extra_id_113>",
1525
+ "lstrip": false,
1526
+ "normalized": false,
1527
+ "rstrip": false,
1528
+ "single_word": false,
1529
+ "special": true
1530
+ },
1531
+ "256187": {
1532
+ "content": "<extra_id_112>",
1533
+ "lstrip": false,
1534
+ "normalized": false,
1535
+ "rstrip": false,
1536
+ "single_word": false,
1537
+ "special": true
1538
+ },
1539
+ "256188": {
1540
+ "content": "<extra_id_111>",
1541
+ "lstrip": false,
1542
+ "normalized": false,
1543
+ "rstrip": false,
1544
+ "single_word": false,
1545
+ "special": true
1546
+ },
1547
+ "256189": {
1548
+ "content": "<extra_id_110>",
1549
+ "lstrip": false,
1550
+ "normalized": false,
1551
+ "rstrip": false,
1552
+ "single_word": false,
1553
+ "special": true
1554
+ },
1555
+ "256190": {
1556
+ "content": "<extra_id_109>",
1557
+ "lstrip": false,
1558
+ "normalized": false,
1559
+ "rstrip": false,
1560
+ "single_word": false,
1561
+ "special": true
1562
+ },
1563
+ "256191": {
1564
+ "content": "<extra_id_108>",
1565
+ "lstrip": false,
1566
+ "normalized": false,
1567
+ "rstrip": false,
1568
+ "single_word": false,
1569
+ "special": true
1570
+ },
1571
+ "256192": {
1572
+ "content": "<extra_id_107>",
1573
+ "lstrip": false,
1574
+ "normalized": false,
1575
+ "rstrip": false,
1576
+ "single_word": false,
1577
+ "special": true
1578
+ },
1579
+ "256193": {
1580
+ "content": "<extra_id_106>",
1581
+ "lstrip": false,
1582
+ "normalized": false,
1583
+ "rstrip": false,
1584
+ "single_word": false,
1585
+ "special": true
1586
+ },
1587
+ "256194": {
1588
+ "content": "<extra_id_105>",
1589
+ "lstrip": false,
1590
+ "normalized": false,
1591
+ "rstrip": false,
1592
+ "single_word": false,
1593
+ "special": true
1594
+ },
1595
+ "256195": {
1596
+ "content": "<extra_id_104>",
1597
+ "lstrip": false,
1598
+ "normalized": false,
1599
+ "rstrip": false,
1600
+ "single_word": false,
1601
+ "special": true
1602
+ },
1603
+ "256196": {
1604
+ "content": "<extra_id_103>",
1605
+ "lstrip": false,
1606
+ "normalized": false,
1607
+ "rstrip": false,
1608
+ "single_word": false,
1609
+ "special": true
1610
+ },
1611
+ "256197": {
1612
+ "content": "<extra_id_102>",
1613
+ "lstrip": false,
1614
+ "normalized": false,
1615
+ "rstrip": false,
1616
+ "single_word": false,
1617
+ "special": true
1618
+ },
1619
+ "256198": {
1620
+ "content": "<extra_id_101>",
1621
+ "lstrip": false,
1622
+ "normalized": false,
1623
+ "rstrip": false,
1624
+ "single_word": false,
1625
+ "special": true
1626
+ },
1627
+ "256199": {
1628
+ "content": "<extra_id_100>",
1629
+ "lstrip": false,
1630
+ "normalized": false,
1631
+ "rstrip": false,
1632
+ "single_word": false,
1633
+ "special": true
1634
+ },
1635
+ "256200": {
1636
+ "content": "<extra_id_99>",
1637
+ "lstrip": false,
1638
+ "normalized": false,
1639
+ "rstrip": false,
1640
+ "single_word": false,
1641
+ "special": true
1642
+ },
1643
+ "256201": {
1644
+ "content": "<extra_id_98>",
1645
+ "lstrip": false,
1646
+ "normalized": false,
1647
+ "rstrip": false,
1648
+ "single_word": false,
1649
+ "special": true
1650
+ },
1651
+ "256202": {
1652
+ "content": "<extra_id_97>",
1653
+ "lstrip": false,
1654
+ "normalized": false,
1655
+ "rstrip": false,
1656
+ "single_word": false,
1657
+ "special": true
1658
+ },
1659
+ "256203": {
1660
+ "content": "<extra_id_96>",
1661
+ "lstrip": false,
1662
+ "normalized": false,
1663
+ "rstrip": false,
1664
+ "single_word": false,
1665
+ "special": true
1666
+ },
1667
+ "256204": {
1668
+ "content": "<extra_id_95>",
1669
+ "lstrip": false,
1670
+ "normalized": false,
1671
+ "rstrip": false,
1672
+ "single_word": false,
1673
+ "special": true
1674
+ },
1675
+ "256205": {
1676
+ "content": "<extra_id_94>",
1677
+ "lstrip": false,
1678
+ "normalized": false,
1679
+ "rstrip": false,
1680
+ "single_word": false,
1681
+ "special": true
1682
+ },
1683
+ "256206": {
1684
+ "content": "<extra_id_93>",
1685
+ "lstrip": false,
1686
+ "normalized": false,
1687
+ "rstrip": false,
1688
+ "single_word": false,
1689
+ "special": true
1690
+ },
1691
+ "256207": {
1692
+ "content": "<extra_id_92>",
1693
+ "lstrip": false,
1694
+ "normalized": false,
1695
+ "rstrip": false,
1696
+ "single_word": false,
1697
+ "special": true
1698
+ },
1699
+ "256208": {
1700
+ "content": "<extra_id_91>",
1701
+ "lstrip": false,
1702
+ "normalized": false,
1703
+ "rstrip": false,
1704
+ "single_word": false,
1705
+ "special": true
1706
+ },
1707
+ "256209": {
1708
+ "content": "<extra_id_90>",
1709
+ "lstrip": false,
1710
+ "normalized": false,
1711
+ "rstrip": false,
1712
+ "single_word": false,
1713
+ "special": true
1714
+ },
1715
+ "256210": {
1716
+ "content": "<extra_id_89>",
1717
+ "lstrip": false,
1718
+ "normalized": false,
1719
+ "rstrip": false,
1720
+ "single_word": false,
1721
+ "special": true
1722
+ },
1723
+ "256211": {
1724
+ "content": "<extra_id_88>",
1725
+ "lstrip": false,
1726
+ "normalized": false,
1727
+ "rstrip": false,
1728
+ "single_word": false,
1729
+ "special": true
1730
+ },
1731
+ "256212": {
1732
+ "content": "<extra_id_87>",
1733
+ "lstrip": false,
1734
+ "normalized": false,
1735
+ "rstrip": false,
1736
+ "single_word": false,
1737
+ "special": true
1738
+ },
1739
+ "256213": {
1740
+ "content": "<extra_id_86>",
1741
+ "lstrip": false,
1742
+ "normalized": false,
1743
+ "rstrip": false,
1744
+ "single_word": false,
1745
+ "special": true
1746
+ },
1747
+ "256214": {
1748
+ "content": "<extra_id_85>",
1749
+ "lstrip": false,
1750
+ "normalized": false,
1751
+ "rstrip": false,
1752
+ "single_word": false,
1753
+ "special": true
1754
+ },
1755
+ "256215": {
1756
+ "content": "<extra_id_84>",
1757
+ "lstrip": false,
1758
+ "normalized": false,
1759
+ "rstrip": false,
1760
+ "single_word": false,
1761
+ "special": true
1762
+ },
1763
+ "256216": {
1764
+ "content": "<extra_id_83>",
1765
+ "lstrip": false,
1766
+ "normalized": false,
1767
+ "rstrip": false,
1768
+ "single_word": false,
1769
+ "special": true
1770
+ },
1771
+ "256217": {
1772
+ "content": "<extra_id_82>",
1773
+ "lstrip": false,
1774
+ "normalized": false,
1775
+ "rstrip": false,
1776
+ "single_word": false,
1777
+ "special": true
1778
+ },
1779
+ "256218": {
1780
+ "content": "<extra_id_81>",
1781
+ "lstrip": false,
1782
+ "normalized": false,
1783
+ "rstrip": false,
1784
+ "single_word": false,
1785
+ "special": true
1786
+ },
1787
+ "256219": {
1788
+ "content": "<extra_id_80>",
1789
+ "lstrip": false,
1790
+ "normalized": false,
1791
+ "rstrip": false,
1792
+ "single_word": false,
1793
+ "special": true
1794
+ },
1795
+ "256220": {
1796
+ "content": "<extra_id_79>",
1797
+ "lstrip": false,
1798
+ "normalized": false,
1799
+ "rstrip": false,
1800
+ "single_word": false,
1801
+ "special": true
1802
+ },
1803
+ "256221": {
1804
+ "content": "<extra_id_78>",
1805
+ "lstrip": false,
1806
+ "normalized": false,
1807
+ "rstrip": false,
1808
+ "single_word": false,
1809
+ "special": true
1810
+ },
1811
+ "256222": {
1812
+ "content": "<extra_id_77>",
1813
+ "lstrip": false,
1814
+ "normalized": false,
1815
+ "rstrip": false,
1816
+ "single_word": false,
1817
+ "special": true
1818
+ },
1819
+ "256223": {
1820
+ "content": "<extra_id_76>",
1821
+ "lstrip": false,
1822
+ "normalized": false,
1823
+ "rstrip": false,
1824
+ "single_word": false,
1825
+ "special": true
1826
+ },
1827
+ "256224": {
1828
+ "content": "<extra_id_75>",
1829
+ "lstrip": false,
1830
+ "normalized": false,
1831
+ "rstrip": false,
1832
+ "single_word": false,
1833
+ "special": true
1834
+ },
1835
+ "256225": {
1836
+ "content": "<extra_id_74>",
1837
+ "lstrip": false,
1838
+ "normalized": false,
1839
+ "rstrip": false,
1840
+ "single_word": false,
1841
+ "special": true
1842
+ },
1843
+ "256226": {
1844
+ "content": "<extra_id_73>",
1845
+ "lstrip": false,
1846
+ "normalized": false,
1847
+ "rstrip": false,
1848
+ "single_word": false,
1849
+ "special": true
1850
+ },
1851
+ "256227": {
1852
+ "content": "<extra_id_72>",
1853
+ "lstrip": false,
1854
+ "normalized": false,
1855
+ "rstrip": false,
1856
+ "single_word": false,
1857
+ "special": true
1858
+ },
1859
+ "256228": {
1860
+ "content": "<extra_id_71>",
1861
+ "lstrip": false,
1862
+ "normalized": false,
1863
+ "rstrip": false,
1864
+ "single_word": false,
1865
+ "special": true
1866
+ },
1867
+ "256229": {
1868
+ "content": "<extra_id_70>",
1869
+ "lstrip": false,
1870
+ "normalized": false,
1871
+ "rstrip": false,
1872
+ "single_word": false,
1873
+ "special": true
1874
+ },
1875
+ "256230": {
1876
+ "content": "<extra_id_69>",
1877
+ "lstrip": false,
1878
+ "normalized": false,
1879
+ "rstrip": false,
1880
+ "single_word": false,
1881
+ "special": true
1882
+ },
1883
+ "256231": {
1884
+ "content": "<extra_id_68>",
1885
+ "lstrip": false,
1886
+ "normalized": false,
1887
+ "rstrip": false,
1888
+ "single_word": false,
1889
+ "special": true
1890
+ },
1891
+ "256232": {
1892
+ "content": "<extra_id_67>",
1893
+ "lstrip": false,
1894
+ "normalized": false,
1895
+ "rstrip": false,
1896
+ "single_word": false,
1897
+ "special": true
1898
+ },
1899
+ "256233": {
1900
+ "content": "<extra_id_66>",
1901
+ "lstrip": false,
1902
+ "normalized": false,
1903
+ "rstrip": false,
1904
+ "single_word": false,
1905
+ "special": true
1906
+ },
1907
+ "256234": {
1908
+ "content": "<extra_id_65>",
1909
+ "lstrip": false,
1910
+ "normalized": false,
1911
+ "rstrip": false,
1912
+ "single_word": false,
1913
+ "special": true
1914
+ },
1915
+ "256235": {
1916
+ "content": "<extra_id_64>",
1917
+ "lstrip": false,
1918
+ "normalized": false,
1919
+ "rstrip": false,
1920
+ "single_word": false,
1921
+ "special": true
1922
+ },
1923
+ "256236": {
1924
+ "content": "<extra_id_63>",
1925
+ "lstrip": false,
1926
+ "normalized": false,
1927
+ "rstrip": false,
1928
+ "single_word": false,
1929
+ "special": true
1930
+ },
1931
+ "256237": {
1932
+ "content": "<extra_id_62>",
1933
+ "lstrip": false,
1934
+ "normalized": false,
1935
+ "rstrip": false,
1936
+ "single_word": false,
1937
+ "special": true
1938
+ },
1939
+ "256238": {
1940
+ "content": "<extra_id_61>",
1941
+ "lstrip": false,
1942
+ "normalized": false,
1943
+ "rstrip": false,
1944
+ "single_word": false,
1945
+ "special": true
1946
+ },
1947
+ "256239": {
1948
+ "content": "<extra_id_60>",
1949
+ "lstrip": false,
1950
+ "normalized": false,
1951
+ "rstrip": false,
1952
+ "single_word": false,
1953
+ "special": true
1954
+ },
1955
+ "256240": {
1956
+ "content": "<extra_id_59>",
1957
+ "lstrip": false,
1958
+ "normalized": false,
1959
+ "rstrip": false,
1960
+ "single_word": false,
1961
+ "special": true
1962
+ },
1963
+ "256241": {
1964
+ "content": "<extra_id_58>",
1965
+ "lstrip": false,
1966
+ "normalized": false,
1967
+ "rstrip": false,
1968
+ "single_word": false,
1969
+ "special": true
1970
+ },
1971
+ "256242": {
1972
+ "content": "<extra_id_57>",
1973
+ "lstrip": false,
1974
+ "normalized": false,
1975
+ "rstrip": false,
1976
+ "single_word": false,
1977
+ "special": true
1978
+ },
1979
+ "256243": {
1980
+ "content": "<extra_id_56>",
1981
+ "lstrip": false,
1982
+ "normalized": false,
1983
+ "rstrip": false,
1984
+ "single_word": false,
1985
+ "special": true
1986
+ },
1987
+ "256244": {
1988
+ "content": "<extra_id_55>",
1989
+ "lstrip": false,
1990
+ "normalized": false,
1991
+ "rstrip": false,
1992
+ "single_word": false,
1993
+ "special": true
1994
+ },
1995
+ "256245": {
1996
+ "content": "<extra_id_54>",
1997
+ "lstrip": false,
1998
+ "normalized": false,
1999
+ "rstrip": false,
2000
+ "single_word": false,
2001
+ "special": true
2002
+ },
2003
+ "256246": {
2004
+ "content": "<extra_id_53>",
2005
+ "lstrip": false,
2006
+ "normalized": false,
2007
+ "rstrip": false,
2008
+ "single_word": false,
2009
+ "special": true
2010
+ },
2011
+ "256247": {
2012
+ "content": "<extra_id_52>",
2013
+ "lstrip": false,
2014
+ "normalized": false,
2015
+ "rstrip": false,
2016
+ "single_word": false,
2017
+ "special": true
2018
+ },
2019
+ "256248": {
2020
+ "content": "<extra_id_51>",
2021
+ "lstrip": false,
2022
+ "normalized": false,
2023
+ "rstrip": false,
2024
+ "single_word": false,
2025
+ "special": true
2026
+ },
2027
+ "256249": {
2028
+ "content": "<extra_id_50>",
2029
+ "lstrip": false,
2030
+ "normalized": false,
2031
+ "rstrip": false,
2032
+ "single_word": false,
2033
+ "special": true
2034
+ },
2035
+ "256250": {
2036
+ "content": "<extra_id_49>",
2037
+ "lstrip": false,
2038
+ "normalized": false,
2039
+ "rstrip": false,
2040
+ "single_word": false,
2041
+ "special": true
2042
+ },
2043
+ "256251": {
2044
+ "content": "<extra_id_48>",
2045
+ "lstrip": false,
2046
+ "normalized": false,
2047
+ "rstrip": false,
2048
+ "single_word": false,
2049
+ "special": true
2050
+ },
2051
+ "256252": {
2052
+ "content": "<extra_id_47>",
2053
+ "lstrip": false,
2054
+ "normalized": false,
2055
+ "rstrip": false,
2056
+ "single_word": false,
2057
+ "special": true
2058
+ },
2059
+ "256253": {
2060
+ "content": "<extra_id_46>",
2061
+ "lstrip": false,
2062
+ "normalized": false,
2063
+ "rstrip": false,
2064
+ "single_word": false,
2065
+ "special": true
2066
+ },
2067
+ "256254": {
2068
+ "content": "<extra_id_45>",
2069
+ "lstrip": false,
2070
+ "normalized": false,
2071
+ "rstrip": false,
2072
+ "single_word": false,
2073
+ "special": true
2074
+ },
2075
+ "256255": {
2076
+ "content": "<extra_id_44>",
2077
+ "lstrip": false,
2078
+ "normalized": false,
2079
+ "rstrip": false,
2080
+ "single_word": false,
2081
+ "special": true
2082
+ },
2083
+ "256256": {
2084
+ "content": "<extra_id_43>",
2085
+ "lstrip": false,
2086
+ "normalized": false,
2087
+ "rstrip": false,
2088
+ "single_word": false,
2089
+ "special": true
2090
+ },
2091
+ "256257": {
2092
+ "content": "<extra_id_42>",
2093
+ "lstrip": false,
2094
+ "normalized": false,
2095
+ "rstrip": false,
2096
+ "single_word": false,
2097
+ "special": true
2098
+ },
2099
+ "256258": {
2100
+ "content": "<extra_id_41>",
2101
+ "lstrip": false,
2102
+ "normalized": false,
2103
+ "rstrip": false,
2104
+ "single_word": false,
2105
+ "special": true
2106
+ },
2107
+ "256259": {
2108
+ "content": "<extra_id_40>",
2109
+ "lstrip": false,
2110
+ "normalized": false,
2111
+ "rstrip": false,
2112
+ "single_word": false,
2113
+ "special": true
2114
+ },
2115
+ "256260": {
2116
+ "content": "<extra_id_39>",
2117
+ "lstrip": false,
2118
+ "normalized": false,
2119
+ "rstrip": false,
2120
+ "single_word": false,
2121
+ "special": true
2122
+ },
2123
+ "256261": {
2124
+ "content": "<extra_id_38>",
2125
+ "lstrip": false,
2126
+ "normalized": false,
2127
+ "rstrip": false,
2128
+ "single_word": false,
2129
+ "special": true
2130
+ },
2131
+ "256262": {
2132
+ "content": "<extra_id_37>",
2133
+ "lstrip": false,
2134
+ "normalized": false,
2135
+ "rstrip": false,
2136
+ "single_word": false,
2137
+ "special": true
2138
+ },
2139
+ "256263": {
2140
+ "content": "<extra_id_36>",
2141
+ "lstrip": false,
2142
+ "normalized": false,
2143
+ "rstrip": false,
2144
+ "single_word": false,
2145
+ "special": true
2146
+ },
2147
+ "256264": {
2148
+ "content": "<extra_id_35>",
2149
+ "lstrip": false,
2150
+ "normalized": false,
2151
+ "rstrip": false,
2152
+ "single_word": false,
2153
+ "special": true
2154
+ },
2155
+ "256265": {
2156
+ "content": "<extra_id_34>",
2157
+ "lstrip": false,
2158
+ "normalized": false,
2159
+ "rstrip": false,
2160
+ "single_word": false,
2161
+ "special": true
2162
+ },
2163
+ "256266": {
2164
+ "content": "<extra_id_33>",
2165
+ "lstrip": false,
2166
+ "normalized": false,
2167
+ "rstrip": false,
2168
+ "single_word": false,
2169
+ "special": true
2170
+ },
2171
+ "256267": {
2172
+ "content": "<extra_id_32>",
2173
+ "lstrip": false,
2174
+ "normalized": false,
2175
+ "rstrip": false,
2176
+ "single_word": false,
2177
+ "special": true
2178
+ },
2179
+ "256268": {
2180
+ "content": "<extra_id_31>",
2181
+ "lstrip": false,
2182
+ "normalized": false,
2183
+ "rstrip": false,
2184
+ "single_word": false,
2185
+ "special": true
2186
+ },
2187
+ "256269": {
2188
+ "content": "<extra_id_30>",
2189
+ "lstrip": false,
2190
+ "normalized": false,
2191
+ "rstrip": false,
2192
+ "single_word": false,
2193
+ "special": true
2194
+ },
2195
+ "256270": {
2196
+ "content": "<extra_id_29>",
2197
+ "lstrip": false,
2198
+ "normalized": false,
2199
+ "rstrip": false,
2200
+ "single_word": false,
2201
+ "special": true
2202
+ },
2203
+ "256271": {
2204
+ "content": "<extra_id_28>",
2205
+ "lstrip": false,
2206
+ "normalized": false,
2207
+ "rstrip": false,
2208
+ "single_word": false,
2209
+ "special": true
2210
+ },
2211
+ "256272": {
2212
+ "content": "<extra_id_27>",
2213
+ "lstrip": false,
2214
+ "normalized": false,
2215
+ "rstrip": false,
2216
+ "single_word": false,
2217
+ "special": true
2218
+ },
2219
+ "256273": {
2220
+ "content": "<extra_id_26>",
2221
+ "lstrip": false,
2222
+ "normalized": false,
2223
+ "rstrip": false,
2224
+ "single_word": false,
2225
+ "special": true
2226
+ },
2227
+ "256274": {
2228
+ "content": "<extra_id_25>",
2229
+ "lstrip": false,
2230
+ "normalized": false,
2231
+ "rstrip": false,
2232
+ "single_word": false,
2233
+ "special": true
2234
+ },
2235
+ "256275": {
2236
+ "content": "<extra_id_24>",
2237
+ "lstrip": false,
2238
+ "normalized": false,
2239
+ "rstrip": false,
2240
+ "single_word": false,
2241
+ "special": true
2242
+ },
2243
+ "256276": {
2244
+ "content": "<extra_id_23>",
2245
+ "lstrip": false,
2246
+ "normalized": false,
2247
+ "rstrip": false,
2248
+ "single_word": false,
2249
+ "special": true
2250
+ },
2251
+ "256277": {
2252
+ "content": "<extra_id_22>",
2253
+ "lstrip": false,
2254
+ "normalized": false,
2255
+ "rstrip": false,
2256
+ "single_word": false,
2257
+ "special": true
2258
+ },
2259
+ "256278": {
2260
+ "content": "<extra_id_21>",
2261
+ "lstrip": false,
2262
+ "normalized": false,
2263
+ "rstrip": false,
2264
+ "single_word": false,
2265
+ "special": true
2266
+ },
2267
+ "256279": {
2268
+ "content": "<extra_id_20>",
2269
+ "lstrip": false,
2270
+ "normalized": false,
2271
+ "rstrip": false,
2272
+ "single_word": false,
2273
+ "special": true
2274
+ },
2275
+ "256280": {
2276
+ "content": "<extra_id_19>",
2277
+ "lstrip": false,
2278
+ "normalized": false,
2279
+ "rstrip": false,
2280
+ "single_word": false,
2281
+ "special": true
2282
+ },
2283
+ "256281": {
2284
+ "content": "<extra_id_18>",
2285
+ "lstrip": false,
2286
+ "normalized": false,
2287
+ "rstrip": false,
2288
+ "single_word": false,
2289
+ "special": true
2290
+ },
2291
+ "256282": {
2292
+ "content": "<extra_id_17>",
2293
+ "lstrip": false,
2294
+ "normalized": false,
2295
+ "rstrip": false,
2296
+ "single_word": false,
2297
+ "special": true
2298
+ },
2299
+ "256283": {
2300
+ "content": "<extra_id_16>",
2301
+ "lstrip": false,
2302
+ "normalized": false,
2303
+ "rstrip": false,
2304
+ "single_word": false,
2305
+ "special": true
2306
+ },
2307
+ "256284": {
2308
+ "content": "<extra_id_15>",
2309
+ "lstrip": false,
2310
+ "normalized": false,
2311
+ "rstrip": false,
2312
+ "single_word": false,
2313
+ "special": true
2314
+ },
2315
+ "256285": {
2316
+ "content": "<extra_id_14>",
2317
+ "lstrip": false,
2318
+ "normalized": false,
2319
+ "rstrip": false,
2320
+ "single_word": false,
2321
+ "special": true
2322
+ },
2323
+ "256286": {
2324
+ "content": "<extra_id_13>",
2325
+ "lstrip": false,
2326
+ "normalized": false,
2327
+ "rstrip": false,
2328
+ "single_word": false,
2329
+ "special": true
2330
+ },
2331
+ "256287": {
2332
+ "content": "<extra_id_12>",
2333
+ "lstrip": false,
2334
+ "normalized": false,
2335
+ "rstrip": false,
2336
+ "single_word": false,
2337
+ "special": true
2338
+ },
2339
+ "256288": {
2340
+ "content": "<extra_id_11>",
2341
+ "lstrip": false,
2342
+ "normalized": false,
2343
+ "rstrip": false,
2344
+ "single_word": false,
2345
+ "special": true
2346
+ },
2347
+ "256289": {
2348
+ "content": "<extra_id_10>",
2349
+ "lstrip": false,
2350
+ "normalized": false,
2351
+ "rstrip": false,
2352
+ "single_word": false,
2353
+ "special": true
2354
+ },
2355
+ "256290": {
2356
+ "content": "<extra_id_9>",
2357
+ "lstrip": false,
2358
+ "normalized": false,
2359
+ "rstrip": false,
2360
+ "single_word": false,
2361
+ "special": true
2362
+ },
2363
+ "256291": {
2364
+ "content": "<extra_id_8>",
2365
+ "lstrip": false,
2366
+ "normalized": false,
2367
+ "rstrip": false,
2368
+ "single_word": false,
2369
+ "special": true
2370
+ },
2371
+ "256292": {
2372
+ "content": "<extra_id_7>",
2373
+ "lstrip": false,
2374
+ "normalized": false,
2375
+ "rstrip": false,
2376
+ "single_word": false,
2377
+ "special": true
2378
+ },
2379
+ "256293": {
2380
+ "content": "<extra_id_6>",
2381
+ "lstrip": false,
2382
+ "normalized": false,
2383
+ "rstrip": false,
2384
+ "single_word": false,
2385
+ "special": true
2386
+ },
2387
+ "256294": {
2388
+ "content": "<extra_id_5>",
2389
+ "lstrip": false,
2390
+ "normalized": false,
2391
+ "rstrip": false,
2392
+ "single_word": false,
2393
+ "special": true
2394
+ },
2395
+ "256295": {
2396
+ "content": "<extra_id_4>",
2397
+ "lstrip": false,
2398
+ "normalized": false,
2399
+ "rstrip": false,
2400
+ "single_word": false,
2401
+ "special": true
2402
+ },
2403
+ "256296": {
2404
+ "content": "<extra_id_3>",
2405
+ "lstrip": false,
2406
+ "normalized": false,
2407
+ "rstrip": false,
2408
+ "single_word": false,
2409
+ "special": true
2410
+ },
2411
+ "256297": {
2412
+ "content": "<extra_id_2>",
2413
+ "lstrip": false,
2414
+ "normalized": false,
2415
+ "rstrip": false,
2416
+ "single_word": false,
2417
+ "special": true
2418
+ },
2419
+ "256298": {
2420
+ "content": "<extra_id_1>",
2421
+ "lstrip": false,
2422
+ "normalized": false,
2423
+ "rstrip": false,
2424
+ "single_word": false,
2425
+ "special": true
2426
+ },
2427
+ "256299": {
2428
+ "content": "<extra_id_0>",
2429
+ "lstrip": false,
2430
+ "normalized": false,
2431
+ "rstrip": false,
2432
+ "single_word": false,
2433
+ "special": true
2434
+ }
2435
+ },
2436
+ "additional_special_tokens": [
2437
+ "<extra_id_0>",
2438
+ "<extra_id_1>",
2439
+ "<extra_id_2>",
2440
+ "<extra_id_3>",
2441
+ "<extra_id_4>",
2442
+ "<extra_id_5>",
2443
+ "<extra_id_6>",
2444
+ "<extra_id_7>",
2445
+ "<extra_id_8>",
2446
+ "<extra_id_9>",
2447
+ "<extra_id_10>",
2448
+ "<extra_id_11>",
2449
+ "<extra_id_12>",
2450
+ "<extra_id_13>",
2451
+ "<extra_id_14>",
2452
+ "<extra_id_15>",
2453
+ "<extra_id_16>",
2454
+ "<extra_id_17>",
2455
+ "<extra_id_18>",
2456
+ "<extra_id_19>",
2457
+ "<extra_id_20>",
2458
+ "<extra_id_21>",
2459
+ "<extra_id_22>",
2460
+ "<extra_id_23>",
2461
+ "<extra_id_24>",
2462
+ "<extra_id_25>",
2463
+ "<extra_id_26>",
2464
+ "<extra_id_27>",
2465
+ "<extra_id_28>",
2466
+ "<extra_id_29>",
2467
+ "<extra_id_30>",
2468
+ "<extra_id_31>",
2469
+ "<extra_id_32>",
2470
+ "<extra_id_33>",
2471
+ "<extra_id_34>",
2472
+ "<extra_id_35>",
2473
+ "<extra_id_36>",
2474
+ "<extra_id_37>",
2475
+ "<extra_id_38>",
2476
+ "<extra_id_39>",
2477
+ "<extra_id_40>",
2478
+ "<extra_id_41>",
2479
+ "<extra_id_42>",
2480
+ "<extra_id_43>",
2481
+ "<extra_id_44>",
2482
+ "<extra_id_45>",
2483
+ "<extra_id_46>",
2484
+ "<extra_id_47>",
2485
+ "<extra_id_48>",
2486
+ "<extra_id_49>",
2487
+ "<extra_id_50>",
2488
+ "<extra_id_51>",
2489
+ "<extra_id_52>",
2490
+ "<extra_id_53>",
2491
+ "<extra_id_54>",
2492
+ "<extra_id_55>",
2493
+ "<extra_id_56>",
2494
+ "<extra_id_57>",
2495
+ "<extra_id_58>",
2496
+ "<extra_id_59>",
2497
+ "<extra_id_60>",
2498
+ "<extra_id_61>",
2499
+ "<extra_id_62>",
2500
+ "<extra_id_63>",
2501
+ "<extra_id_64>",
2502
+ "<extra_id_65>",
2503
+ "<extra_id_66>",
2504
+ "<extra_id_67>",
2505
+ "<extra_id_68>",
2506
+ "<extra_id_69>",
2507
+ "<extra_id_70>",
2508
+ "<extra_id_71>",
2509
+ "<extra_id_72>",
2510
+ "<extra_id_73>",
2511
+ "<extra_id_74>",
2512
+ "<extra_id_75>",
2513
+ "<extra_id_76>",
2514
+ "<extra_id_77>",
2515
+ "<extra_id_78>",
2516
+ "<extra_id_79>",
2517
+ "<extra_id_80>",
2518
+ "<extra_id_81>",
2519
+ "<extra_id_82>",
2520
+ "<extra_id_83>",
2521
+ "<extra_id_84>",
2522
+ "<extra_id_85>",
2523
+ "<extra_id_86>",
2524
+ "<extra_id_87>",
2525
+ "<extra_id_88>",
2526
+ "<extra_id_89>",
2527
+ "<extra_id_90>",
2528
+ "<extra_id_91>",
2529
+ "<extra_id_92>",
2530
+ "<extra_id_93>",
2531
+ "<extra_id_94>",
2532
+ "<extra_id_95>",
2533
+ "<extra_id_96>",
2534
+ "<extra_id_97>",
2535
+ "<extra_id_98>",
2536
+ "<extra_id_99>",
2537
+ "<extra_id_100>",
2538
+ "<extra_id_101>",
2539
+ "<extra_id_102>",
2540
+ "<extra_id_103>",
2541
+ "<extra_id_104>",
2542
+ "<extra_id_105>",
2543
+ "<extra_id_106>",
2544
+ "<extra_id_107>",
2545
+ "<extra_id_108>",
2546
+ "<extra_id_109>",
2547
+ "<extra_id_110>",
2548
+ "<extra_id_111>",
2549
+ "<extra_id_112>",
2550
+ "<extra_id_113>",
2551
+ "<extra_id_114>",
2552
+ "<extra_id_115>",
2553
+ "<extra_id_116>",
2554
+ "<extra_id_117>",
2555
+ "<extra_id_118>",
2556
+ "<extra_id_119>",
2557
+ "<extra_id_120>",
2558
+ "<extra_id_121>",
2559
+ "<extra_id_122>",
2560
+ "<extra_id_123>",
2561
+ "<extra_id_124>",
2562
+ "<extra_id_125>",
2563
+ "<extra_id_126>",
2564
+ "<extra_id_127>",
2565
+ "<extra_id_128>",
2566
+ "<extra_id_129>",
2567
+ "<extra_id_130>",
2568
+ "<extra_id_131>",
2569
+ "<extra_id_132>",
2570
+ "<extra_id_133>",
2571
+ "<extra_id_134>",
2572
+ "<extra_id_135>",
2573
+ "<extra_id_136>",
2574
+ "<extra_id_137>",
2575
+ "<extra_id_138>",
2576
+ "<extra_id_139>",
2577
+ "<extra_id_140>",
2578
+ "<extra_id_141>",
2579
+ "<extra_id_142>",
2580
+ "<extra_id_143>",
2581
+ "<extra_id_144>",
2582
+ "<extra_id_145>",
2583
+ "<extra_id_146>",
2584
+ "<extra_id_147>",
2585
+ "<extra_id_148>",
2586
+ "<extra_id_149>",
2587
+ "<extra_id_150>",
2588
+ "<extra_id_151>",
2589
+ "<extra_id_152>",
2590
+ "<extra_id_153>",
2591
+ "<extra_id_154>",
2592
+ "<extra_id_155>",
2593
+ "<extra_id_156>",
2594
+ "<extra_id_157>",
2595
+ "<extra_id_158>",
2596
+ "<extra_id_159>",
2597
+ "<extra_id_160>",
2598
+ "<extra_id_161>",
2599
+ "<extra_id_162>",
2600
+ "<extra_id_163>",
2601
+ "<extra_id_164>",
2602
+ "<extra_id_165>",
2603
+ "<extra_id_166>",
2604
+ "<extra_id_167>",
2605
+ "<extra_id_168>",
2606
+ "<extra_id_169>",
2607
+ "<extra_id_170>",
2608
+ "<extra_id_171>",
2609
+ "<extra_id_172>",
2610
+ "<extra_id_173>",
2611
+ "<extra_id_174>",
2612
+ "<extra_id_175>",
2613
+ "<extra_id_176>",
2614
+ "<extra_id_177>",
2615
+ "<extra_id_178>",
2616
+ "<extra_id_179>",
2617
+ "<extra_id_180>",
2618
+ "<extra_id_181>",
2619
+ "<extra_id_182>",
2620
+ "<extra_id_183>",
2621
+ "<extra_id_184>",
2622
+ "<extra_id_185>",
2623
+ "<extra_id_186>",
2624
+ "<extra_id_187>",
2625
+ "<extra_id_188>",
2626
+ "<extra_id_189>",
2627
+ "<extra_id_190>",
2628
+ "<extra_id_191>",
2629
+ "<extra_id_192>",
2630
+ "<extra_id_193>",
2631
+ "<extra_id_194>",
2632
+ "<extra_id_195>",
2633
+ "<extra_id_196>",
2634
+ "<extra_id_197>",
2635
+ "<extra_id_198>",
2636
+ "<extra_id_199>",
2637
+ "<extra_id_200>",
2638
+ "<extra_id_201>",
2639
+ "<extra_id_202>",
2640
+ "<extra_id_203>",
2641
+ "<extra_id_204>",
2642
+ "<extra_id_205>",
2643
+ "<extra_id_206>",
2644
+ "<extra_id_207>",
2645
+ "<extra_id_208>",
2646
+ "<extra_id_209>",
2647
+ "<extra_id_210>",
2648
+ "<extra_id_211>",
2649
+ "<extra_id_212>",
2650
+ "<extra_id_213>",
2651
+ "<extra_id_214>",
2652
+ "<extra_id_215>",
2653
+ "<extra_id_216>",
2654
+ "<extra_id_217>",
2655
+ "<extra_id_218>",
2656
+ "<extra_id_219>",
2657
+ "<extra_id_220>",
2658
+ "<extra_id_221>",
2659
+ "<extra_id_222>",
2660
+ "<extra_id_223>",
2661
+ "<extra_id_224>",
2662
+ "<extra_id_225>",
2663
+ "<extra_id_226>",
2664
+ "<extra_id_227>",
2665
+ "<extra_id_228>",
2666
+ "<extra_id_229>",
2667
+ "<extra_id_230>",
2668
+ "<extra_id_231>",
2669
+ "<extra_id_232>",
2670
+ "<extra_id_233>",
2671
+ "<extra_id_234>",
2672
+ "<extra_id_235>",
2673
+ "<extra_id_236>",
2674
+ "<extra_id_237>",
2675
+ "<extra_id_238>",
2676
+ "<extra_id_239>",
2677
+ "<extra_id_240>",
2678
+ "<extra_id_241>",
2679
+ "<extra_id_242>",
2680
+ "<extra_id_243>",
2681
+ "<extra_id_244>",
2682
+ "<extra_id_245>",
2683
+ "<extra_id_246>",
2684
+ "<extra_id_247>",
2685
+ "<extra_id_248>",
2686
+ "<extra_id_249>",
2687
+ "<extra_id_250>",
2688
+ "<extra_id_251>",
2689
+ "<extra_id_252>",
2690
+ "<extra_id_253>",
2691
+ "<extra_id_254>",
2692
+ "<extra_id_255>",
2693
+ "<extra_id_256>",
2694
+ "<extra_id_257>",
2695
+ "<extra_id_258>",
2696
+ "<extra_id_259>",
2697
+ "<extra_id_260>",
2698
+ "<extra_id_261>",
2699
+ "<extra_id_262>",
2700
+ "<extra_id_263>",
2701
+ "<extra_id_264>",
2702
+ "<extra_id_265>",
2703
+ "<extra_id_266>",
2704
+ "<extra_id_267>",
2705
+ "<extra_id_268>",
2706
+ "<extra_id_269>",
2707
+ "<extra_id_270>",
2708
+ "<extra_id_271>",
2709
+ "<extra_id_272>",
2710
+ "<extra_id_273>",
2711
+ "<extra_id_274>",
2712
+ "<extra_id_275>",
2713
+ "<extra_id_276>",
2714
+ "<extra_id_277>",
2715
+ "<extra_id_278>",
2716
+ "<extra_id_279>",
2717
+ "<extra_id_280>",
2718
+ "<extra_id_281>",
2719
+ "<extra_id_282>",
2720
+ "<extra_id_283>",
2721
+ "<extra_id_284>",
2722
+ "<extra_id_285>",
2723
+ "<extra_id_286>",
2724
+ "<extra_id_287>",
2725
+ "<extra_id_288>",
2726
+ "<extra_id_289>",
2727
+ "<extra_id_290>",
2728
+ "<extra_id_291>",
2729
+ "<extra_id_292>",
2730
+ "<extra_id_293>",
2731
+ "<extra_id_294>",
2732
+ "<extra_id_295>",
2733
+ "<extra_id_296>",
2734
+ "<extra_id_297>",
2735
+ "<extra_id_298>",
2736
+ "<extra_id_299>"
2737
+ ],
2738
+ "bos_token": "<s>",
2739
+ "clean_up_tokenization_spaces": true,
2740
+ "eos_token": "</s>",
2741
+ "extra_ids": 300,
2742
+ "model_max_length": 1000000000000000019884624838656,
2743
+ "pad_token": "<pad>",
2744
+ "sp_model_kwargs": {},
2745
+ "spaces_between_special_tokens": false,
2746
+ "tokenizer_class": "T5Tokenizer",
2747
+ "unk_token": "<unk>"
2748
+ }
models_t5_umt5-xxl-enc-bf16.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cace0da2b446bbbbc57d031ab6cf163a3d59b366da94e5afe36745b746fd81d
3
+ size 11361920418