Upload folder using huggingface_hub
Browse files- .gitattributes +9 -0
- chat_template.jinja +52 -0
- config.json +48 -0
- generation_config.json +8 -0
- onnx/model.onnx +3 -0
- onnx/model.onnx_data +3 -0
- onnx/model.onnx_data_1 +3 -0
- onnx/model.onnx_data_2 +3 -0
- onnx/model.onnx_data_3 +3 -0
- onnx/model_fp16.onnx +3 -0
- onnx/model_fp16.onnx_data +3 -0
- onnx/model_fp16.onnx_data_1 +3 -0
- onnx/model_q4.onnx +3 -0
- onnx/model_q4.onnx_data +3 -0
- onnx/model_q4.onnx_data_1 +3 -0
- onnx/model_q4f16.onnx +3 -0
- onnx/model_q4f16.onnx_data +3 -0
- special_tokens_map.json +32 -0
- tokenizer.json +0 -0
- tokenizer_config.json +97 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
onnx/model.onnx_data filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
onnx/model.onnx_data_1 filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
onnx/model.onnx_data_2 filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
onnx/model.onnx_data_3 filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
onnx/model_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
onnx/model_fp16.onnx_data_1 filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
onnx/model_q4.onnx_data filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
onnx/model_q4.onnx_data_1 filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
onnx/model_q4f16.onnx_data filter=lfs diff=lfs merge=lfs -text
|
chat_template.jinja
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{{- bos_token }}
|
| 2 |
+
{%- if messages[0]['role'] == 'system' -%}
|
| 3 |
+
{%- if messages[1] is not defined or messages[1]['role'] != 'user' -%}
|
| 4 |
+
{{- raise_exception("System message must be followed by a user message") }}
|
| 5 |
+
{%- endif -%}
|
| 6 |
+
{%- set system_content = messages[0]['content'] -%}
|
| 7 |
+
{%- set loop_messages = messages[1:] -%}
|
| 8 |
+
{%- set first_user_has_system = true -%}
|
| 9 |
+
{%- else -%}
|
| 10 |
+
{%- set system_content = "" -%}
|
| 11 |
+
{%- set loop_messages = messages -%}
|
| 12 |
+
{%- set first_user_has_system = false -%}
|
| 13 |
+
{%- endif -%}
|
| 14 |
+
{%- for message in loop_messages -%}
|
| 15 |
+
{%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}
|
| 16 |
+
{{- raise_exception("Conversation roles must alternate user/assistant/user/assistant/...") }}
|
| 17 |
+
{%- endif -%}
|
| 18 |
+
{%- if message['role'] == 'user' -%}
|
| 19 |
+
{{- '<|user_start|>' }}
|
| 20 |
+
{%- if loop.first and first_user_has_system -%}
|
| 21 |
+
{{- system_content + '\n\n' }}
|
| 22 |
+
{%- endif -%}
|
| 23 |
+
{%- if message['content'] is not string -%}
|
| 24 |
+
{{- raise_exception("User messages must contain string content") }}
|
| 25 |
+
{%- endif -%}
|
| 26 |
+
{{- message['content'] }}
|
| 27 |
+
{{- '<|user_end|>' }}
|
| 28 |
+
{%- elif message['role'] == 'assistant' -%}
|
| 29 |
+
{{- '<|assistant_start|>' }}
|
| 30 |
+
{%- if message['content'] is string -%}
|
| 31 |
+
{{- message['content'] }}
|
| 32 |
+
{%- elif message['content'] is iterable -%}
|
| 33 |
+
{%- for part in message['content'] -%}
|
| 34 |
+
{%- if part['type'] == 'text' -%}
|
| 35 |
+
{{- part.get('text', '') }}
|
| 36 |
+
{%- elif part['type'] == 'python' -%}
|
| 37 |
+
{{- '<|python_start|>' + part.get('text', '') + '<|python_end|>' }}
|
| 38 |
+
{%- elif part['type'] == 'python_output' -%}
|
| 39 |
+
{{- '<|output_start|>' + part.get('text', '') + '<|output_end|>' }}
|
| 40 |
+
{%- else -%}
|
| 41 |
+
{{- raise_exception("Unknown assistant content part: " + part['type']) }}
|
| 42 |
+
{%- endif -%}
|
| 43 |
+
{%- endfor -%}
|
| 44 |
+
{%- else -%}
|
| 45 |
+
{{- raise_exception("Unsupported assistant content type") }}
|
| 46 |
+
{%- endif -%}
|
| 47 |
+
{{- '<|assistant_end|>' }}
|
| 48 |
+
{%- endif -%}
|
| 49 |
+
{%- endfor -%}
|
| 50 |
+
{%- if add_generation_prompt -%}
|
| 51 |
+
{{- '<|assistant_start|>' }}
|
| 52 |
+
{%- endif -%}
|
config.json
ADDED
|
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"NanoChatForCausalLM"
|
| 4 |
+
],
|
| 5 |
+
"attention_dropout": 0.0,
|
| 6 |
+
"bos_token": "<|bos|>",
|
| 7 |
+
"bos_token_id": 0,
|
| 8 |
+
"chat_template": "{% if messages[0]['role'] == 'system' %}<|bos|><|user_start|>{{ messages[0]['content'] }}\n\n{{ messages[1]['content'] }}<|user_end|>{% set messages = messages[2:] %}{% else %}<|bos|>{% endif %}{% for message in messages %}{% if loop.index0 % 2 == 0 %}<|user_start|>{{ message['content'] }}<|user_end|>{% else %}<|assistant_start|>{{ message['content'] }}<|assistant_end|>{% endif %}{% endfor %}",
|
| 9 |
+
"dtype": "bfloat16",
|
| 10 |
+
"eos_token": "<|assistant_end|>",
|
| 11 |
+
"eos_token_id": 65531,
|
| 12 |
+
"hidden_act": "relu2",
|
| 13 |
+
"hidden_size": 2048,
|
| 14 |
+
"initializer_range": 0.02,
|
| 15 |
+
"intermediate_size": 8192,
|
| 16 |
+
"logits_soft_cap": 15.0,
|
| 17 |
+
"max_position_embeddings": 2048,
|
| 18 |
+
"model_type": "nanochat",
|
| 19 |
+
"n_embd": 2048,
|
| 20 |
+
"n_head": 16,
|
| 21 |
+
"n_layer": 32,
|
| 22 |
+
"num_attention_heads": 16,
|
| 23 |
+
"num_hidden_layers": 32,
|
| 24 |
+
"num_key_value_heads": 16,
|
| 25 |
+
"pad_token": "<|assistant_end|>",
|
| 26 |
+
"pad_token_id": 65531,
|
| 27 |
+
"qkv_bias": false,
|
| 28 |
+
"resid_dropout": 0.0,
|
| 29 |
+
"rms_norm_eps": 1e-06,
|
| 30 |
+
"rope_scaling": null,
|
| 31 |
+
"rope_theta": 10000.0,
|
| 32 |
+
"tie_word_embeddings": false,
|
| 33 |
+
"transformers_version": "4.57.0",
|
| 34 |
+
"use_cache": true,
|
| 35 |
+
"vocab_size": 65536,
|
| 36 |
+
"transformers.js_config": {
|
| 37 |
+
"use_external_data_format": {
|
| 38 |
+
"model.onnx": 4,
|
| 39 |
+
"model_fp16.onnx": 2,
|
| 40 |
+
"model_q4.onnx": 2,
|
| 41 |
+
"model_q4f16.onnx": 1
|
| 42 |
+
},
|
| 43 |
+
"kv_cache_dtype": {
|
| 44 |
+
"q4f16": "float16",
|
| 45 |
+
"fp16": "float16"
|
| 46 |
+
}
|
| 47 |
+
}
|
| 48 |
+
}
|
generation_config.json
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token_id": 50256,
|
| 3 |
+
"eos_token_id": [
|
| 4 |
+
50256,
|
| 5 |
+
65531
|
| 6 |
+
],
|
| 7 |
+
"transformers_version": "4.57.0"
|
| 8 |
+
}
|
onnx/model.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c677af793b923e543aad0342a4b061c14e1d6be0ca8d924764279bbb0c598cc4
|
| 3 |
+
size 465675
|
onnx/model.onnx_data
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cd405d4e79f9fbecef3f263ba5b0e9111237436235f7d45d9eff0131a38052ee
|
| 3 |
+
size 2081562624
|
onnx/model.onnx_data_1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:17ec1a458031beb774bd714dc8498a58171605486eaabe988c08e5a8159b5e56
|
| 3 |
+
size 2080546816
|
onnx/model.onnx_data_2
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:65d5d764574b10c62ba75efc9c8b8de859bf752a49ed001e02fde05becf05bf2
|
| 3 |
+
size 2080546816
|
onnx/model.onnx_data_3
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0eaa7769e344e834be8cfe3d7ff9d1bffc6e6d927b5c50b39134dc6b594528fe
|
| 3 |
+
size 1275125760
|
onnx/model_fp16.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:19b66a4e151601442ee3e14976dd20c13053c3a4bd357ae987e1e7f44151df3d
|
| 3 |
+
size 560294
|
onnx/model_fp16.onnx_data
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3a0a1d5ec6e921fc912def8f7162609230bf4c9be4509cc1e136d9bbc87c04fe
|
| 3 |
+
size 2089598976
|
onnx/model_fp16.onnx_data_1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ef4617f06937f7c88e2a60619f2d900c08508b574a72719707c494831052f9c8
|
| 3 |
+
size 1669562368
|
onnx/model_q4.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9e94739ecf6a1e3ff12e420dd870c971fb6bcfe1d1a6e6277e16780e7bdfcfa1
|
| 3 |
+
size 543776
|
onnx/model_q4.onnx_data
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f11242e6687f1fe27d3621052caf112d76ab6d9f36d1324a51c8be7d4c93c086
|
| 3 |
+
size 2095529984
|
onnx/model_q4.onnx_data_1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:897e660a4658272e64020ca4ee403a3e9ac9176dda8e0ca5003897c335eb9095
|
| 3 |
+
size 744292352
|
onnx/model_q4f16.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:729175f0c09338c6cbfbd12be4b9a3eff69e2aa32f85a759295666675b7b617c
|
| 3 |
+
size 638756
|
onnx/model_q4f16.onnx_data
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:876fe7d956f5f968eb9957199ef612e05428180ce02a11c6ad93e4bebabf0717
|
| 3 |
+
size 1761296384
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"additional_special_tokens": [
|
| 3 |
+
"<|user_start|>",
|
| 4 |
+
"<|user_end|>",
|
| 5 |
+
"<|assistant_start|>",
|
| 6 |
+
"<|python_start|>",
|
| 7 |
+
"<|python_end|>",
|
| 8 |
+
"<|output_start|>",
|
| 9 |
+
"<|output_end|>"
|
| 10 |
+
],
|
| 11 |
+
"bos_token": {
|
| 12 |
+
"content": "<|bos|>",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false
|
| 17 |
+
},
|
| 18 |
+
"eos_token": {
|
| 19 |
+
"content": "<|assistant_end|>",
|
| 20 |
+
"lstrip": false,
|
| 21 |
+
"normalized": false,
|
| 22 |
+
"rstrip": false,
|
| 23 |
+
"single_word": false
|
| 24 |
+
},
|
| 25 |
+
"pad_token": {
|
| 26 |
+
"content": "<|assistant_end|>",
|
| 27 |
+
"lstrip": false,
|
| 28 |
+
"normalized": false,
|
| 29 |
+
"rstrip": false,
|
| 30 |
+
"single_word": false
|
| 31 |
+
}
|
| 32 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,97 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"65527": {
|
| 4 |
+
"content": "<|bos|>",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"65528": {
|
| 12 |
+
"content": "<|user_start|>",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"65529": {
|
| 20 |
+
"content": "<|user_end|>",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"65530": {
|
| 28 |
+
"content": "<|assistant_start|>",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"65531": {
|
| 36 |
+
"content": "<|assistant_end|>",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
},
|
| 43 |
+
"65532": {
|
| 44 |
+
"content": "<|python_start|>",
|
| 45 |
+
"lstrip": false,
|
| 46 |
+
"normalized": false,
|
| 47 |
+
"rstrip": false,
|
| 48 |
+
"single_word": false,
|
| 49 |
+
"special": true
|
| 50 |
+
},
|
| 51 |
+
"65533": {
|
| 52 |
+
"content": "<|python_end|>",
|
| 53 |
+
"lstrip": false,
|
| 54 |
+
"normalized": false,
|
| 55 |
+
"rstrip": false,
|
| 56 |
+
"single_word": false,
|
| 57 |
+
"special": true
|
| 58 |
+
},
|
| 59 |
+
"65534": {
|
| 60 |
+
"content": "<|output_start|>",
|
| 61 |
+
"lstrip": false,
|
| 62 |
+
"normalized": false,
|
| 63 |
+
"rstrip": false,
|
| 64 |
+
"single_word": false,
|
| 65 |
+
"special": true
|
| 66 |
+
},
|
| 67 |
+
"65535": {
|
| 68 |
+
"content": "<|output_end|>",
|
| 69 |
+
"lstrip": false,
|
| 70 |
+
"normalized": false,
|
| 71 |
+
"rstrip": false,
|
| 72 |
+
"single_word": false,
|
| 73 |
+
"special": true
|
| 74 |
+
}
|
| 75 |
+
},
|
| 76 |
+
"additional_special_tokens": [
|
| 77 |
+
"<|user_start|>",
|
| 78 |
+
"<|user_end|>",
|
| 79 |
+
"<|assistant_start|>",
|
| 80 |
+
"<|python_start|>",
|
| 81 |
+
"<|python_end|>",
|
| 82 |
+
"<|output_start|>",
|
| 83 |
+
"<|output_end|>"
|
| 84 |
+
],
|
| 85 |
+
"bos_token": "<|bos|>",
|
| 86 |
+
"clean_up_tokenization_spaces": false,
|
| 87 |
+
"eos_token": "<|assistant_end|>",
|
| 88 |
+
"extra_special_tokens": {},
|
| 89 |
+
"model_input_names": [
|
| 90 |
+
"input_ids",
|
| 91 |
+
"attention_mask"
|
| 92 |
+
],
|
| 93 |
+
"model_max_length": 1000000000000000019884624838656,
|
| 94 |
+
"pad_token": "<|assistant_end|>",
|
| 95 |
+
"tokenizer_class": "PreTrainedTokenizerFast",
|
| 96 |
+
"chat_template": "{{- bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[1] is not defined or messages[1]['role'] != 'user' -%}\n {{- raise_exception(\"System message must be followed by a user message\") }}\n {%- endif -%}\n {%- set system_content = messages[0]['content'] -%}\n {%- set loop_messages = messages[1:] -%}\n {%- set first_user_has_system = true -%}\n{%- else -%}\n {%- set system_content = \"\" -%}\n {%- set loop_messages = messages -%}\n {%- set first_user_has_system = false -%}\n{%- endif -%}\n{%- for message in loop_messages -%}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}\n {{- raise_exception(\"Conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif -%}\n {%- if message['role'] == 'user' -%}\n {{- '<|user_start|>' }}\n {%- if loop.first and first_user_has_system -%}\n {{- system_content + '\\n\\n' }}\n {%- endif -%}\n {%- if message['content'] is not string -%}\n {{- raise_exception(\"User messages must contain string content\") }}\n {%- endif -%}\n {{- message['content'] }}\n {{- '<|user_end|>' }}\n {%- elif message['role'] == 'assistant' -%}\n {{- '<|assistant_start|>' }}\n {%- if message['content'] is string -%}\n {{- message['content'] }}\n {%- elif message['content'] is iterable -%}\n {%- for part in message['content'] -%}\n {%- if part['type'] == 'text' -%}\n {{- part.get('text', '') }}\n {%- elif part['type'] == 'python' -%}\n {{- '<|python_start|>' + part.get('text', '') + '<|python_end|>' }}\n {%- elif part['type'] == 'python_output' -%}\n {{- '<|output_start|>' + part.get('text', '') + '<|output_end|>' }}\n {%- else -%}\n {{- raise_exception(\"Unknown assistant content part: \" + part['type']) }}\n {%- endif -%}\n {%- endfor -%}\n {%- else -%}\n {{- raise_exception(\"Unsupported assistant content type\") }}\n {%- endif -%}\n {{- '<|assistant_end|>' }}\n {%- endif -%}\n{%- endfor -%}\n{%- if add_generation_prompt -%}\n {{- '<|assistant_start|>' }}\n{%- endif -%}\n"
|
| 97 |
+
}
|