Xenova HF Staff commited on
Commit
5e500c2
·
verified ·
1 Parent(s): 50fa719

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ onnx/model.onnx_data filter=lfs diff=lfs merge=lfs -text
37
+ onnx/model.onnx_data_1 filter=lfs diff=lfs merge=lfs -text
38
+ onnx/model.onnx_data_2 filter=lfs diff=lfs merge=lfs -text
39
+ onnx/model.onnx_data_3 filter=lfs diff=lfs merge=lfs -text
40
+ onnx/model_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
41
+ onnx/model_fp16.onnx_data_1 filter=lfs diff=lfs merge=lfs -text
42
+ onnx/model_q4.onnx_data filter=lfs diff=lfs merge=lfs -text
43
+ onnx/model_q4.onnx_data_1 filter=lfs diff=lfs merge=lfs -text
44
+ onnx/model_q4f16.onnx_data filter=lfs diff=lfs merge=lfs -text
chat_template.jinja ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {{- bos_token }}
2
+ {%- if messages[0]['role'] == 'system' -%}
3
+ {%- if messages[1] is not defined or messages[1]['role'] != 'user' -%}
4
+ {{- raise_exception("System message must be followed by a user message") }}
5
+ {%- endif -%}
6
+ {%- set system_content = messages[0]['content'] -%}
7
+ {%- set loop_messages = messages[1:] -%}
8
+ {%- set first_user_has_system = true -%}
9
+ {%- else -%}
10
+ {%- set system_content = "" -%}
11
+ {%- set loop_messages = messages -%}
12
+ {%- set first_user_has_system = false -%}
13
+ {%- endif -%}
14
+ {%- for message in loop_messages -%}
15
+ {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}
16
+ {{- raise_exception("Conversation roles must alternate user/assistant/user/assistant/...") }}
17
+ {%- endif -%}
18
+ {%- if message['role'] == 'user' -%}
19
+ {{- '<|user_start|>' }}
20
+ {%- if loop.first and first_user_has_system -%}
21
+ {{- system_content + '\n\n' }}
22
+ {%- endif -%}
23
+ {%- if message['content'] is not string -%}
24
+ {{- raise_exception("User messages must contain string content") }}
25
+ {%- endif -%}
26
+ {{- message['content'] }}
27
+ {{- '<|user_end|>' }}
28
+ {%- elif message['role'] == 'assistant' -%}
29
+ {{- '<|assistant_start|>' }}
30
+ {%- if message['content'] is string -%}
31
+ {{- message['content'] }}
32
+ {%- elif message['content'] is iterable -%}
33
+ {%- for part in message['content'] -%}
34
+ {%- if part['type'] == 'text' -%}
35
+ {{- part.get('text', '') }}
36
+ {%- elif part['type'] == 'python' -%}
37
+ {{- '<|python_start|>' + part.get('text', '') + '<|python_end|>' }}
38
+ {%- elif part['type'] == 'python_output' -%}
39
+ {{- '<|output_start|>' + part.get('text', '') + '<|output_end|>' }}
40
+ {%- else -%}
41
+ {{- raise_exception("Unknown assistant content part: " + part['type']) }}
42
+ {%- endif -%}
43
+ {%- endfor -%}
44
+ {%- else -%}
45
+ {{- raise_exception("Unsupported assistant content type") }}
46
+ {%- endif -%}
47
+ {{- '<|assistant_end|>' }}
48
+ {%- endif -%}
49
+ {%- endfor -%}
50
+ {%- if add_generation_prompt -%}
51
+ {{- '<|assistant_start|>' }}
52
+ {%- endif -%}
config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "NanoChatForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token": "<|bos|>",
7
+ "bos_token_id": 0,
8
+ "chat_template": "{% if messages[0]['role'] == 'system' %}<|bos|><|user_start|>{{ messages[0]['content'] }}\n\n{{ messages[1]['content'] }}<|user_end|>{% set messages = messages[2:] %}{% else %}<|bos|>{% endif %}{% for message in messages %}{% if loop.index0 % 2 == 0 %}<|user_start|>{{ message['content'] }}<|user_end|>{% else %}<|assistant_start|>{{ message['content'] }}<|assistant_end|>{% endif %}{% endfor %}",
9
+ "dtype": "bfloat16",
10
+ "eos_token": "<|assistant_end|>",
11
+ "eos_token_id": 65531,
12
+ "hidden_act": "relu2",
13
+ "hidden_size": 2048,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 8192,
16
+ "logits_soft_cap": 15.0,
17
+ "max_position_embeddings": 2048,
18
+ "model_type": "nanochat",
19
+ "n_embd": 2048,
20
+ "n_head": 16,
21
+ "n_layer": 32,
22
+ "num_attention_heads": 16,
23
+ "num_hidden_layers": 32,
24
+ "num_key_value_heads": 16,
25
+ "pad_token": "<|assistant_end|>",
26
+ "pad_token_id": 65531,
27
+ "qkv_bias": false,
28
+ "resid_dropout": 0.0,
29
+ "rms_norm_eps": 1e-06,
30
+ "rope_scaling": null,
31
+ "rope_theta": 10000.0,
32
+ "tie_word_embeddings": false,
33
+ "transformers_version": "4.57.0",
34
+ "use_cache": true,
35
+ "vocab_size": 65536,
36
+ "transformers.js_config": {
37
+ "use_external_data_format": {
38
+ "model.onnx": 4,
39
+ "model_fp16.onnx": 2,
40
+ "model_q4.onnx": 2,
41
+ "model_q4f16.onnx": 1
42
+ },
43
+ "kv_cache_dtype": {
44
+ "q4f16": "float16",
45
+ "fp16": "float16"
46
+ }
47
+ }
48
+ }
generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 50256,
3
+ "eos_token_id": [
4
+ 50256,
5
+ 65531
6
+ ],
7
+ "transformers_version": "4.57.0"
8
+ }
onnx/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c677af793b923e543aad0342a4b061c14e1d6be0ca8d924764279bbb0c598cc4
3
+ size 465675
onnx/model.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd405d4e79f9fbecef3f263ba5b0e9111237436235f7d45d9eff0131a38052ee
3
+ size 2081562624
onnx/model.onnx_data_1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17ec1a458031beb774bd714dc8498a58171605486eaabe988c08e5a8159b5e56
3
+ size 2080546816
onnx/model.onnx_data_2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65d5d764574b10c62ba75efc9c8b8de859bf752a49ed001e02fde05becf05bf2
3
+ size 2080546816
onnx/model.onnx_data_3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eaa7769e344e834be8cfe3d7ff9d1bffc6e6d927b5c50b39134dc6b594528fe
3
+ size 1275125760
onnx/model_fp16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19b66a4e151601442ee3e14976dd20c13053c3a4bd357ae987e1e7f44151df3d
3
+ size 560294
onnx/model_fp16.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a0a1d5ec6e921fc912def8f7162609230bf4c9be4509cc1e136d9bbc87c04fe
3
+ size 2089598976
onnx/model_fp16.onnx_data_1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef4617f06937f7c88e2a60619f2d900c08508b574a72719707c494831052f9c8
3
+ size 1669562368
onnx/model_q4.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e94739ecf6a1e3ff12e420dd870c971fb6bcfe1d1a6e6277e16780e7bdfcfa1
3
+ size 543776
onnx/model_q4.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f11242e6687f1fe27d3621052caf112d76ab6d9f36d1324a51c8be7d4c93c086
3
+ size 2095529984
onnx/model_q4.onnx_data_1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:897e660a4658272e64020ca4ee403a3e9ac9176dda8e0ca5003897c335eb9095
3
+ size 744292352
onnx/model_q4f16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:729175f0c09338c6cbfbd12be4b9a3eff69e2aa32f85a759295666675b7b617c
3
+ size 638756
onnx/model_q4f16.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:876fe7d956f5f968eb9957199ef612e05428180ce02a11c6ad93e4bebabf0717
3
+ size 1761296384
special_tokens_map.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|user_start|>",
4
+ "<|user_end|>",
5
+ "<|assistant_start|>",
6
+ "<|python_start|>",
7
+ "<|python_end|>",
8
+ "<|output_start|>",
9
+ "<|output_end|>"
10
+ ],
11
+ "bos_token": {
12
+ "content": "<|bos|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false
17
+ },
18
+ "eos_token": {
19
+ "content": "<|assistant_end|>",
20
+ "lstrip": false,
21
+ "normalized": false,
22
+ "rstrip": false,
23
+ "single_word": false
24
+ },
25
+ "pad_token": {
26
+ "content": "<|assistant_end|>",
27
+ "lstrip": false,
28
+ "normalized": false,
29
+ "rstrip": false,
30
+ "single_word": false
31
+ }
32
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "65527": {
4
+ "content": "<|bos|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "65528": {
12
+ "content": "<|user_start|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "65529": {
20
+ "content": "<|user_end|>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "65530": {
28
+ "content": "<|assistant_start|>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "65531": {
36
+ "content": "<|assistant_end|>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "65532": {
44
+ "content": "<|python_start|>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "65533": {
52
+ "content": "<|python_end|>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "65534": {
60
+ "content": "<|output_start|>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "65535": {
68
+ "content": "<|output_end|>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ }
75
+ },
76
+ "additional_special_tokens": [
77
+ "<|user_start|>",
78
+ "<|user_end|>",
79
+ "<|assistant_start|>",
80
+ "<|python_start|>",
81
+ "<|python_end|>",
82
+ "<|output_start|>",
83
+ "<|output_end|>"
84
+ ],
85
+ "bos_token": "<|bos|>",
86
+ "clean_up_tokenization_spaces": false,
87
+ "eos_token": "<|assistant_end|>",
88
+ "extra_special_tokens": {},
89
+ "model_input_names": [
90
+ "input_ids",
91
+ "attention_mask"
92
+ ],
93
+ "model_max_length": 1000000000000000019884624838656,
94
+ "pad_token": "<|assistant_end|>",
95
+ "tokenizer_class": "PreTrainedTokenizerFast",
96
+ "chat_template": "{{- bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[1] is not defined or messages[1]['role'] != 'user' -%}\n {{- raise_exception(\"System message must be followed by a user message\") }}\n {%- endif -%}\n {%- set system_content = messages[0]['content'] -%}\n {%- set loop_messages = messages[1:] -%}\n {%- set first_user_has_system = true -%}\n{%- else -%}\n {%- set system_content = \"\" -%}\n {%- set loop_messages = messages -%}\n {%- set first_user_has_system = false -%}\n{%- endif -%}\n{%- for message in loop_messages -%}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}\n {{- raise_exception(\"Conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif -%}\n {%- if message['role'] == 'user' -%}\n {{- '<|user_start|>' }}\n {%- if loop.first and first_user_has_system -%}\n {{- system_content + '\\n\\n' }}\n {%- endif -%}\n {%- if message['content'] is not string -%}\n {{- raise_exception(\"User messages must contain string content\") }}\n {%- endif -%}\n {{- message['content'] }}\n {{- '<|user_end|>' }}\n {%- elif message['role'] == 'assistant' -%}\n {{- '<|assistant_start|>' }}\n {%- if message['content'] is string -%}\n {{- message['content'] }}\n {%- elif message['content'] is iterable -%}\n {%- for part in message['content'] -%}\n {%- if part['type'] == 'text' -%}\n {{- part.get('text', '') }}\n {%- elif part['type'] == 'python' -%}\n {{- '<|python_start|>' + part.get('text', '') + '<|python_end|>' }}\n {%- elif part['type'] == 'python_output' -%}\n {{- '<|output_start|>' + part.get('text', '') + '<|output_end|>' }}\n {%- else -%}\n {{- raise_exception(\"Unknown assistant content part: \" + part['type']) }}\n {%- endif -%}\n {%- endfor -%}\n {%- else -%}\n {{- raise_exception(\"Unsupported assistant content type\") }}\n {%- endif -%}\n {{- '<|assistant_end|>' }}\n {%- endif -%}\n{%- endfor -%}\n{%- if add_generation_prompt -%}\n {{- '<|assistant_start|>' }}\n{%- endif -%}\n"
97
+ }