TianshengHuang commited on
Commit
64cebd0
·
verified ·
1 Parent(s): 72f1efd

Model save

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: TianshengHuang/s1k
3
  library_name: transformers
4
  model_name: _sft_sft_5
5
  tags:
@@ -11,7 +11,7 @@ licence: license
11
 
12
  # Model Card for _sft_sft_5
13
 
14
- This model is a fine-tuned version of [TianshengHuang/s1k](https://huggingface.co/TianshengHuang/s1k).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
1
  ---
2
+ base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
3
  library_name: transformers
4
  model_name: _sft_sft_5
5
  tags:
 
11
 
12
  # Model Card for _sft_sft_5
13
 
14
+ This model is a fine-tuned version of [deepseek-ai/DeepSeek-R1-Distill-Qwen-32B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
config.json CHANGED
@@ -1,22 +1,22 @@
1
  {
2
- "_name_or_path": "TianshengHuang/s1k",
3
  "architectures": [
4
  "Qwen2ForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 151643,
8
- "eos_token_id": 151645,
9
  "hidden_act": "silu",
10
  "hidden_size": 5120,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 27648,
13
- "max_position_embeddings": 32768,
14
- "max_window_layers": 70,
15
  "model_type": "qwen2",
16
  "num_attention_heads": 40,
17
  "num_hidden_layers": 64,
18
  "num_key_value_heads": 8,
19
- "rms_norm_eps": 1e-06,
20
  "rope_scaling": null,
21
  "rope_theta": 1000000.0,
22
  "sliding_window": null,
 
1
  {
2
+ "_name_or_path": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B",
3
  "architectures": [
4
  "Qwen2ForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
  "hidden_act": "silu",
10
  "hidden_size": 5120,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 27648,
13
+ "max_position_embeddings": 131072,
14
+ "max_window_layers": 64,
15
  "model_type": "qwen2",
16
  "num_attention_heads": 40,
17
  "num_hidden_layers": 64,
18
  "num_key_value_heads": 8,
19
+ "rms_norm_eps": 1e-05,
20
  "rope_scaling": null,
21
  "rope_theta": 1000000.0,
22
  "sliding_window": null,
generation_config.json CHANGED
@@ -1,14 +1,9 @@
1
  {
2
- "bos_token_id": 151643,
 
3
  "do_sample": true,
4
- "eos_token_id": [
5
- 151645,
6
- 151643
7
- ],
8
- "pad_token_id": 151643,
9
- "repetition_penalty": 1.05,
10
- "temperature": 0.7,
11
- "top_k": 20,
12
- "top_p": 0.8,
13
  "transformers_version": "4.46.1"
14
  }
 
1
  {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151646,
4
  "do_sample": true,
5
+ "eos_token_id": 151643,
6
+ "temperature": 0.6,
7
+ "top_p": 0.95,
 
 
 
 
 
 
8
  "transformers_version": "4.46.1"
9
  }
model-00001-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d797f70ba7f1b3fb4a2c971d77896c59369c7d634c09701ff321374ecba4eaa
3
  size 4891730992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32b302cbcf810efef02f1f796b7b429b647bb46dbfe27bdd0794e323c25d06d5
3
  size 4891730992
model-00002-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:261b71b88deead7ac53233764a6055ab10165cb6b5aed7391cce1f52ea128834
3
  size 4876059352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f31635e59a73214edaa09c3ae3f16d9a042ef61dc366dd47a80a1286b93df31
3
  size 4876059352
model-00003-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f852efdb550af10f4e743411354a7e7d8f7af855bcc496c0a75d249a0b8b052
3
  size 4876059384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51941ae259a3458153c01d69ca44c7ecb1460a6dda0c8585807256101695a35a
3
  size 4876059384
model-00004-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f12d7185a37cff385b12e232168ca6111188d69e8e6cd297e7de4d0bda6a16ba
3
  size 4876059416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b8945ba7bf120acfe84403c65126947a6fd3701f5f1d4df17b25708d8050e9f
3
  size 4876059416
model-00005-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a90c43faf7910d1659af1fceb42cbbe6bc4f8ae68b68543b427e01177a9311e2
3
  size 4876059416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c4fcf1f208159793a48fa86a00081b1e48ce5daa4c346bb5012f8482a5b067e
3
  size 4876059416
model-00006-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cedbe456c2053e1f7fb2c7be6f1bde928131cacd68120bca2b07e3beb8a75718
3
  size 4876059416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05858e15319a18ac9fbf89244c860817a71de9dc81841a9388ea8ec53693fd05
3
  size 4876059416
model-00007-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:faa920f188ec3a93741ffcb5f3d63970a5031d78e9c4ef7bf4b1f3a521370a72
3
  size 4876059416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9166ef18fb3d2ca8b8a89f2e53965a1d47fe26a586e8b53427b9677351f9c6d5
3
  size 4876059416
model-00008-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:814d660f09932fbcdee9f13b97836e5d471ad3e6d61ad85e4439af43edb9b831
3
  size 4876059416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f822daacd802e02d78ecdded1777e3ee30142ee12ddff14f1b873ffd7480f651
3
  size 4876059416
model-00009-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1b0af36cfcea91335b526ffb69f37918d7d0f1d92bb0ac4b31c0067dd93a8ec
3
  size 4876059416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d8fc34ad71b574d5bd25815dc2441feb93d7e2c9bba1bb8b0f50960ef35da0a
3
  size 4876059416
model-00010-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2cdb588e227b7a4cba9853e9338c0040471fcd1c59ab50b4c10161e3a43bc2a
3
  size 4876059416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cf064c509eba81e2d252637d7c787d0239244976c4f4e6c1d856b9407abd585
3
  size 4876059416
model-00011-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:495a2e6daaca88b15caae5b3113ebf2ee873dd605c345212b0a050ae4f6ed5df
3
  size 4876059416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fff6dfb64bbe5e722a5d19ae3242d55e367b973dc0239318adce6d5162b4435
3
  size 4876059416
model-00012-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4faa71b9f19c4d695e19e005d8965f0602fbf9ef51d85fb51b1ae62f4d595a7
3
  size 4876059416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b4cf1d664a5cad0d453eeb5d9d909856b465c63d843f9d1d5884b795faca590
3
  size 4876059416
model-00013-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:119d87ff5a65fe3c3ee4b49cc58ebca44b9e70bde8ec4d3db5c4b5bd2c11f7cc
3
  size 4876059416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08874d84adb319f3acff090b4258571e770687d460212d7d8a884bbe7b1e7437
3
  size 4876059416
model-00014-of-00014.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bfa6be68455fa0c1569a8e0519dac90b7f5dcfe274b6c6f67ea928f54e32a1bd
3
  size 2123397800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22d385ce64bdfbc0406a62120b7455bb08fa45e768b05c8c7dfbfb37aae421bb
3
  size 2123397800
special_tokens_map.json CHANGED
@@ -1,28 +1,20 @@
1
  {
2
- "additional_special_tokens": [
3
- "<|im_start|>",
4
- "<|im_end|>",
5
- "<|object_ref_start|>",
6
- "<|object_ref_end|>",
7
- "<|box_start|>",
8
- "<|box_end|>",
9
- "<|quad_start|>",
10
- "<|quad_end|>",
11
- "<|vision_start|>",
12
- "<|vision_end|>",
13
- "<|vision_pad|>",
14
- "<|image_pad|>",
15
- "<|video_pad|>"
16
- ],
17
  "eos_token": {
18
- "content": "<|im_end|>",
19
  "lstrip": false,
20
  "normalized": false,
21
  "rstrip": false,
22
  "single_word": false
23
  },
24
  "pad_token": {
25
- "content": "<|endoftext|>",
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
 
1
  {
2
+ "bos_token": {
3
+ "content": "<|begin▁of▁sentence|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
 
 
 
 
 
 
 
 
9
  "eos_token": {
10
+ "content": "<|end▁of▁sentence|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "pad_token": {
17
+ "content": "<|end▁of▁sentence|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68301b5308a972f0c3e44b7251787e603564d598575dab87d8068b0ef893996d
3
- size 11421996
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8749954a732f5b46cd497790be3031e5199a8f6e7f72862d5e1d26513bac9564
3
+ size 11422878
tokenizer_config.json CHANGED
@@ -1,9 +1,10 @@
1
  {
2
- "add_bos_token": false,
3
- "add_prefix_space": false,
 
4
  "added_tokens_decoder": {
5
  "151643": {
6
- "content": "<|endoftext|>",
7
  "lstrip": false,
8
  "normalized": false,
9
  "rstrip": false,
@@ -11,23 +12,23 @@
11
  "special": true
12
  },
13
  "151644": {
14
- "content": "<|im_start|>",
15
  "lstrip": false,
16
  "normalized": false,
17
  "rstrip": false,
18
  "single_word": false,
19
- "special": true
20
  },
21
  "151645": {
22
- "content": "<|im_end|>",
23
  "lstrip": false,
24
  "normalized": false,
25
  "rstrip": false,
26
  "single_word": false,
27
- "special": true
28
  },
29
  "151646": {
30
- "content": "<|object_ref_start|>",
31
  "lstrip": false,
32
  "normalized": false,
33
  "rstrip": false,
@@ -35,28 +36,28 @@
35
  "special": true
36
  },
37
  "151647": {
38
- "content": "<|object_ref_end|>",
39
  "lstrip": false,
40
  "normalized": false,
41
  "rstrip": false,
42
  "single_word": false,
43
- "special": true
44
  },
45
  "151648": {
46
- "content": "<|box_start|>",
47
  "lstrip": false,
48
  "normalized": false,
49
  "rstrip": false,
50
  "single_word": false,
51
- "special": true
52
  },
53
  "151649": {
54
- "content": "<|box_end|>",
55
  "lstrip": false,
56
  "normalized": false,
57
  "rstrip": false,
58
  "single_word": false,
59
- "special": true
60
  },
61
  "151650": {
62
  "content": "<|quad_start|>",
@@ -179,34 +180,15 @@
179
  "special": false
180
  }
181
  },
182
- "additional_special_tokens": [
183
- "<|im_start|>",
184
- "<|im_end|>",
185
- "<|object_ref_start|>",
186
- "<|object_ref_end|>",
187
- "<|box_start|>",
188
- "<|box_end|>",
189
- "<|quad_start|>",
190
- "<|quad_end|>",
191
- "<|vision_start|>",
192
- "<|vision_end|>",
193
- "<|vision_pad|>",
194
- "<|image_pad|>",
195
- "<|video_pad|>"
196
- ],
197
- "bos_token": null,
198
- "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
  "clean_up_tokenization_spaces": false,
200
- "eos_token": "<|im_end|>",
201
- "errors": "replace",
202
- "extra_special_tokens": {},
203
- "max_length": 20000,
204
- "model_max_length": 131072,
205
- "pad_token": "<|endoftext|>",
206
- "split_special_tokens": false,
207
- "stride": 0,
208
- "tokenizer_class": "Qwen2Tokenizer",
209
- "truncation_side": "right",
210
- "truncation_strategy": "longest_first",
211
- "unk_token": null
212
  }
 
1
  {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "151643": {
7
+ "content": "<|end▁of▁sentence|>",
8
  "lstrip": false,
9
  "normalized": false,
10
  "rstrip": false,
 
12
  "special": true
13
  },
14
  "151644": {
15
+ "content": "<|User|>",
16
  "lstrip": false,
17
  "normalized": false,
18
  "rstrip": false,
19
  "single_word": false,
20
+ "special": false
21
  },
22
  "151645": {
23
+ "content": "<|Assistant|>",
24
  "lstrip": false,
25
  "normalized": false,
26
  "rstrip": false,
27
  "single_word": false,
28
+ "special": false
29
  },
30
  "151646": {
31
+ "content": "<|begin▁of▁sentence|>",
32
  "lstrip": false,
33
  "normalized": false,
34
  "rstrip": false,
 
36
  "special": true
37
  },
38
  "151647": {
39
+ "content": "<|EOT|>",
40
  "lstrip": false,
41
  "normalized": false,
42
  "rstrip": false,
43
  "single_word": false,
44
+ "special": false
45
  },
46
  "151648": {
47
+ "content": "<think>",
48
  "lstrip": false,
49
  "normalized": false,
50
  "rstrip": false,
51
  "single_word": false,
52
+ "special": false
53
  },
54
  "151649": {
55
+ "content": "</think>",
56
  "lstrip": false,
57
  "normalized": false,
58
  "rstrip": false,
59
  "single_word": false,
60
+ "special": false
61
  },
62
  "151650": {
63
  "content": "<|quad_start|>",
 
180
  "special": false
181
  }
182
  },
183
+ "bos_token": "<|begin▁of▁sentence|>",
184
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set ns = namespace(is_first=false, is_tool=false, is_output_first=true, system_prompt='') %}{%- for message in messages %}{%- if message['role'] == 'system' %}{% set ns.system_prompt = message['content'] %}{%- endif %}{%- endfor %}{{bos_token}}{{ns.system_prompt}}{%- for message in messages %}{%- if message['role'] == 'user' %}{%- set ns.is_tool = false -%}{{'<|User|>' + message['content']}}{%- endif %}{%- if message['role'] == 'assistant' and message['content'] is none %}{%- set ns.is_tool = false -%}{%- for tool in message['tool_calls']%}{%- if not ns.is_first %}{{'<|Assistant|><|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\\n' + '```json' + '\\n' + tool['function']['arguments'] + '\\n' + '```' + '<|tool▁call▁end|>'}}{%- set ns.is_first = true -%}{%- else %}{{'\\n' + '<|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\\n' + '```json' + '\\n' + tool['function']['arguments'] + '\\n' + '```' + '<|tool▁call▁end|>'}}{{'<|tool▁calls▁end|><|end▁of▁sentence|>'}}{%- endif %}{%- endfor %}{%- endif %}{%- if message['role'] == 'assistant' and message['content'] is not none %}{%- if ns.is_tool %}{{'<|tool▁outputs▁end|>' + message['content'] + '<|end▁of▁sentence|>'}}{%- set ns.is_tool = false -%}{%- else %}{% set content = message['content'] %}{% if '</think>' in content %}{% set content = content.split('</think>')[-1] %}{% endif %}{{'<|Assistant|>' + content + '<|end▁of▁sentence|>'}}{%- endif %}{%- endif %}{%- if message['role'] == 'tool' %}{%- set ns.is_tool = true -%}{%- if ns.is_output_first %}{{'<|tool▁outputs▁begin|><|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- set ns.is_output_first = false %}{%- else %}{{'\\n<|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- endif %}{%- endif %}{%- endfor -%}{% if ns.is_tool %}{{'<|tool▁outputs▁end|>'}}{% endif %}{% if add_generation_prompt and not ns.is_tool %}{{'<|Assistant|><think>\\n'}}{% endif %}",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
185
  "clean_up_tokenization_spaces": false,
186
+ "eos_token": "<|end▁of▁sentence|>",
187
+ "legacy": true,
188
+ "model_max_length": 16384,
189
+ "pad_token": "<|end▁of▁sentence|>",
190
+ "sp_model_kwargs": {},
191
+ "tokenizer_class": "LlamaTokenizer",
192
+ "unk_token": null,
193
+ "use_default_system_prompt": false
 
 
 
 
194
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a9836d5824aaeeca875e0b2e9d5fdb7e9ac54ca2961f3c5cbe0f2cdbc25bb45
3
- size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93df8358164e9e3490f9367d2a961f98b3c4b9c4f5470d9c7e3baf8d22f17975
3
+ size 5816