Training in progress, epoch 1
Browse files- chat_template.jinja +54 -0
- config.json +31 -2
- model-00001-of-00004.safetensors +1 -1
- model-00002-of-00004.safetensors +1 -1
- model-00003-of-00004.safetensors +1 -1
- model-00004-of-00004.safetensors +1 -1
- model.safetensors.index.json +1 -0
- special_tokens_map.json +1 -1
- tokenizer_config.json +3 -3
- trainer_log.jsonl +0 -0
- training_args.bin +2 -2
    	
        chat_template.jinja
    ADDED
    
    | @@ -0,0 +1,54 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {%- if tools %}
         | 
| 2 | 
            +
                {{- '<|im_start|>system\n' }}
         | 
| 3 | 
            +
                {%- if messages[0]['role'] == 'system' %}
         | 
| 4 | 
            +
                    {{- messages[0]['content'] }}
         | 
| 5 | 
            +
                {%- else %}
         | 
| 6 | 
            +
                    {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}
         | 
| 7 | 
            +
                {%- endif %}
         | 
| 8 | 
            +
                {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
         | 
| 9 | 
            +
                {%- for tool in tools %}
         | 
| 10 | 
            +
                    {{- "\n" }}
         | 
| 11 | 
            +
                    {{- tool | tojson }}
         | 
| 12 | 
            +
                {%- endfor %}
         | 
| 13 | 
            +
                {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
         | 
| 14 | 
            +
            {%- else %}
         | 
| 15 | 
            +
                {%- if messages[0]['role'] == 'system' %}
         | 
| 16 | 
            +
                    {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
         | 
| 17 | 
            +
                {%- else %}
         | 
| 18 | 
            +
                    {{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }}
         | 
| 19 | 
            +
                {%- endif %}
         | 
| 20 | 
            +
            {%- endif %}
         | 
| 21 | 
            +
            {%- for message in messages %}
         | 
| 22 | 
            +
                {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %}
         | 
| 23 | 
            +
                    {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
         | 
| 24 | 
            +
                {%- elif message.role == "assistant" %}
         | 
| 25 | 
            +
                    {{- '<|im_start|>' + message.role }}
         | 
| 26 | 
            +
                    {%- if message.content %}
         | 
| 27 | 
            +
                        {{- '\n' + message.content }}
         | 
| 28 | 
            +
                    {%- endif %}
         | 
| 29 | 
            +
                    {%- for tool_call in message.tool_calls %}
         | 
| 30 | 
            +
                        {%- if tool_call.function is defined %}
         | 
| 31 | 
            +
                            {%- set tool_call = tool_call.function %}
         | 
| 32 | 
            +
                        {%- endif %}
         | 
| 33 | 
            +
                        {{- '\n<tool_call>\n{"name": "' }}
         | 
| 34 | 
            +
                        {{- tool_call.name }}
         | 
| 35 | 
            +
                        {{- '", "arguments": ' }}
         | 
| 36 | 
            +
                        {{- tool_call.arguments | tojson }}
         | 
| 37 | 
            +
                        {{- '}\n</tool_call>' }}
         | 
| 38 | 
            +
                    {%- endfor %}
         | 
| 39 | 
            +
                    {{- '<|im_end|>\n' }}
         | 
| 40 | 
            +
                {%- elif message.role == "tool" %}
         | 
| 41 | 
            +
                    {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
         | 
| 42 | 
            +
                        {{- '<|im_start|>user' }}
         | 
| 43 | 
            +
                    {%- endif %}
         | 
| 44 | 
            +
                    {{- '\n<tool_response>\n' }}
         | 
| 45 | 
            +
                    {{- message.content }}
         | 
| 46 | 
            +
                    {{- '\n</tool_response>' }}
         | 
| 47 | 
            +
                    {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
         | 
| 48 | 
            +
                        {{- '<|im_end|>\n' }}
         | 
| 49 | 
            +
                    {%- endif %}
         | 
| 50 | 
            +
                {%- endif %}
         | 
| 51 | 
            +
            {%- endfor %}
         | 
| 52 | 
            +
            {%- if add_generation_prompt %}
         | 
| 53 | 
            +
                {{- '<|im_start|>assistant\n' }}
         | 
| 54 | 
            +
            {%- endif %}
         | 
    	
        config.json
    CHANGED
    
    | @@ -1,5 +1,4 @@ | |
| 1 | 
             
            {
         | 
| 2 | 
            -
              "_name_or_path": "Qwen/Qwen2.5-7B-Instruct",
         | 
| 3 | 
             
              "architectures": [
         | 
| 4 | 
             
                "Qwen2ForCausalLM"
         | 
| 5 | 
             
              ],
         | 
| @@ -10,6 +9,36 @@ | |
| 10 | 
             
              "hidden_size": 3584,
         | 
| 11 | 
             
              "initializer_range": 0.02,
         | 
| 12 | 
             
              "intermediate_size": 18944,
         | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 13 | 
             
              "max_position_embeddings": 32768,
         | 
| 14 | 
             
              "max_window_layers": 28,
         | 
| 15 | 
             
              "model_type": "qwen2",
         | 
| @@ -22,7 +51,7 @@ | |
| 22 | 
             
              "sliding_window": null,
         | 
| 23 | 
             
              "tie_word_embeddings": false,
         | 
| 24 | 
             
              "torch_dtype": "bfloat16",
         | 
| 25 | 
            -
              "transformers_version": "4. | 
| 26 | 
             
              "use_cache": false,
         | 
| 27 | 
             
              "use_sliding_window": false,
         | 
| 28 | 
             
              "vocab_size": 152064
         | 
|  | |
| 1 | 
             
            {
         | 
|  | |
| 2 | 
             
              "architectures": [
         | 
| 3 | 
             
                "Qwen2ForCausalLM"
         | 
| 4 | 
             
              ],
         | 
|  | |
| 9 | 
             
              "hidden_size": 3584,
         | 
| 10 | 
             
              "initializer_range": 0.02,
         | 
| 11 | 
             
              "intermediate_size": 18944,
         | 
| 12 | 
            +
              "layer_types": [
         | 
| 13 | 
            +
                "full_attention",
         | 
| 14 | 
            +
                "full_attention",
         | 
| 15 | 
            +
                "full_attention",
         | 
| 16 | 
            +
                "full_attention",
         | 
| 17 | 
            +
                "full_attention",
         | 
| 18 | 
            +
                "full_attention",
         | 
| 19 | 
            +
                "full_attention",
         | 
| 20 | 
            +
                "full_attention",
         | 
| 21 | 
            +
                "full_attention",
         | 
| 22 | 
            +
                "full_attention",
         | 
| 23 | 
            +
                "full_attention",
         | 
| 24 | 
            +
                "full_attention",
         | 
| 25 | 
            +
                "full_attention",
         | 
| 26 | 
            +
                "full_attention",
         | 
| 27 | 
            +
                "full_attention",
         | 
| 28 | 
            +
                "full_attention",
         | 
| 29 | 
            +
                "full_attention",
         | 
| 30 | 
            +
                "full_attention",
         | 
| 31 | 
            +
                "full_attention",
         | 
| 32 | 
            +
                "full_attention",
         | 
| 33 | 
            +
                "full_attention",
         | 
| 34 | 
            +
                "full_attention",
         | 
| 35 | 
            +
                "full_attention",
         | 
| 36 | 
            +
                "full_attention",
         | 
| 37 | 
            +
                "full_attention",
         | 
| 38 | 
            +
                "full_attention",
         | 
| 39 | 
            +
                "full_attention",
         | 
| 40 | 
            +
                "full_attention"
         | 
| 41 | 
            +
              ],
         | 
| 42 | 
             
              "max_position_embeddings": 32768,
         | 
| 43 | 
             
              "max_window_layers": 28,
         | 
| 44 | 
             
              "model_type": "qwen2",
         | 
|  | |
| 51 | 
             
              "sliding_window": null,
         | 
| 52 | 
             
              "tie_word_embeddings": false,
         | 
| 53 | 
             
              "torch_dtype": "bfloat16",
         | 
| 54 | 
            +
              "transformers_version": "4.55.0",
         | 
| 55 | 
             
              "use_cache": false,
         | 
| 56 | 
             
              "use_sliding_window": false,
         | 
| 57 | 
             
              "vocab_size": 152064
         | 
    	
        model-00001-of-00004.safetensors
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
             
            size 4877660776
         | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:3acb83f66cd2e73f93085e12af05da0ce1837fecca46a3959c91dac5b0d81888
         | 
| 3 | 
             
            size 4877660776
         | 
    	
        model-00002-of-00004.safetensors
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
             
            size 4932751008
         | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:b81154270761f1a26fadaf1abd4a10a225f1ad77ee4e108664d4403cea2f13c4
         | 
| 3 | 
             
            size 4932751008
         | 
    	
        model-00003-of-00004.safetensors
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
             
            size 4330865200
         | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:df8b7a39a9ed56197f28fa52bc50d328428665a48b5e37190cbe0519b8d1b2ea
         | 
| 3 | 
             
            size 4330865200
         | 
    	
        model-00004-of-00004.safetensors
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
             
            size 1089994880
         | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:5ef03a5c463a20df64e168be85a724e55929cf9d1ab6d9882fcc0a8c84d21557
         | 
| 3 | 
             
            size 1089994880
         | 
    	
        model.safetensors.index.json
    CHANGED
    
    | @@ -1,5 +1,6 @@ | |
| 1 | 
             
            {
         | 
| 2 | 
             
              "metadata": {
         | 
|  | |
| 3 | 
             
                "total_size": 15231233024
         | 
| 4 | 
             
              },
         | 
| 5 | 
             
              "weight_map": {
         | 
|  | |
| 1 | 
             
            {
         | 
| 2 | 
             
              "metadata": {
         | 
| 3 | 
            +
                "total_parameters": 333312,
         | 
| 4 | 
             
                "total_size": 15231233024
         | 
| 5 | 
             
              },
         | 
| 6 | 
             
              "weight_map": {
         | 
    	
        special_tokens_map.json
    CHANGED
    
    | @@ -15,7 +15,7 @@ | |
| 15 | 
             
                "<|video_pad|>"
         | 
| 16 | 
             
              ],
         | 
| 17 | 
             
              "eos_token": {
         | 
| 18 | 
            -
                "content": "<| | 
| 19 | 
             
                "lstrip": false,
         | 
| 20 | 
             
                "normalized": false,
         | 
| 21 | 
             
                "rstrip": false,
         | 
|  | |
| 15 | 
             
                "<|video_pad|>"
         | 
| 16 | 
             
              ],
         | 
| 17 | 
             
              "eos_token": {
         | 
| 18 | 
            +
                "content": "<|im_end|>",
         | 
| 19 | 
             
                "lstrip": false,
         | 
| 20 | 
             
                "normalized": false,
         | 
| 21 | 
             
                "rstrip": false,
         | 
    	
        tokenizer_config.json
    CHANGED
    
    | @@ -195,11 +195,11 @@ | |
| 195 | 
             
                "<|video_pad|>"
         | 
| 196 | 
             
              ],
         | 
| 197 | 
             
              "bos_token": null,
         | 
| 198 | 
            -
              "chat_template": "{%- if tools %}\n    {{- '<|im_start|>system\\n' }}\n    {%- if messages[0]['role'] == 'system' %}\n        {{- messages[0]['content'] }}\n    {%- else %}\n        {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n    {%- endif %}\n    {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n    {%- for tool in tools %}\n        {{- \"\\n\" }}\n        {{- tool | tojson }}\n    {%- endfor %}\n    {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n    {%- if messages[0]['role'] == 'system' %}\n        {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n    {%- else %}\n        {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n    {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n    {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n        {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n    {%- elif message.role == \"assistant\" %}\n        {{- '<|im_start|>' + message.role }}\n        {%- if message.content %}\n            {{- '\\n' + message.content }}\n        {%- endif %}\n        {%- for tool_call in message.tool_calls %}\n            {%- if tool_call.function is defined %}\n                {%- set tool_call = tool_call.function %}\n            {%- endif %}\n            {{- '\\n<tool_call>\\n{\"name\": \"' }}\n            {{- tool_call.name }}\n            {{- '\", \"arguments\": ' }}\n            {{- tool_call.arguments | tojson }}\n            {{- '}\\n</tool_call>' }}\n        {%- endfor %}\n        {{- '<|im_end|>\\n' }}\n    {%- elif message.role == \"tool\" %}\n        {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n            {{- '<|im_start|>user' }}\n        {%- endif %}\n        {{- '\\n<tool_response>\\n' }}\n        {{- message.content }}\n        {{- '\\n</tool_response>' }}\n        {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n            {{- '<|im_end|>\\n' }}\n        {%- endif %}\n    {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n    {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
         | 
| 199 | 
             
              "clean_up_tokenization_spaces": false,
         | 
| 200 | 
            -
              "eos_token": "<| | 
| 201 | 
             
              "errors": "replace",
         | 
| 202 | 
            -
              " | 
|  | |
| 203 | 
             
              "pad_token": "<|endoftext|>",
         | 
| 204 | 
             
              "padding_side": "right",
         | 
| 205 | 
             
              "split_special_tokens": false,
         | 
|  | |
| 195 | 
             
                "<|video_pad|>"
         | 
| 196 | 
             
              ],
         | 
| 197 | 
             
              "bos_token": null,
         | 
|  | |
| 198 | 
             
              "clean_up_tokenization_spaces": false,
         | 
| 199 | 
            +
              "eos_token": "<|im_end|>",
         | 
| 200 | 
             
              "errors": "replace",
         | 
| 201 | 
            +
              "extra_special_tokens": {},
         | 
| 202 | 
            +
              "model_max_length": 16384,
         | 
| 203 | 
             
              "pad_token": "<|endoftext|>",
         | 
| 204 | 
             
              "padding_side": "right",
         | 
| 205 | 
             
              "split_special_tokens": false,
         | 
    	
        trainer_log.jsonl
    CHANGED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        training_args.bin
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
            -
            size  | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:31d00beb6c7822b727035828821276d6a1e137aa38169392711d52a5c616ae8f
         | 
| 3 | 
            +
            size 8465
         | 
