File size: 2,695 Bytes
812540e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
{
  "architectures": [
    "Qwen3ForCausalLM"
  ],
  "auto_map": {
    "AutoModelForCausalLM": "Qwen3ForCausalLM"
  },
  "model_type": "qwen3",
  "quantization_config": {
    "quant_method": "gguf",
    "bits": 8,
    "group_size": 32,
    "zero_point": true,
    "desc_act": false
  },
  "gguf_metadata": {
    "general": {
      "name": "Qwen3 4B Function Calling Pro",
      "architecture": "qwen3",
      "file_type": 7,
      "quantization_version": 2,
      "size_label": "4B"
    },
    "qwen3": {
      "context_length": 262144,
      "embedding_length": 2560,
      "block_count": 36,
      "feed_forward_length": 9728,
      "attention": {
        "head_count": 32,
        "head_count_kv": 8,
        "key_length": 128,
        "value_length": 128,
        "layer_norm_rms_epsilon": 1e-06
      },
      "rope": {
        "freq_base": 5000000.0
      }
    }
  },
  "vocab_size": 151936,
  "hidden_size": 2560,
  "intermediate_size": 9728,
  "num_hidden_layers": 36,
  "num_attention_heads": 32,
  "num_key_value_heads": 8,
  "hidden_act": "silu",
  "max_position_embeddings": 262144,
  "initializer_range": 0.02,
  "rms_norm_eps": 1e-06,
  "use_cache": true,
  "rope_theta": 5000000.0,
  "attention_bias": false,
  "attention_dropout": 0.0,
  "torch_dtype": "float16",
  "transformers_version": "4.45.0",
  "special_tokens": {
    "bos_token": "<|endoftext|>",
    "eos_token": "<|im_end|>",
    "pad_token": "<|im_end|>",
    "additional_special_tokens": [
      "<|im_start|>",
      "<|im_end|>",
      "<|object_ref_start|>",
      "<|object_ref_end|>",
      "<|box_start|>",
      "<|box_end|>",
      "<|quad_start|>",
      "<|quad_end|>",
      "<|vision_start|>",
      "<|vision_end|>",
      "<|vision_pad|>",
      "<|image_pad|>",
      "<|video_pad|>",
      "<tool_call>",
      "</tool_call>",
      "<tool_response>",
      "</tool_response>"
    ]
  },
  "chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{ '<start_of_turn>user\n' + message['content'] + '<end_of_turn>\n' }}{% elif message['role'] == 'system' %}{{ '<start_of_turn>system\n' + message['content'] + '<end_of_turn>\n' }}{% elif message['role'] == 'assistant' %}{{ '<start_of_turn>model\n' + message['content'] + '<end_of_turn>\n' }}{% endif %}{% if loop.last and add_generation_prompt %}{{ '<start_of_turn>model\n' }}{% endif %}{% endfor %}",
  "generation_config": {
    "do_sample": true,
    "temperature": 0.7,
    "top_p": 0.8,
    "top_k": 40,
    "repetition_penalty": 1.1,
    "max_new_tokens": 512,
    "stop": [
      "<|im_end|>",
      "<|im_start|>",
      "<tool_call>",
      "</tool_call>",
      "<tool_response>",
      "</tool_response>"
    ]
  }
}