| default_stage: | |
| default_modifiers: | |
| AWQModifier: | |
| config_groups: | |
| group_0: | |
| targets: [Linear] | |
| weights: | |
| num_bits: 4 | |
| type: int | |
| symmetric: true | |
| group_size: 32 | |
| strategy: group | |
| block_structure: null | |
| dynamic: false | |
| actorder: null | |
| observer: mse | |
| observer_kwargs: {} | |
| input_activations: null | |
| output_activations: null | |
| format: null | |
| targets: [Linear] | |
| ignore: [model.embed_tokens, 're:.*input_layernorm$', 're:.*linear_attn.*', 're:.*norm.*', | |
| 're:.*RMSNorm.*', 're:.*rotary.*', 're:.*shared_expert.*', 're:.*shared_expert_gate$', | |
| 're:.*mlp[.]gate$', 're:.*router.*', 're:.*post_attention_layernorm$', 're:.*self_attn.*', | |
| 're:mtp.*', lm_head] | |
| mappings: | |
| - smooth_layer: re:.*input_layernorm$ | |
| balance_layers: ['re:.*self_attn[.]q_proj$', 're:.*self_attn[.]k_proj$', 're:.*self_attn[.]v_proj$', | |
| 're:.*linear_attn[.]in_proj_qkvz$', 're:.*linear_attn[.]in_proj_ba$'] | |
| - smooth_layer: re:.*self_attn[.]v_proj$ | |
| balance_layers: ['re:.*self_attn[.]o_proj$'] | |
| - smooth_layer: re:.*post_attention_layernorm$ | |
| balance_layers: ['re:.*gate_proj$', 're:.*up_proj$'] | |
| - smooth_layer: re:.*up_proj$ | |
| balance_layers: ['re:.*down_proj$'] | |
| - smooth_layer: re:.*linear_attn[.]norm$ | |
| balance_layers: ['re:.*linear_attn[.]out_proj$'] | |
| duo_scaling: true | |