Upload folder using huggingface_hub
Browse files- .gitattributes +1 -0
- chat_template.json +3 -0
- config.json +37 -0
- generation_config.json +8 -0
- model-00001-of-00012.safetensors +3 -0
- model-00002-of-00012.safetensors +3 -0
- model-00003-of-00012.safetensors +3 -0
- model-00004-of-00012.safetensors +3 -0
- model-00005-of-00012.safetensors +3 -0
- model-00006-of-00012.safetensors +3 -0
- model-00007-of-00012.safetensors +3 -0
- model-00008-of-00012.safetensors +3 -0
- model-00009-of-00012.safetensors +3 -0
- model-00010-of-00012.safetensors +3 -0
- model-00011-of-00012.safetensors +3 -0
- model-00012-of-00012.safetensors +3 -0
- model.safetensors.index.json +0 -0
- preprocessor_config.json +29 -0
- processor_config.json +4 -0
- special_tokens_map.json +33 -0
- tokenizer.json +3 -0
- tokenizer_config.json +0 -0
    	
        .gitattributes
    CHANGED
    
    | @@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text | |
| 33 | 
             
            *.zip filter=lfs diff=lfs merge=lfs -text
         | 
| 34 | 
             
            *.zst filter=lfs diff=lfs merge=lfs -text
         | 
| 35 | 
             
            *tfevents* filter=lfs diff=lfs merge=lfs -text
         | 
|  | 
|  | |
| 33 | 
             
            *.zip filter=lfs diff=lfs merge=lfs -text
         | 
| 34 | 
             
            *.zst filter=lfs diff=lfs merge=lfs -text
         | 
| 35 | 
             
            *tfevents* filter=lfs diff=lfs merge=lfs -text
         | 
| 36 | 
            +
            tokenizer.json filter=lfs diff=lfs merge=lfs -text
         | 
    	
        chat_template.json
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n    {%- if messages[0]['content'] is string -%}\n        {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n    {%- else -%}\n        {%- set first_user_prefix = messages[0]['content'][0]['text'] + '\n\n' -%}\n    {%- endif -%}\n    {%- set loop_messages = messages[1:] -%}\n{%- else -%}\n    {%- set first_user_prefix = \"\" -%}\n    {%- set loop_messages = messages -%}\n{%- endif -%}\n{%- for message in loop_messages -%}\n    {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}\n        {{ raise_exception(\"Conversation roles must alternate user/assistant/user/assistant/...\") }}\n    {%- endif -%}\n    {%- if (message['role'] == 'assistant') -%}\n        {%- set role = \"model\" -%}\n    {%- else -%}\n        {%- set role = message['role'] -%}\n    {%- endif -%}\n    {{ '<start_of_turn>' + role + '\n' + (first_user_prefix if loop.first else \"\") }}\n    {%- if message['content'] is string -%}\n        {{ message['content'] | trim }}\n    {%- elif message['content'] is iterable -%}\n        {%- for item in message['content'] -%}\n            {%- if item['type'] == 'image' -%}\n                {{ '<start_of_image>' }}\n            {%- elif item['type'] == 'text' -%}\n                {{ item['text'] | trim }}\n            {%- endif -%}\n        {%- endfor -%}\n    {%- else -%}\n        {{ raise_exception(\"Invalid content type\") }}\n    {%- endif -%}\n    {{ '<end_of_turn>\n' }}\n{%- endfor -%}\n{%- if add_generation_prompt -%}\n    {{ '<start_of_turn>model\n' }}\n{%- endif -%}\n"
         | 
| 3 | 
            +
            }
         | 
    	
        config.json
    ADDED
    
    | @@ -0,0 +1,37 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "architectures": [
         | 
| 3 | 
            +
                "Gemma3ForCausalLM"
         | 
| 4 | 
            +
              ],
         | 
| 5 | 
            +
              "attention_bias": false,
         | 
| 6 | 
            +
              "attention_dropout": 0.0,
         | 
| 7 | 
            +
              "attn_logit_softcapping": null,
         | 
| 8 | 
            +
              "bos_token_id": 2,
         | 
| 9 | 
            +
              "cache_implementation": "hybrid",
         | 
| 10 | 
            +
              "eos_token_id": 1,
         | 
| 11 | 
            +
              "final_logit_softcapping": null,
         | 
| 12 | 
            +
              "head_dim": 128,
         | 
| 13 | 
            +
              "hidden_activation": "gelu_pytorch_tanh",
         | 
| 14 | 
            +
              "hidden_size": 5376,
         | 
| 15 | 
            +
              "initializer_range": 0.02,
         | 
| 16 | 
            +
              "intermediate_size": 21504,
         | 
| 17 | 
            +
              "max_position_embeddings": 131072,
         | 
| 18 | 
            +
              "model_type": "gemma3_text",
         | 
| 19 | 
            +
              "num_attention_heads": 32,
         | 
| 20 | 
            +
              "num_hidden_layers": 62,
         | 
| 21 | 
            +
              "num_key_value_heads": 16,
         | 
| 22 | 
            +
              "pad_token_id": 0,
         | 
| 23 | 
            +
              "query_pre_attn_scalar": 168,
         | 
| 24 | 
            +
              "rms_norm_eps": 1e-06,
         | 
| 25 | 
            +
              "rope_local_base_freq": 10000.0,
         | 
| 26 | 
            +
              "rope_scaling": {
         | 
| 27 | 
            +
                "factor": 8.0,
         | 
| 28 | 
            +
                "rope_type": "linear"
         | 
| 29 | 
            +
              },
         | 
| 30 | 
            +
              "rope_theta": 1000000.0,
         | 
| 31 | 
            +
              "sliding_window": 1024,
         | 
| 32 | 
            +
              "sliding_window_pattern": 6,
         | 
| 33 | 
            +
              "torch_dtype": "bfloat16",
         | 
| 34 | 
            +
              "transformers_version": "4.50.0.dev0",
         | 
| 35 | 
            +
              "use_cache": true,
         | 
| 36 | 
            +
              "vocab_size": 262208
         | 
| 37 | 
            +
            }
         | 
    	
        generation_config.json
    ADDED
    
    | @@ -0,0 +1,8 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "_from_model_config": true,
         | 
| 3 | 
            +
              "bos_token_id": 2,
         | 
| 4 | 
            +
              "cache_implementation": "hybrid",
         | 
| 5 | 
            +
              "eos_token_id": 1,
         | 
| 6 | 
            +
              "pad_token_id": 0,
         | 
| 7 | 
            +
              "transformers_version": "4.50.0.dev0"
         | 
| 8 | 
            +
            }
         | 
    	
        model-00001-of-00012.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:03d268429104654462aa95fb53736c463e45c96f99e624ce74b4b9201dc0195e
         | 
| 3 | 
            +
            size 4854573696
         | 
    	
        model-00002-of-00012.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:dc28d18305bf2d66109a60241e13daf302aa0b91ff9cbab4f6e70f17149a6f1f
         | 
| 3 | 
            +
            size 4954792944
         | 
    	
        model-00003-of-00012.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:d4487bb69c930bff3fea0d36fa2d190d43b238f09353a09ef64cd17bcf52d771
         | 
| 3 | 
            +
            size 4954792976
         | 
    	
        model-00004-of-00012.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:13115107a8b161d1d432be6e18ae354bf7c1a9e3825162cfce056097e33c5b5e
         | 
| 3 | 
            +
            size 4954793016
         | 
    	
        model-00005-of-00012.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:0eb3a6e04d1ae71cd84e5cd03dc763bcbb3d1ed3bd6e12894eabf213f5e83386
         | 
| 3 | 
            +
            size 4954793016
         | 
    	
        model-00006-of-00012.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:f797a0223433649c5550f906c1a2ee5426543a7db405c62379323c220c8bba29
         | 
| 3 | 
            +
            size 4954793016
         | 
    	
        model-00007-of-00012.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:615f80d3ee17afb844a269f56f9af69b80669102c046d45d665ebb7f3a9c1eb5
         | 
| 3 | 
            +
            size 4954793016
         | 
    	
        model-00008-of-00012.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:c8a7e4c43f17e6f40b6d36e66a68d2e831d9dd359a03df736b911291ed2aeaf5
         | 
| 3 | 
            +
            size 4954793016
         | 
    	
        model-00009-of-00012.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:4ec8b6aecf457f355ffc605a939427dc740ed20f9aad0163c182f2e2909d25a5
         | 
| 3 | 
            +
            size 4954793016
         | 
    	
        model-00010-of-00012.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:f3c2fb6e9ee9693ee7dcc9a3795d23dfaec33948badeb4558c505dc3e3f14709
         | 
| 3 | 
            +
            size 4954793016
         | 
    	
        model-00011-of-00012.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:4b38e3052f1084917f3b04b3b47851be563b0753bf329ae1adfd5064a974eaf6
         | 
| 3 | 
            +
            size 4954793016
         | 
    	
        model-00012-of-00012.safetensors
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:90a80fd46efef9c54dc6dffb1a8fef5febc1cbaa41318eafd37d0a86179b7df7
         | 
| 3 | 
            +
            size 462476696
         | 
    	
        model.safetensors.index.json
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
    	
        preprocessor_config.json
    ADDED
    
    | @@ -0,0 +1,29 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "do_convert_rgb": null,
         | 
| 3 | 
            +
              "do_normalize": true,
         | 
| 4 | 
            +
              "do_pan_and_scan": null,
         | 
| 5 | 
            +
              "do_rescale": true,
         | 
| 6 | 
            +
              "do_resize": true,
         | 
| 7 | 
            +
              "image_mean": [
         | 
| 8 | 
            +
                0.5,
         | 
| 9 | 
            +
                0.5,
         | 
| 10 | 
            +
                0.5
         | 
| 11 | 
            +
              ],
         | 
| 12 | 
            +
              "image_processor_type": "Gemma3ImageProcessor",
         | 
| 13 | 
            +
              "image_seq_length": 256,
         | 
| 14 | 
            +
              "image_std": [
         | 
| 15 | 
            +
                0.5,
         | 
| 16 | 
            +
                0.5,
         | 
| 17 | 
            +
                0.5
         | 
| 18 | 
            +
              ],
         | 
| 19 | 
            +
              "pan_and_scan_max_num_crops": null,
         | 
| 20 | 
            +
              "pan_and_scan_min_crop_size": null,
         | 
| 21 | 
            +
              "pan_and_scan_min_ratio_to_activate": null,
         | 
| 22 | 
            +
              "processor_class": "Gemma3Processor",
         | 
| 23 | 
            +
              "resample": 2,
         | 
| 24 | 
            +
              "rescale_factor": 0.00392156862745098,
         | 
| 25 | 
            +
              "size": {
         | 
| 26 | 
            +
                "height": 896,
         | 
| 27 | 
            +
                "width": 896
         | 
| 28 | 
            +
              }
         | 
| 29 | 
            +
            }
         | 
    	
        processor_config.json
    ADDED
    
    | @@ -0,0 +1,4 @@ | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "image_seq_length": 256,
         | 
| 3 | 
            +
              "processor_class": "Gemma3Processor"
         | 
| 4 | 
            +
            }
         | 
    	
        special_tokens_map.json
    ADDED
    
    | @@ -0,0 +1,33 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "boi_token": "<start_of_image>",
         | 
| 3 | 
            +
              "bos_token": {
         | 
| 4 | 
            +
                "content": "<bos>",
         | 
| 5 | 
            +
                "lstrip": false,
         | 
| 6 | 
            +
                "normalized": false,
         | 
| 7 | 
            +
                "rstrip": false,
         | 
| 8 | 
            +
                "single_word": false
         | 
| 9 | 
            +
              },
         | 
| 10 | 
            +
              "eoi_token": "<end_of_image>",
         | 
| 11 | 
            +
              "eos_token": {
         | 
| 12 | 
            +
                "content": "<end_of_turn>",
         | 
| 13 | 
            +
                "lstrip": false,
         | 
| 14 | 
            +
                "normalized": false,
         | 
| 15 | 
            +
                "rstrip": false,
         | 
| 16 | 
            +
                "single_word": false
         | 
| 17 | 
            +
              },
         | 
| 18 | 
            +
              "image_token": "<image_soft_token>",
         | 
| 19 | 
            +
              "pad_token": {
         | 
| 20 | 
            +
                "content": "<pad>",
         | 
| 21 | 
            +
                "lstrip": false,
         | 
| 22 | 
            +
                "normalized": false,
         | 
| 23 | 
            +
                "rstrip": false,
         | 
| 24 | 
            +
                "single_word": false
         | 
| 25 | 
            +
              },
         | 
| 26 | 
            +
              "unk_token": {
         | 
| 27 | 
            +
                "content": "<unk>",
         | 
| 28 | 
            +
                "lstrip": false,
         | 
| 29 | 
            +
                "normalized": false,
         | 
| 30 | 
            +
                "rstrip": false,
         | 
| 31 | 
            +
                "single_word": false
         | 
| 32 | 
            +
              }
         | 
| 33 | 
            +
            }
         | 
    	
        tokenizer.json
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:4667f2089529e8e7657cfb6d1c19910ae71ff5f28aa7ab2ff2763330affad795
         | 
| 3 | 
            +
            size 33384568
         | 
    	
        tokenizer_config.json
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
