pawasthy commited on
Commit
54c7e47
·
verified ·
1 Parent(s): fb2c3be

Upload 5 files

Browse files
Files changed (4) hide show
  1. config.json +2 -4
  2. model.safetensors +1 -1
  3. tokenizer.json +1 -1
  4. tokenizer_config.json +2 -2
config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "granite-embedding-reranker-english-r2",
3
  "architectures": [
4
  "ModernBertForSequenceClassification"
5
  ],
@@ -34,7 +33,7 @@
34
  },
35
  "layer_norm_eps": 1e-05,
36
  "local_attention": 128,
37
- "local_rope_theta": 40000.0,
38
  "max_position_embeddings": 8192,
39
  "mlp_bias": false,
40
  "mlp_dropout": 0.0,
@@ -45,7 +44,6 @@
45
  "num_hidden_layers": 22,
46
  "pad_token_id": 50283,
47
  "position_embedding_type": "relative_key_query",
48
- "reference_compile": true,
49
  "repad_logits_with_grad": false,
50
  "sentence_transformers": {
51
  "activation_fn": "torch.nn.modules.activation.Sigmoid",
@@ -55,6 +53,6 @@
55
  "sparse_pred_ignore_index": -100,
56
  "sparse_prediction": false,
57
  "torch_dtype": "float32",
58
- "transformers_version": "4.48.2",
59
  "vocab_size": 50368
60
  }
 
1
  {
 
2
  "architectures": [
3
  "ModernBertForSequenceClassification"
4
  ],
 
33
  },
34
  "layer_norm_eps": 1e-05,
35
  "local_attention": 128,
36
+ "local_rope_theta": 40000,
37
  "max_position_embeddings": 8192,
38
  "mlp_bias": false,
39
  "mlp_dropout": 0.0,
 
44
  "num_hidden_layers": 22,
45
  "pad_token_id": 50283,
46
  "position_embedding_type": "relative_key_query",
 
47
  "repad_logits_with_grad": false,
48
  "sentence_transformers": {
49
  "activation_fn": "torch.nn.modules.activation.Sigmoid",
 
53
  "sparse_pred_ignore_index": -100,
54
  "sparse_prediction": false,
55
  "torch_dtype": "float32",
56
+ "transformers_version": "4.53.0",
57
  "vocab_size": 50368
58
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e522b6559192b6ff8aa79fa669123a4b6a9481cd7fd45b8c31fd4c00f61f7ccb
3
  size 598436708
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54306925a8ed2de4db707d8a282edb9d21cc10d5e89d4062dac5b5e8087ff73a
3
  size 598436708
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 640,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 1124,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
tokenizer_config.json CHANGED
@@ -933,12 +933,12 @@
933
  "cls_token": "[CLS]",
934
  "extra_special_tokens": {},
935
  "mask_token": "[MASK]",
936
- "max_length": 640,
937
  "model_input_names": [
938
  "input_ids",
939
  "attention_mask"
940
  ],
941
- "model_max_length": 640,
942
  "pad_to_multiple_of": null,
943
  "pad_token": "[PAD]",
944
  "pad_token_type_id": 0,
 
933
  "cls_token": "[CLS]",
934
  "extra_special_tokens": {},
935
  "mask_token": "[MASK]",
936
+ "max_length": 1124,
937
  "model_input_names": [
938
  "input_ids",
939
  "attention_mask"
940
  ],
941
+ "model_max_length": 8192,
942
  "pad_to_multiple_of": null,
943
  "pad_token": "[PAD]",
944
  "pad_token_type_id": 0,