MagistrTheOne commited on
Commit
bd60771
·
verified ·
1 Parent(s): 40651cc

Create working RadonSAI with real weights

Browse files
README.md CHANGED
@@ -3,13 +3,19 @@ license: apache-2.0
3
  tags:
4
  - radon
5
  - gpt2
6
- - 2000mb
7
- - fixed
8
  ---
9
 
10
- # RadonSAI (Fixed)
11
 
12
- Исправленная версия RadonSAI с рабочей конфигурацией.
 
 
 
 
 
 
13
 
14
  ## Использование
15
  ```python
@@ -17,4 +23,26 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
17
 
18
  model = AutoModelForCausalLM.from_pretrained("MagistrTheOne/RadonSAI")
19
  tokenizer = AutoTokenizer.from_pretrained("MagistrTheOne/RadonSAI")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  ```
 
3
  tags:
4
  - radon
5
  - gpt2
6
+ - working
7
+ - 1024d
8
  ---
9
 
10
+ # RadonSAI (Working)
11
 
12
+ Полностью рабочая версия RadonSAI с реальными весами.
13
+
14
+ ## Технические характеристики
15
+ - Параметры: ~203,668,480
16
+ - Размерность: 1024
17
+ - Слои: 12
18
+ - Головы внимания: 16
19
 
20
  ## Использование
21
  ```python
 
23
 
24
  model = AutoModelForCausalLM.from_pretrained("MagistrTheOne/RadonSAI")
25
  tokenizer = AutoTokenizer.from_pretrained("MagistrTheOne/RadonSAI")
26
+
27
+ # Генерация текста
28
+ input_text = "Hello, how are you?"
29
+ inputs = tokenizer(input_text, return_tensors="pt")
30
+ outputs = model.generate(**inputs, max_length=50)
31
+ result = tokenizer.decode(outputs[0], skip_special_tokens=True)
32
+ print(result)
33
+ ```
34
+
35
+ ## Тестирование
36
+ ```python
37
+ # Быстрый тест
38
+ import torch
39
+ model = AutoModelForCausalLM.from_pretrained("MagistrTheOne/RadonSAI")
40
+ tokenizer = AutoTokenizer.from_pretrained("MagistrTheOne/RadonSAI")
41
+
42
+ # Простая генерация
43
+ text = "The future of AI is"
44
+ inputs = tokenizer(text, return_tensors="pt")
45
+ with torch.no_grad():
46
+ outputs = model.generate(**inputs, max_length=30, do_sample=True, temperature=0.7)
47
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
48
  ```
config.json CHANGED
@@ -1,28 +1,31 @@
1
  {
 
2
  "architectures": [
3
  "GPT2LMHeadModel"
4
  ],
5
- "model_type": "gpt2",
6
- "n_ctx": 1024,
7
- "n_embd": 1024,
8
- "n_head": 16,
9
- "n_layer": 12,
10
- "n_positions": 1024,
11
- "vocab_size": 50257,
12
- "torch_dtype": "float16",
13
- "transformers_version": "4.36.2",
14
- "use_cache": true,
15
- "attention_dropout": 0.0,
16
  "attn_pdrop": 0.1,
17
  "bos_token_id": 50256,
18
- "eos_token_id": 50256,
19
  "embd_pdrop": 0.1,
 
20
  "initializer_range": 0.02,
21
  "layer_norm_epsilon": 1e-05,
 
 
 
 
 
 
 
22
  "resid_pdrop": 0.1,
 
 
23
  "summary_activation": null,
24
  "summary_first_dropout": 0.1,
25
  "summary_proj_to_labels": true,
26
  "summary_type": "cls_index",
27
- "summary_use_proj": true
28
- }
 
 
 
 
 
1
  {
2
+ "activation_function": "gelu_new",
3
  "architectures": [
4
  "GPT2LMHeadModel"
5
  ],
 
 
 
 
 
 
 
 
 
 
 
6
  "attn_pdrop": 0.1,
7
  "bos_token_id": 50256,
 
8
  "embd_pdrop": 0.1,
9
+ "eos_token_id": 50256,
10
  "initializer_range": 0.02,
11
  "layer_norm_epsilon": 1e-05,
12
+ "model_type": "gpt2",
13
+ "n_embd": 1024,
14
+ "n_head": 16,
15
+ "n_inner": null,
16
+ "n_layer": 12,
17
+ "n_positions": 1024,
18
+ "reorder_and_upcast_attn": false,
19
  "resid_pdrop": 0.1,
20
+ "scale_attn_by_inverse_layer_idx": false,
21
+ "scale_attn_weights": true,
22
  "summary_activation": null,
23
  "summary_first_dropout": 0.1,
24
  "summary_proj_to_labels": true,
25
  "summary_type": "cls_index",
26
+ "summary_use_proj": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.36.2",
29
+ "use_cache": true,
30
+ "vocab_size": 50257
31
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "4.36.2"
6
+ }
merges.txt CHANGED
The diff for this file is too large to render. See raw diff
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3323be5474d086a52bf5c73dee21b5438b501c1f4b007342edbe6cada51e25c9
3
- size 131278312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e4cb54b9ebcdc8a72ebb3265c871023c38189e7d6851a87641fe6aeafce66ca
3
+ size 814689096
special_tokens_map.json CHANGED
@@ -13,6 +13,7 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
16
  "unk_token": {
17
  "content": "<|endoftext|>",
18
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": "<|endoftext|>",
17
  "unk_token": {
18
  "content": "<|endoftext|>",
19
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -1,14 +1,22 @@
1
  {
2
- "auto_map": {
3
- "AutoTokenizer": [
4
- "gpt2",
5
- null
6
- ]
 
 
 
 
 
 
7
  },
8
  "bos_token": "<|endoftext|>",
 
9
  "eos_token": "<|endoftext|>",
 
10
  "model_max_length": 1024,
11
  "pad_token": "<|endoftext|>",
12
  "tokenizer_class": "GPT2Tokenizer",
13
  "unk_token": "<|endoftext|>"
14
- }
 
1
  {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "50256": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ }
13
  },
14
  "bos_token": "<|endoftext|>",
15
+ "clean_up_tokenization_spaces": true,
16
  "eos_token": "<|endoftext|>",
17
+ "errors": "replace",
18
  "model_max_length": 1024,
19
  "pad_token": "<|endoftext|>",
20
  "tokenizer_class": "GPT2Tokenizer",
21
  "unk_token": "<|endoftext|>"
22
+ }
vocab.json CHANGED
The diff for this file is too large to render. See raw diff