--- base_model: gpt2-large inference: parameters: do_sample: true max_new_tokens: 256 temperature: 0.7 top_p: 0.9 language: - en - ru library_name: transformers license: apache-2.0 model_type: gpt2 pipeline_tag: text-generation tags: - safetensors - text-generation - conversational - machine-learning - nlp - transformer - russian - english - gpt2 - large --- # RadonSAI ## Overview RadonSAI is a variant of the Radon model family, based on the GPT2LMHeadModel architecture. ## Model Details - **Source Model**: gpt2-large - **Architecture**: GPT2LMHeadModel - **Parameters**: 772.2M - **Model Type**: gpt2 ## Usage ```python from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("MagistrTheOne/RadonSAI") model = AutoModelForCausalLM.from_pretrained("MagistrTheOne/RadonSAI") prompt = "Hello, how are you?" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=50) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ## Model Information - **Languages**: English, Russian - **License**: Apache 2.0 - **Format**: Safetensors - **Library**: Transformers ## Citation If you use this model, please cite the original source model and the Radon project.