Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,8 +1,9 @@
|
|
| 1 |
import torch
|
| 2 |
from transformers import GPT2Tokenizer, GPT2LMHeadModel
|
|
|
|
| 3 |
|
| 4 |
# Load the custom model and tokenizer
|
| 5 |
-
model_path =
|
| 6 |
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
|
| 7 |
model = GPT2LMHeadModel.from_pretrained(model_path)
|
| 8 |
|
|
@@ -21,10 +22,10 @@ def generate_response(prompt, model, tokenizer, max_length=100, num_beams=1, tem
|
|
| 21 |
max_length=max_length,
|
| 22 |
num_return_sequences=1,
|
| 23 |
pad_token_id=tokenizer.eos_token_id,
|
| 24 |
-
num_beams=num_beams,
|
| 25 |
temperature=temperature,
|
| 26 |
top_p=top_p,
|
| 27 |
-
repetition_penalty=repetition_penalty,
|
| 28 |
early_stopping=True
|
| 29 |
)
|
| 30 |
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
|
@@ -39,7 +40,7 @@ def generate_response(prompt, model, tokenizer, max_length=100, num_beams=1, tem
|
|
| 39 |
response = ' '.join(clean_response)
|
| 40 |
return response.strip()
|
| 41 |
|
| 42 |
-
def respond(message, history
|
| 43 |
# Prepare the prompt from the history and the new message
|
| 44 |
system_message = "You are a friendly chatbot."
|
| 45 |
conversation = system_message + "\n"
|
|
@@ -56,10 +57,10 @@ def respond(message, history: list[tuple[str, str]]):
|
|
| 56 |
|
| 57 |
return response
|
| 58 |
|
| 59 |
-
# Gradio Chat Interface
|
| 60 |
demo = gr.ChatInterface(
|
| 61 |
respond
|
| 62 |
)
|
| 63 |
|
| 64 |
if __name__ == "__main__":
|
| 65 |
-
demo.launch()
|
|
|
|
| 1 |
import torch
|
| 2 |
from transformers import GPT2Tokenizer, GPT2LMHeadModel
|
| 3 |
+
import gradio as gr
|
| 4 |
|
| 5 |
# Load the custom model and tokenizer
|
| 6 |
+
model_path = 'redael/model_udc'
|
| 7 |
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
|
| 8 |
model = GPT2LMHeadModel.from_pretrained(model_path)
|
| 9 |
|
|
|
|
| 22 |
max_length=max_length,
|
| 23 |
num_return_sequences=1,
|
| 24 |
pad_token_id=tokenizer.eos_token_id,
|
| 25 |
+
num_beams=num_beams,
|
| 26 |
temperature=temperature,
|
| 27 |
top_p=top_p,
|
| 28 |
+
repetition_penalty=repetition_penalty,
|
| 29 |
early_stopping=True
|
| 30 |
)
|
| 31 |
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
|
|
|
| 40 |
response = ' '.join(clean_response)
|
| 41 |
return response.strip()
|
| 42 |
|
| 43 |
+
def respond(message, history):
|
| 44 |
# Prepare the prompt from the history and the new message
|
| 45 |
system_message = "You are a friendly chatbot."
|
| 46 |
conversation = system_message + "\n"
|
|
|
|
| 57 |
|
| 58 |
return response
|
| 59 |
|
| 60 |
+
# Gradio Chat Interface
|
| 61 |
demo = gr.ChatInterface(
|
| 62 |
respond
|
| 63 |
)
|
| 64 |
|
| 65 |
if __name__ == "__main__":
|
| 66 |
+
demo.launch()
|