teojerah commited on
Commit
6877757
·
verified ·
1 Parent(s): c5db5fa

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +27 -27
app.py CHANGED
@@ -1,27 +1,27 @@
1
- import gradio as gr
2
- from transformers import AutoModelForCausalLM, AutoTokenizer
3
-
4
- # Load Dolly 2.0 model and tokenizer
5
- model_name = "databricks/dolly-v2-3b"
6
- model = AutoModelForCausalLM.from_pretrained(model_name)
7
- tokenizer = AutoTokenizer.from_pretrained(model_name)
8
-
9
- # Define a function to generate responses
10
- def dolly_response(prompt):
11
- inputs = tokenizer(prompt, return_tensors="pt")
12
- outputs = model.generate(
13
- inputs["input_ids"],
14
- max_length=100,
15
- num_beams=5,
16
- no_repeat_ngram_size=2,
17
- early_stopping=True
18
- )
19
- return tokenizer.decode(outputs[0], skip_special_tokens=True)
20
-
21
- # Create Gradio interface
22
- gr.Interface(
23
- fn=dolly_response,
24
- inputs="text",
25
- outputs="text",
26
- title="Dolly 2.0 Chatbot"
27
- ).launch()
 
1
+ import gradio as gr
2
+ from transformers import AutoModelForCausalLM, AutoTokenizer
3
+
4
+ # Load Dolly 2.0 model and tokenizer
5
+ model_name = "databricks/dolly-v2-3b"
6
+ model = AutoModelForCausalLM.from_pretrained(model_name)
7
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
8
+
9
+ # Define a function to generate responses
10
+ def dolly_response(prompt):
11
+ inputs = tokenizer(prompt, return_tensors="pt")
12
+ outputs = model.generate(
13
+ inputs["input_ids"],
14
+ max_length=50,
15
+ num_beams=3,
16
+ no_repeat_ngram_size=2,
17
+ early_stopping=True
18
+ )
19
+ return tokenizer.decode(outputs[0], skip_special_tokens=True)
20
+
21
+ # Create Gradio interface
22
+ gr.Interface(
23
+ fn=dolly_response,
24
+ inputs="text",
25
+ outputs="text",
26
+ title="Dolly 2.0 Chatbot"
27
+ ).launch()