Spaces:
Runtime error
Runtime error
Commit
·
77582ad
1
Parent(s):
6a590e1
07/01/23
Browse files
app.py
CHANGED
|
@@ -2,13 +2,7 @@ import re
|
|
| 2 |
import gradio as gr
|
| 3 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 4 |
|
| 5 |
-
model = AutoModelForCausalLM.from_pretrained(
|
| 6 |
-
'parsak/codegen-350M-mono-lora-instruction',
|
| 7 |
-
)
|
| 8 |
-
tokenizer = AutoTokenizer.from_pretrained('Salesforce/codegen-350M-mono')
|
| 9 |
|
| 10 |
-
tokenizer.pad_token_id = 0 # different to <eos>
|
| 11 |
-
tokenizer.padding_side = "left" # Allow batched inference
|
| 12 |
|
| 13 |
def extract_code(input_text):
|
| 14 |
pattern = r"'''py\n(.*?)'''"
|
|
@@ -19,28 +13,44 @@ def extract_code(input_text):
|
|
| 19 |
else:
|
| 20 |
return None # Return None if no match is found
|
| 21 |
|
| 22 |
-
def generate_code(input_text):
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
|
| 28 |
def respond(message, chat_history, additional_inputs):
|
| 29 |
-
return f"Here's an example code:\n\n```python\n{generate_code(message)}\n```"
|
| 30 |
|
| 31 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 32 |
|
| 33 |
|
|
|
|
| 34 |
|
| 35 |
-
|
| 36 |
-
gr.ChatInterface(respond,
|
| 37 |
retry_btn= gr.Button(value="Retry"),
|
| 38 |
undo_btn=None, clear_btn=gr.Button(value="Clear"),
|
| 39 |
additional_inputs=[
|
| 40 |
-
|
| 41 |
]
|
| 42 |
-
)
|
| 43 |
-
|
| 44 |
|
| 45 |
|
|
|
|
|
|
|
| 46 |
|
|
|
|
| 2 |
import gradio as gr
|
| 3 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 4 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
|
|
|
|
|
|
|
| 6 |
|
| 7 |
def extract_code(input_text):
|
| 8 |
pattern = r"'''py\n(.*?)'''"
|
|
|
|
| 13 |
else:
|
| 14 |
return None # Return None if no match is found
|
| 15 |
|
| 16 |
+
def generate_code(input_text,modelName):
|
| 17 |
+
if(modelName == "codegen-350M"):
|
| 18 |
+
input_ids = codeGenTokenizer(input_text, return_tensors="pt").input_ids
|
| 19 |
+
generated_ids = codeGenModel.generate(input_ids, max_length=128)
|
| 20 |
+
result = codeGenTokenizer.decode(generated_ids[0], skip_special_tokens=True)
|
| 21 |
+
return extract_code(result)
|
| 22 |
+
elif(modelName == "mistral-7b"):
|
| 23 |
+
input_ids = mistralTokenizer(input_text, return_tensors="pt").input_ids
|
| 24 |
+
generated_ids = mistralModel.generate(input_ids, max_length=128)
|
| 25 |
+
result = mistralTokenizer.decode(generated_ids[0], skip_special_tokens=True)
|
| 26 |
+
return result
|
| 27 |
+
else:
|
| 28 |
+
return None
|
| 29 |
+
|
| 30 |
|
| 31 |
def respond(message, chat_history, additional_inputs):
|
| 32 |
+
return f"Here's an example code:\n\n```python\n{generate_code(message,additional_inputs)}\n```"
|
| 33 |
|
| 34 |
|
| 35 |
+
codeGenModel = AutoModelForCausalLM.from_pretrained('parsak/codegen-350M-mono-lora-instruction')
|
| 36 |
+
mistralModel = AutoModelForCausalLM.from_pretrained('parsak/mistral-code-7b-instruct')
|
| 37 |
+
codeGenTokenizer = AutoTokenizer.from_pretrained('Salesforce/codegen-350M-mono')
|
| 38 |
+
mistralTokenizer = AutoTokenizer.from_pretrained('parsak/mistral-code-7b-instruct')
|
| 39 |
+
codeGenTokenizer.pad_token_id = 0
|
| 40 |
+
codeGenTokenizer.padding_side = "left"
|
| 41 |
|
| 42 |
|
| 43 |
+
dropdown = gr.Dropdown(label="Models",choices=["codegen-350M", "mistral-7b"], value="codegen-350M")
|
| 44 |
|
| 45 |
+
interface = gr.ChatInterface(respond,
|
|
|
|
| 46 |
retry_btn= gr.Button(value="Retry"),
|
| 47 |
undo_btn=None, clear_btn=gr.Button(value="Clear"),
|
| 48 |
additional_inputs=[
|
| 49 |
+
dropdown
|
| 50 |
]
|
| 51 |
+
)
|
|
|
|
| 52 |
|
| 53 |
|
| 54 |
+
if __name__ == "__main__":
|
| 55 |
+
interface.launch()
|
| 56 |
|