Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -9,9 +9,10 @@ def cged_correction(sentence, function):
|
|
| 9 |
prompt = {"錯別字校正":"糾正句子中的錯字:", "文法校正":"糾正句子中的錯誤:",
|
| 10 |
"文本重構":"在不改動文意的情況下改寫句子:", "文本簡化":"在不改動文意的情況下改寫句子:", "整體校正":"修改句子的錯誤或使其更通順:"}
|
| 11 |
#input_ids = tokenizer(prompt[function] + sentence, return_tensors="pt").input_ids
|
| 12 |
-
|
|
|
|
| 13 |
#edited_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
| 14 |
-
return
|
| 15 |
|
| 16 |
with gr.Blocks() as demo:
|
| 17 |
gr.Markdown(
|
|
|
|
| 9 |
prompt = {"錯別字校正":"糾正句子中的錯字:", "文法校正":"糾正句子中的錯誤:",
|
| 10 |
"文本重構":"在不改動文意的情況下改寫句子:", "文本簡化":"在不改動文意的情況下改寫句子:", "整體校正":"修改句子的錯誤或使其更通順:"}
|
| 11 |
#input_ids = tokenizer(prompt[function] + sentence, return_tensors="pt").input_ids
|
| 12 |
+
for _ in range(3):
|
| 13 |
+
sentence = model.predict([prompt[function] + sentence + "_輸出句:"])[0]
|
| 14 |
#edited_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
| 15 |
+
return sentence
|
| 16 |
|
| 17 |
with gr.Blocks() as demo:
|
| 18 |
gr.Markdown(
|