File size: 1,226 Bytes
6c3f4f2 522fc1f 7295e60 6c3f4f2 7295e60 522fc1f d69d539 ccc8077 7295e60 947fb35 522fc1f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 |
# from transformers import MarianMTModel, MarianTokenizer
# def translate_text(text, src_lang="es", tgt_lang="en"):
# model_name = f"Helsinki-NLP/opus-mt-{src_lang}-{tgt_lang}"
# tokenizer = MarianTokenizer.from_pretrained(model_name)
# model = MarianMTModel.from_pretrained(model_name)
# inputs = tokenizer(text, return_tensors="pt", padding=True)
# translated = model.generate(**inputs)
# return tokenizer.decode(translated[0], skip_special_tokens=True)
# if __name__ == "__main__":
# input_text = "¿Cómo estás?"
# print("Translated Text:", translate_text(input_text, src_lang="es", tgt_lang="en"))
import spaces
from transformers import MarianMTModel, MarianTokenizer
# Preload the translation model globally
model_name = "Helsinki-NLP/opus-mt-mul-en"
tokenizer = MarianTokenizer.from_pretrained(model_name)
translation_model = MarianMTModel.from_pretrained(model_name)
@spaces.GPU
def translate_text(text, src_lang="auto", tgt_lang="en"):
"""Translate text from any language to English."""
inputs = tokenizer(text, return_tensors="pt", padding=True)
translated = translation_model.generate(**inputs)
return tokenizer.decode(translated[0], skip_special_tokens=True)
|