import os import gradio as gr from transformers import AutoTokenizer, AutoModelForSeq2SeqLM, pipeline def text(input): from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-en-de" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) mname = "facebook/wmt19-de-en" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input_ids = tokenizer.encode(decoded, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) return decoded iface = gr.Interface(fn=text, inputs=[ gr.inputs.Textbox( lines=2, placeholder=None, label='Sentence'), ], outputs=[gr.outputs.JSON(label=None)]) iface.launch()