from transformers import AutoModelForSeq2SeqLM, AutoTokenizer # Load model and tokenizer model_name = "hima06varshini/english-to-telugu-translation" model = AutoModelForSeq2SeqLM.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name) def translate(text): inputs = tokenizer(text, return_tensors="pt") outputs = model.generate(**inputs) return tokenizer.decode(outputs[0], skip_special_tokens=True) # Example usage text = "Hello, how are you?" translation = translate(text) print("Translated text:", translation)