--- base_model: unsloth/gemma-2-9b-bnb-4bit library_name: peft license: apache-2.0 language: - en - mr --- # Model Card for Gemma2 7B - English to Marathi Translation ## Model Details ### Model Description This model is a fine-tuned variant of **Unsloth's Gemma2 7B**, trained for high-quality English-to-Marathi translations. Built on a robust transformer architecture, the model handles complex translations, idiomatic expressions, and long-context paragraphs effectively. It is optimized for efficient inference using 4-bit quantization. - **Developed by:** Devavrat Samak - **Model type:** Causal Language Model, fine-tuned for translation tasks. - **Language(s) (NLP):** English (en), Marathi (mr) - **License:** Apache-2.0 - **Finetuned from model:** unsloth/gemma-2-9b-bnb-4bit ### Model Sources - **Repository:** [https://github.com/Devsam2898/Gemma2-Marathi] ## Uses ### Direct Use The model can be directly used for English-to-Marathi translations, including handling long-context paragraphs, noisy inputs, and code-mixed sentences. ### Downstream Use The model can be integrated into applications for: - Chatbots with multilingual support. - Translating historical texts for research. - Localization of content for Marathi-speaking audiences. ### Out-of-Scope Use - The model is not designed for real-time, high-speed translation in latency-critical systems. - It may not generalize well for highly domain-specific jargon without additional fine-tuning. ## Bias, Risks, and Limitations - The model's translations might occasionally lose nuance or context in culturally significant expressions. - Performance may degrade for noisy data or highly informal text. ### Recommendations - Users should validate translations in sensitive domains to ensure accuracy. - Consider additional fine-tuning for domain-specific tasks. ## How to Get Started with the Model ```python from transformers import AutoTokenizer from unsloth import Gemma2 # Load model and tokenizer model = Gemma2.from_pretrained("unsloth/gemma-2-9b-bnb-4bit") tokenizer = AutoTokenizer.from_pretrained("unsloth/gemma-2-9b-bnb-4bit") # Input and inference input_text = "The golden age of the Peshwas brought cultural and political prosperity to Maharashtra." inputs = tokenizer(input_text, return_tensors="pt") outputs = model.generate(inputs["input_ids"], max_length=128) translated_text = tokenizer.decode(outputs[0], skip_special_tokens=True) print(translated_text)