SwahiliGPT
This is a Mistral model that has been fine-tuned on the Wikipedia Swahili dataset.
Usage
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("mwitiderrick/SwahiliGPT_v0.1")
model = AutoModelForCausalLM.from_pretrained("mwitiderrick/SwahiliGPT_v0.1", device_map="auto")
inputs = tokenizer("Hapo zamani za kale", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200, do_sample=True, repetition_penalty=1.1)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
"""
Hapo zamani za kale katika historia ya jamii, ambavyo sehemu moja hutazama historia ile inayopendekezwa au inayojulikana, na sehemu nyingine inafanya history ambalai hainajulikana.
Utaifishaji unaleta utata kwanza mambo ya karne zilizoandamana, na seconda matokeo yanatokana na vipitio vya maisha muhimu ambavyo haivyo vitakuva mahitaji katika jamii hiyo (hunajua wakiweka mitindo katakatani). Ni kinyume kingine kwamba kuna sifa ambayo umechukizwa vitu hivi vilitengenezwa zaidi.
Katika Afrika Magharibi, historia huitwa ngan
"""
- Downloads last month
- 9
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Model tree for mwitiderrick/SwahiliGPT_v0.1
Base model
mistralai/Mistral-7B-v0.1