SwahiliGPT
This is a Mistral model that has been fine-tuned on the Wikipedia Swahili dataset.
Usage
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("mwitiderrick/SwahiliGPT_v0.1")
model = AutoModelForCausalLM.from_pretrained("mwitiderrick/SwahiliGPT_v0.1", device_map="auto")
inputs = tokenizer("Hapo zamani za kale", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200, do_sample=True, repetition_penalty=1.1)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
"""
Hapo zamani za kale katika historia ya jamii, ambavyo sehemu moja hutazama historia ile inayopendekezwa au inayojulikana, na sehemu nyingine inafanya history ambalai hainajulikana.
Utaifishaji unaleta utata kwanza mambo ya karne zilizoandamana, na seconda matokeo yanatokana na vipitio vya maisha muhimu ambavyo haivyo vitakuva mahitaji katika jamii hiyo (hunajua wakiweka mitindo katakatani). Ni kinyume kingine kwamba kuna sifa ambayo umechukizwa vitu hivi vilitengenezwa zaidi.
Katika Afrika Magharibi, historia huitwa ngan
"""
- Downloads last month
- 23
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Model tree for mwitiderrick/SwahiliGPT_v0.1
Base model
mistralai/Mistral-7B-v0.1