NYTK
/

PULI-LlumiX-Llama-3.1 8B base (8.03B billion parameter)

Dataset for continued pretraining

  • Hungarian (8.08 billion words): documents (763K) that exceed 5000 words in length + Hungarian Wikipedia
  • English: Long Context QA (2 billion words), BookSum (78 million words)

Limitations

  • max_seq_length = 16 384
  • bfloat16

Usage with pipeline

from transformers import pipeline, LlamaForCausalLM, AutoTokenizer

model = LlamaForCausalLM.from_pretrained("NYTK/PULI-LlumiX-Llama-3.1")
tokenizer = AutoTokenizer.from_pretrained("NYTK/PULI-LlumiX-Llama-3.1")
prompt = "Elmes茅lek egy t枚rt茅netet a nyelvtechnol贸gi谩r贸l."
generator = pipeline(task="text-generation", model=model, tokenizer=tokenizer, device=0)

print(generator(prompt, max_new_tokens=30)[0]["generated_text"])
Downloads last month
33
Safetensors
Model size
8.03B params
Tensor type
BF16
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for NYTK/PULI-LlumiX-Llama-3.1

Quantizations
1 model