NYTK
/

PULI-LlumiX-Llama-3.1 8B base (8.03B billion parameter)

Dataset for continued pretraining

  • Hungarian (8.08 billion words): documents (763K) that exceed 5000 words in length + Hungarian Wikipedia
  • English: Long Context QA (2 billion words), BookSum (78 million words)

Limitations

  • max_seq_length = 16 384
  • bfloat16

Usage with pipeline

from transformers import pipeline, LlamaForCausalLM, AutoTokenizer

model = LlamaForCausalLM.from_pretrained("NYTK/PULI-LlumiX-Llama-3.1")
tokenizer = AutoTokenizer.from_pretrained("NYTK/PULI-LlumiX-Llama-3.1")
prompt = "Elmesélek egy történetet a nyelvtechnológiáról."
generator = pipeline(task="text-generation", model=model, tokenizer=tokenizer, device=0)

print(generator(prompt, max_new_tokens=30)[0]["generated_text"])
Downloads last month
58
Safetensors
Model size
8.03B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for NYTK/PULI-LlumiX-Llama-3.1

Quantizations
1 model